var/home/core/zuul-output/0000755000175000017500000000000015070710752014531 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015070722237015476 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000004617430415070722231017705 0ustar rootrootOct 06 09:58:09 crc systemd[1]: Starting Kubernetes Kubelet... Oct 06 09:58:09 crc restorecon[4665]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:09 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 06 09:58:10 crc restorecon[4665]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 06 09:58:10 crc restorecon[4665]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Oct 06 09:58:11 crc kubenswrapper[4824]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 06 09:58:11 crc kubenswrapper[4824]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Oct 06 09:58:11 crc kubenswrapper[4824]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 06 09:58:11 crc kubenswrapper[4824]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 06 09:58:11 crc kubenswrapper[4824]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Oct 06 09:58:11 crc kubenswrapper[4824]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.015946 4824 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.020216 4824 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.020234 4824 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.020239 4824 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.020243 4824 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.020247 4824 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.020253 4824 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.020257 4824 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.020262 4824 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.020267 4824 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.020272 4824 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.020277 4824 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.020282 4824 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.020287 4824 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.020300 4824 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.020305 4824 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.020310 4824 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.020316 4824 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.020321 4824 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.020326 4824 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.020331 4824 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.020335 4824 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.020339 4824 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.020344 4824 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.020348 4824 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.020352 4824 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.020356 4824 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.020360 4824 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.020365 4824 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.020369 4824 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.020373 4824 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.020378 4824 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.020382 4824 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.020387 4824 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.020391 4824 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.020394 4824 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.020399 4824 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.020403 4824 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.020407 4824 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.020411 4824 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.020415 4824 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.020419 4824 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.020425 4824 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.020431 4824 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.020435 4824 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.020440 4824 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.020444 4824 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.020449 4824 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.020453 4824 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.020457 4824 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.020461 4824 feature_gate.go:330] unrecognized feature gate: Example Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.020465 4824 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.020468 4824 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.020472 4824 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.020476 4824 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.020479 4824 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.020483 4824 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.020486 4824 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.020490 4824 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.020494 4824 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.020497 4824 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.020501 4824 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.020505 4824 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.020509 4824 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.020513 4824 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.020517 4824 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.020521 4824 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.020525 4824 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.020529 4824 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.020534 4824 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.020539 4824 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.020544 4824 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.020669 4824 flags.go:64] FLAG: --address="0.0.0.0" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.020679 4824 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.020687 4824 flags.go:64] FLAG: --anonymous-auth="true" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.020695 4824 flags.go:64] FLAG: --application-metrics-count-limit="100" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.020702 4824 flags.go:64] FLAG: --authentication-token-webhook="false" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.020707 4824 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.020714 4824 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.020720 4824 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.020725 4824 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.020731 4824 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.020737 4824 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.020743 4824 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.020748 4824 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.020753 4824 flags.go:64] FLAG: --cgroup-root="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.020759 4824 flags.go:64] FLAG: --cgroups-per-qos="true" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.020764 4824 flags.go:64] FLAG: --client-ca-file="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.020769 4824 flags.go:64] FLAG: --cloud-config="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.020774 4824 flags.go:64] FLAG: --cloud-provider="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.020779 4824 flags.go:64] FLAG: --cluster-dns="[]" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.020786 4824 flags.go:64] FLAG: --cluster-domain="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.020791 4824 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.020796 4824 flags.go:64] FLAG: --config-dir="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.020801 4824 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.020807 4824 flags.go:64] FLAG: --container-log-max-files="5" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.020814 4824 flags.go:64] FLAG: --container-log-max-size="10Mi" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.020820 4824 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.020825 4824 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.020832 4824 flags.go:64] FLAG: --containerd-namespace="k8s.io" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.020839 4824 flags.go:64] FLAG: --contention-profiling="false" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.020844 4824 flags.go:64] FLAG: --cpu-cfs-quota="true" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.020851 4824 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.020857 4824 flags.go:64] FLAG: --cpu-manager-policy="none" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.020862 4824 flags.go:64] FLAG: --cpu-manager-policy-options="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.020869 4824 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.020874 4824 flags.go:64] FLAG: --enable-controller-attach-detach="true" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.020880 4824 flags.go:64] FLAG: --enable-debugging-handlers="true" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.020884 4824 flags.go:64] FLAG: --enable-load-reader="false" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.020889 4824 flags.go:64] FLAG: --enable-server="true" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.020894 4824 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.020900 4824 flags.go:64] FLAG: --event-burst="100" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.020905 4824 flags.go:64] FLAG: --event-qps="50" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.020910 4824 flags.go:64] FLAG: --event-storage-age-limit="default=0" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.020914 4824 flags.go:64] FLAG: --event-storage-event-limit="default=0" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.020919 4824 flags.go:64] FLAG: --eviction-hard="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.020925 4824 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.020929 4824 flags.go:64] FLAG: --eviction-minimum-reclaim="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.020934 4824 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.020939 4824 flags.go:64] FLAG: --eviction-soft="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.020943 4824 flags.go:64] FLAG: --eviction-soft-grace-period="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.020947 4824 flags.go:64] FLAG: --exit-on-lock-contention="false" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.020952 4824 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.020956 4824 flags.go:64] FLAG: --experimental-mounter-path="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.020960 4824 flags.go:64] FLAG: --fail-cgroupv1="false" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.020965 4824 flags.go:64] FLAG: --fail-swap-on="true" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.020969 4824 flags.go:64] FLAG: --feature-gates="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.020990 4824 flags.go:64] FLAG: --file-check-frequency="20s" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.020995 4824 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.021000 4824 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.021005 4824 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.021010 4824 flags.go:64] FLAG: --healthz-port="10248" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.021015 4824 flags.go:64] FLAG: --help="false" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.021020 4824 flags.go:64] FLAG: --hostname-override="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.021025 4824 flags.go:64] FLAG: --housekeeping-interval="10s" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.021030 4824 flags.go:64] FLAG: --http-check-frequency="20s" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.021036 4824 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.021040 4824 flags.go:64] FLAG: --image-credential-provider-config="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.021044 4824 flags.go:64] FLAG: --image-gc-high-threshold="85" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.021052 4824 flags.go:64] FLAG: --image-gc-low-threshold="80" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.021057 4824 flags.go:64] FLAG: --image-service-endpoint="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.021062 4824 flags.go:64] FLAG: --kernel-memcg-notification="false" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.021067 4824 flags.go:64] FLAG: --kube-api-burst="100" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.021071 4824 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.021076 4824 flags.go:64] FLAG: --kube-api-qps="50" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.021081 4824 flags.go:64] FLAG: --kube-reserved="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.021085 4824 flags.go:64] FLAG: --kube-reserved-cgroup="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.021090 4824 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.021094 4824 flags.go:64] FLAG: --kubelet-cgroups="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.021099 4824 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.021103 4824 flags.go:64] FLAG: --lock-file="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.021108 4824 flags.go:64] FLAG: --log-cadvisor-usage="false" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.021122 4824 flags.go:64] FLAG: --log-flush-frequency="5s" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.021127 4824 flags.go:64] FLAG: --log-json-info-buffer-size="0" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.021134 4824 flags.go:64] FLAG: --log-json-split-stream="false" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.021139 4824 flags.go:64] FLAG: --log-text-info-buffer-size="0" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.021143 4824 flags.go:64] FLAG: --log-text-split-stream="false" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.021148 4824 flags.go:64] FLAG: --logging-format="text" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.021152 4824 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.021157 4824 flags.go:64] FLAG: --make-iptables-util-chains="true" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.021162 4824 flags.go:64] FLAG: --manifest-url="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.021167 4824 flags.go:64] FLAG: --manifest-url-header="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.021174 4824 flags.go:64] FLAG: --max-housekeeping-interval="15s" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.021180 4824 flags.go:64] FLAG: --max-open-files="1000000" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.021187 4824 flags.go:64] FLAG: --max-pods="110" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.021192 4824 flags.go:64] FLAG: --maximum-dead-containers="-1" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.021198 4824 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.021203 4824 flags.go:64] FLAG: --memory-manager-policy="None" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.021208 4824 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.021213 4824 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.021217 4824 flags.go:64] FLAG: --node-ip="192.168.126.11" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.021224 4824 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.021236 4824 flags.go:64] FLAG: --node-status-max-images="50" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.021241 4824 flags.go:64] FLAG: --node-status-update-frequency="10s" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.021245 4824 flags.go:64] FLAG: --oom-score-adj="-999" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.021250 4824 flags.go:64] FLAG: --pod-cidr="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.021254 4824 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.021262 4824 flags.go:64] FLAG: --pod-manifest-path="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.021266 4824 flags.go:64] FLAG: --pod-max-pids="-1" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.021271 4824 flags.go:64] FLAG: --pods-per-core="0" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.021275 4824 flags.go:64] FLAG: --port="10250" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.021280 4824 flags.go:64] FLAG: --protect-kernel-defaults="false" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.021284 4824 flags.go:64] FLAG: --provider-id="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.021288 4824 flags.go:64] FLAG: --qos-reserved="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.021292 4824 flags.go:64] FLAG: --read-only-port="10255" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.021296 4824 flags.go:64] FLAG: --register-node="true" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.021301 4824 flags.go:64] FLAG: --register-schedulable="true" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.021305 4824 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.021313 4824 flags.go:64] FLAG: --registry-burst="10" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.021318 4824 flags.go:64] FLAG: --registry-qps="5" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.021322 4824 flags.go:64] FLAG: --reserved-cpus="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.021327 4824 flags.go:64] FLAG: --reserved-memory="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.021333 4824 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.021338 4824 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.021342 4824 flags.go:64] FLAG: --rotate-certificates="false" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.021348 4824 flags.go:64] FLAG: --rotate-server-certificates="false" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.021352 4824 flags.go:64] FLAG: --runonce="false" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.021357 4824 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.021362 4824 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.021366 4824 flags.go:64] FLAG: --seccomp-default="false" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.021371 4824 flags.go:64] FLAG: --serialize-image-pulls="true" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.021375 4824 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.021380 4824 flags.go:64] FLAG: --storage-driver-db="cadvisor" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.021385 4824 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.021390 4824 flags.go:64] FLAG: --storage-driver-password="root" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.021395 4824 flags.go:64] FLAG: --storage-driver-secure="false" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.021400 4824 flags.go:64] FLAG: --storage-driver-table="stats" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.021405 4824 flags.go:64] FLAG: --storage-driver-user="root" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.021409 4824 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.021414 4824 flags.go:64] FLAG: --sync-frequency="1m0s" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.021418 4824 flags.go:64] FLAG: --system-cgroups="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.021423 4824 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.021429 4824 flags.go:64] FLAG: --system-reserved-cgroup="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.021434 4824 flags.go:64] FLAG: --tls-cert-file="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.021439 4824 flags.go:64] FLAG: --tls-cipher-suites="[]" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.021444 4824 flags.go:64] FLAG: --tls-min-version="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.021449 4824 flags.go:64] FLAG: --tls-private-key-file="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.021454 4824 flags.go:64] FLAG: --topology-manager-policy="none" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.021458 4824 flags.go:64] FLAG: --topology-manager-policy-options="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.021463 4824 flags.go:64] FLAG: --topology-manager-scope="container" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.021468 4824 flags.go:64] FLAG: --v="2" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.021475 4824 flags.go:64] FLAG: --version="false" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.021483 4824 flags.go:64] FLAG: --vmodule="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.021489 4824 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.021495 4824 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.021620 4824 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.021626 4824 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.021631 4824 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.021636 4824 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.021640 4824 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.021645 4824 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.021649 4824 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.021652 4824 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.021656 4824 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.021660 4824 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.021664 4824 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.021669 4824 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.021674 4824 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.021679 4824 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.021684 4824 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.021688 4824 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.021694 4824 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.021700 4824 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.021705 4824 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.021710 4824 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.021715 4824 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.021719 4824 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.021726 4824 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.021731 4824 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.021738 4824 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.021744 4824 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.021749 4824 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.021755 4824 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.021761 4824 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.021766 4824 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.021771 4824 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.021776 4824 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.021780 4824 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.021784 4824 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.021789 4824 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.021793 4824 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.021797 4824 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.021801 4824 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.021806 4824 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.021810 4824 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.021814 4824 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.021819 4824 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.021823 4824 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.021828 4824 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.021832 4824 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.021836 4824 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.021839 4824 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.021843 4824 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.021847 4824 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.021851 4824 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.021854 4824 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.021858 4824 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.021862 4824 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.021865 4824 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.021869 4824 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.021873 4824 feature_gate.go:330] unrecognized feature gate: Example Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.021877 4824 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.021881 4824 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.021885 4824 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.021889 4824 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.021894 4824 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.021899 4824 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.021904 4824 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.021907 4824 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.021911 4824 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.021915 4824 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.021919 4824 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.021922 4824 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.021926 4824 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.021930 4824 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.021934 4824 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.023043 4824 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.037100 4824 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.037127 4824 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.037283 4824 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.037293 4824 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.037299 4824 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.037305 4824 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.037309 4824 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.037316 4824 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.037322 4824 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.037329 4824 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.037344 4824 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.037350 4824 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.037355 4824 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.037359 4824 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.037364 4824 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.037369 4824 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.037373 4824 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.037377 4824 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.037382 4824 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.037387 4824 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.037391 4824 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.037395 4824 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.037405 4824 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.037410 4824 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.037414 4824 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.037419 4824 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.037423 4824 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.037428 4824 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.037432 4824 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.037436 4824 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.037442 4824 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.037447 4824 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.037451 4824 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.037457 4824 feature_gate.go:330] unrecognized feature gate: Example Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.037462 4824 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.037471 4824 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.037478 4824 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.037483 4824 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.037489 4824 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.037494 4824 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.037499 4824 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.037504 4824 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.037509 4824 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.037514 4824 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.037518 4824 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.037523 4824 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.037528 4824 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.037538 4824 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.037543 4824 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.037549 4824 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.037554 4824 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.037559 4824 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.037565 4824 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.037571 4824 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.037576 4824 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.037581 4824 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.037586 4824 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.037590 4824 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.037595 4824 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.037600 4824 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.037610 4824 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.037614 4824 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.037619 4824 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.037624 4824 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.037628 4824 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.037633 4824 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.037637 4824 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.037642 4824 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.037646 4824 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.037651 4824 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.037656 4824 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.037661 4824 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.037668 4824 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.037679 4824 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.038101 4824 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.038154 4824 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.038170 4824 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.038184 4824 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.038195 4824 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.038206 4824 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.038214 4824 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.038223 4824 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.038232 4824 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.038252 4824 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.038261 4824 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.038269 4824 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.038278 4824 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.038328 4824 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.038338 4824 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.038351 4824 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.038775 4824 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.038786 4824 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.038794 4824 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.038802 4824 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.038811 4824 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.038819 4824 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.038828 4824 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.038837 4824 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.038845 4824 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.038854 4824 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.038862 4824 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.038870 4824 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.038878 4824 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.038886 4824 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.038900 4824 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.038911 4824 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.038922 4824 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.038932 4824 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.038940 4824 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.038948 4824 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.038956 4824 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.038965 4824 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.038975 4824 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.039012 4824 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.039020 4824 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.039028 4824 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.039041 4824 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.039052 4824 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.039062 4824 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.039070 4824 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.039079 4824 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.039087 4824 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.039095 4824 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.039103 4824 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.039111 4824 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.039119 4824 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.039152 4824 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.039160 4824 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.039168 4824 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.039176 4824 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.039184 4824 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.039192 4824 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.039200 4824 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.039207 4824 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.039215 4824 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.039223 4824 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.039231 4824 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.039239 4824 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.039249 4824 feature_gate.go:330] unrecognized feature gate: Example Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.039257 4824 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.039265 4824 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.039274 4824 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.039281 4824 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.039290 4824 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.039298 4824 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.039315 4824 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.040714 4824 server.go:940] "Client rotation is on, will bootstrap in background" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.046777 4824 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.046927 4824 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.048967 4824 server.go:997] "Starting client certificate rotation" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.049025 4824 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.049294 4824 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-12-15 13:41:46.712300752 +0000 UTC Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.049392 4824 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 1683h43m35.662913488s for next certificate rotation Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.081561 4824 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.085333 4824 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.104397 4824 log.go:25] "Validated CRI v1 runtime API" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.144840 4824 log.go:25] "Validated CRI v1 image API" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.146971 4824 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.154108 4824 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-10-06-09-53-51-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.154154 4824 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:45 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:42 fsType:tmpfs blockSize:0}] Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.173635 4824 manager.go:217] Machine: {Timestamp:2025-10-06 09:58:11.170694502 +0000 UTC m=+0.535117383 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2799998 MemoryCapacity:33654132736 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:77448683-227c-4c6b-a334-eb212350758a BootID:740c2e7e-c94c-411f-b080-eb42c9aecdca Filesystems:[{Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:45 Capacity:3365412864 Type:vfs Inodes:821634 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:42 Capacity:1073741824 Type:vfs Inodes:4108170 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827064320 Type:vfs Inodes:4108170 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730829824 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827068416 Type:vfs Inodes:1048576 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:b9:31:53 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:b9:31:53 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:bc:42:90 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:f1:99:1b Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:13:d0:07 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:45:09:2b Speed:-1 Mtu:1496} {Name:eth10 MacAddress:9e:4e:54:43:f6:2d Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:5e:35:ec:cd:bf:2c Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654132736 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.173854 4824 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.173970 4824 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.175155 4824 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.175312 4824 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.175354 4824 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.175577 4824 topology_manager.go:138] "Creating topology manager with none policy" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.175588 4824 container_manager_linux.go:303] "Creating device plugin manager" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.176177 4824 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.176213 4824 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.176503 4824 state_mem.go:36] "Initialized new in-memory state store" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.176581 4824 server.go:1245] "Using root directory" path="/var/lib/kubelet" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.183024 4824 kubelet.go:418] "Attempting to sync node with API server" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.183049 4824 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.183092 4824 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.183107 4824 kubelet.go:324] "Adding apiserver pod source" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.183118 4824 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.189859 4824 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.191840 4824 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.193158 4824 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.50:6443: connect: connection refused Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.193181 4824 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.50:6443: connect: connection refused Oct 06 09:58:11 crc kubenswrapper[4824]: E1006 09:58:11.193250 4824 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.50:6443: connect: connection refused" logger="UnhandledError" Oct 06 09:58:11 crc kubenswrapper[4824]: E1006 09:58:11.193299 4824 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.50:6443: connect: connection refused" logger="UnhandledError" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.194128 4824 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.195866 4824 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.195909 4824 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.195925 4824 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.195938 4824 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.195960 4824 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.195974 4824 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.196013 4824 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.196059 4824 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.196077 4824 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.196091 4824 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.196110 4824 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.196123 4824 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.198309 4824 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.199009 4824 server.go:1280] "Started kubelet" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.200242 4824 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.200546 4824 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.200837 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.200871 4824 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.201015 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-17 00:21:46.955490994 +0000 UTC Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.201064 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 2462h23m35.754430699s for next certificate rotation Oct 06 09:58:11 crc systemd[1]: Started Kubernetes Kubelet. Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.201189 4824 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Oct 06 09:58:11 crc kubenswrapper[4824]: E1006 09:58:11.201750 4824 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.202015 4824 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.201642 4824 volume_manager.go:287] "The desired_state_of_world populator starts" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.202352 4824 volume_manager.go:289] "Starting Kubelet Volume Manager" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.202896 4824 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.50:6443: connect: connection refused Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.203585 4824 factory.go:153] Registering CRI-O factory Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.203641 4824 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.50:6443: connect: connection refused Oct 06 09:58:11 crc kubenswrapper[4824]: E1006 09:58:11.203761 4824 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.50:6443: connect: connection refused" logger="UnhandledError" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.204312 4824 factory.go:221] Registration of the crio container factory successfully Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.204422 4824 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.204445 4824 factory.go:55] Registering systemd factory Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.204458 4824 factory.go:221] Registration of the systemd container factory successfully Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.204486 4824 factory.go:103] Registering Raw factory Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.204507 4824 manager.go:1196] Started watching for new ooms in manager Oct 06 09:58:11 crc kubenswrapper[4824]: E1006 09:58:11.204885 4824 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.50:6443: connect: connection refused" interval="200ms" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.205267 4824 server.go:460] "Adding debug handlers to kubelet server" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.205418 4824 manager.go:319] Starting recovery of all containers Oct 06 09:58:11 crc kubenswrapper[4824]: E1006 09:58:11.212946 4824 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.50:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.186bde791e4ae506 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-10-06 09:58:11.198944518 +0000 UTC m=+0.563367409,LastTimestamp:2025-10-06 09:58:11.198944518 +0000 UTC m=+0.563367409,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.226710 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.226791 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.226813 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.226833 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.226853 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.226872 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.226891 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.226911 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.226936 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.226957 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.227004 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.227023 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.227073 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.227097 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.227116 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.227133 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.230250 4824 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.230319 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.230348 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.230371 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.230390 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.230411 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.230429 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.230449 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.230468 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.230491 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.230509 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.230532 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.230555 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.230577 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.230596 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.230615 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.230643 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.230661 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.230678 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.230727 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.230746 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.230763 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.230805 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.230825 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.230845 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.230866 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.230883 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.230920 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.230938 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.230956 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.231057 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.231077 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.231099 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.231121 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.231138 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.231157 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.231175 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.231199 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.231219 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.231241 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.231262 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.231283 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.231300 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.231318 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.231367 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.231388 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.231406 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.231423 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.231442 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.231475 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.231495 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.231516 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.231534 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.231553 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.231571 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.231589 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.231608 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.231628 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.231646 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.231664 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.231683 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.231700 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.231718 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.231736 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.231757 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.231777 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.231796 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.231816 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.231835 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.231855 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.231876 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.231896 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.231914 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.231934 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.231955 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.231975 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.232022 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.232041 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.232059 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.232078 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.232097 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.232116 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.232135 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.232154 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.232191 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.232209 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.232231 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.232250 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.232268 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.232295 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.232319 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.232340 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.232363 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.232383 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.232404 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.232425 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.232445 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.232469 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.232490 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.232508 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.232526 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.232546 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.232565 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.232583 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.232601 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.232623 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.232641 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.232660 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.232681 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.232703 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.232723 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.232743 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.232762 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.232779 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.232799 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.232817 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.232835 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.232854 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.232873 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.232892 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.232912 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.232932 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.232950 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.232968 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.233012 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.233031 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.233051 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.233070 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.233087 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.233105 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.233124 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.233144 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.233165 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.233183 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.233200 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.233219 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.233236 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.233257 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.233275 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.233312 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.233331 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.233349 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.233368 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.233388 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.233407 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.233424 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.233443 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.233460 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.233478 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.233497 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.233514 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.233534 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.233554 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.233575 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.233594 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.233613 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.233630 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.233647 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.233665 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.233683 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.233700 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.233721 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.233742 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.233760 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.233777 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.233795 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.233814 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.233833 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.233853 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.233871 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.233888 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.233906 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.233924 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.233971 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.234014 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.234032 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.234051 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.234069 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.234087 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.234105 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.234122 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.234140 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.234166 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.234183 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.234202 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.234221 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.234239 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.234257 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.234274 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.234291 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.234310 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.234329 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.234346 4824 reconstruct.go:97] "Volume reconstruction finished" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.234360 4824 reconciler.go:26] "Reconciler: start to sync state" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.239339 4824 manager.go:324] Recovery completed Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.256154 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.258118 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.258171 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.258183 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.263965 4824 cpu_manager.go:225] "Starting CPU manager" policy="none" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.264017 4824 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.264121 4824 state_mem.go:36] "Initialized new in-memory state store" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.270047 4824 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.272694 4824 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.272740 4824 status_manager.go:217] "Starting to sync pod status with apiserver" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.272776 4824 kubelet.go:2335] "Starting kubelet main sync loop" Oct 06 09:58:11 crc kubenswrapper[4824]: E1006 09:58:11.272861 4824 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.273522 4824 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.50:6443: connect: connection refused Oct 06 09:58:11 crc kubenswrapper[4824]: E1006 09:58:11.273581 4824 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.50:6443: connect: connection refused" logger="UnhandledError" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.288402 4824 policy_none.go:49] "None policy: Start" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.289563 4824 memory_manager.go:170] "Starting memorymanager" policy="None" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.289622 4824 state_mem.go:35] "Initializing new in-memory state store" Oct 06 09:58:11 crc kubenswrapper[4824]: E1006 09:58:11.301896 4824 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.348229 4824 manager.go:334] "Starting Device Plugin manager" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.348306 4824 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.348329 4824 server.go:79] "Starting device plugin registration server" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.348952 4824 eviction_manager.go:189] "Eviction manager: starting control loop" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.349010 4824 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.349198 4824 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.349333 4824 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.349349 4824 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Oct 06 09:58:11 crc kubenswrapper[4824]: E1006 09:58:11.362621 4824 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.373102 4824 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc"] Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.373192 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.374071 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.374101 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.374109 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.374212 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.374354 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.374408 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.374844 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.374876 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.374886 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.375048 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.375233 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.375286 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.375847 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.376040 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.376087 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.376096 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.376889 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.376927 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.377511 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.377560 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.377669 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.377712 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.377738 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.377757 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.379849 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.379887 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.379905 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.380399 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.380564 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.380588 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.380875 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.381164 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.381319 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.382725 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.382798 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.382812 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.383051 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.383079 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.383282 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.383334 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.383347 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.383932 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.383957 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.383968 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:11 crc kubenswrapper[4824]: E1006 09:58:11.406312 4824 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.50:6443: connect: connection refused" interval="400ms" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.435953 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.436041 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.436090 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.436182 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.436240 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.436276 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.436306 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.436335 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.436367 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.436396 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.436425 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.436498 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.436547 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.436585 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.436617 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.449872 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.451159 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.451223 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.451245 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.451283 4824 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 06 09:58:11 crc kubenswrapper[4824]: E1006 09:58:11.451911 4824 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.50:6443: connect: connection refused" node="crc" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.538262 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.538322 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.538353 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.538442 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.538446 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.538514 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.538508 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.538583 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.538605 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.538613 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.538613 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.538667 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.538687 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.538724 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.538751 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.538754 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.538779 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.538806 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.538826 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.538849 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.538832 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.538959 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.539031 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.539063 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.539074 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.539145 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.539152 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.539117 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.539097 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.539275 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.652043 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.653762 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.653820 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.653839 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.653873 4824 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 06 09:58:11 crc kubenswrapper[4824]: E1006 09:58:11.654434 4824 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.50:6443: connect: connection refused" node="crc" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.710393 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.716143 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.741508 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.758732 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 06 09:58:11 crc kubenswrapper[4824]: I1006 09:58:11.762967 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.763476 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-e0d2515016635e68d7a4538e4f7a066c2843c631ea202dcd3bd6f2c9569ffb10 WatchSource:0}: Error finding container e0d2515016635e68d7a4538e4f7a066c2843c631ea202dcd3bd6f2c9569ffb10: Status 404 returned error can't find the container with id e0d2515016635e68d7a4538e4f7a066c2843c631ea202dcd3bd6f2c9569ffb10 Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.765456 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-23255223e0cd844a9116f2753e87f62c3faff3abf71336f3eed46c8277b64799 WatchSource:0}: Error finding container 23255223e0cd844a9116f2753e87f62c3faff3abf71336f3eed46c8277b64799: Status 404 returned error can't find the container with id 23255223e0cd844a9116f2753e87f62c3faff3abf71336f3eed46c8277b64799 Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.791790 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-8f3b856c9fd035a6d54da34c8163eaad45db094f1951930e4691de86735a8d74 WatchSource:0}: Error finding container 8f3b856c9fd035a6d54da34c8163eaad45db094f1951930e4691de86735a8d74: Status 404 returned error can't find the container with id 8f3b856c9fd035a6d54da34c8163eaad45db094f1951930e4691de86735a8d74 Oct 06 09:58:11 crc kubenswrapper[4824]: E1006 09:58:11.807150 4824 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.50:6443: connect: connection refused" interval="800ms" Oct 06 09:58:11 crc kubenswrapper[4824]: W1006 09:58:11.998350 4824 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.50:6443: connect: connection refused Oct 06 09:58:11 crc kubenswrapper[4824]: E1006 09:58:11.998451 4824 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.50:6443: connect: connection refused" logger="UnhandledError" Oct 06 09:58:12 crc kubenswrapper[4824]: I1006 09:58:12.054968 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 06 09:58:12 crc kubenswrapper[4824]: I1006 09:58:12.057084 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:12 crc kubenswrapper[4824]: I1006 09:58:12.057133 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:12 crc kubenswrapper[4824]: I1006 09:58:12.057151 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:12 crc kubenswrapper[4824]: I1006 09:58:12.057184 4824 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 06 09:58:12 crc kubenswrapper[4824]: E1006 09:58:12.057643 4824 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.50:6443: connect: connection refused" node="crc" Oct 06 09:58:12 crc kubenswrapper[4824]: I1006 09:58:12.204474 4824 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.50:6443: connect: connection refused Oct 06 09:58:12 crc kubenswrapper[4824]: I1006 09:58:12.278459 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"8f3b856c9fd035a6d54da34c8163eaad45db094f1951930e4691de86735a8d74"} Oct 06 09:58:12 crc kubenswrapper[4824]: I1006 09:58:12.279836 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"ffd06906c9304005fcc44b691810db2a36a6918775a59be8393a22ede43bb49c"} Oct 06 09:58:12 crc kubenswrapper[4824]: I1006 09:58:12.280516 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"e0d2515016635e68d7a4538e4f7a066c2843c631ea202dcd3bd6f2c9569ffb10"} Oct 06 09:58:12 crc kubenswrapper[4824]: I1006 09:58:12.281545 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"23255223e0cd844a9116f2753e87f62c3faff3abf71336f3eed46c8277b64799"} Oct 06 09:58:12 crc kubenswrapper[4824]: I1006 09:58:12.282832 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"1f6cb59a9dc55b1ce3009676b2a3853ccc5c2c8b53193792b71b94a0dc7e31c8"} Oct 06 09:58:12 crc kubenswrapper[4824]: E1006 09:58:12.607785 4824 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.50:6443: connect: connection refused" interval="1.6s" Oct 06 09:58:12 crc kubenswrapper[4824]: W1006 09:58:12.704326 4824 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.50:6443: connect: connection refused Oct 06 09:58:12 crc kubenswrapper[4824]: E1006 09:58:12.704459 4824 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.50:6443: connect: connection refused" logger="UnhandledError" Oct 06 09:58:12 crc kubenswrapper[4824]: W1006 09:58:12.752526 4824 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.50:6443: connect: connection refused Oct 06 09:58:12 crc kubenswrapper[4824]: E1006 09:58:12.752639 4824 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.50:6443: connect: connection refused" logger="UnhandledError" Oct 06 09:58:12 crc kubenswrapper[4824]: W1006 09:58:12.795488 4824 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.50:6443: connect: connection refused Oct 06 09:58:12 crc kubenswrapper[4824]: E1006 09:58:12.795552 4824 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.50:6443: connect: connection refused" logger="UnhandledError" Oct 06 09:58:12 crc kubenswrapper[4824]: I1006 09:58:12.858728 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 06 09:58:12 crc kubenswrapper[4824]: I1006 09:58:12.860434 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:12 crc kubenswrapper[4824]: I1006 09:58:12.860508 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:12 crc kubenswrapper[4824]: I1006 09:58:12.860520 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:12 crc kubenswrapper[4824]: I1006 09:58:12.860543 4824 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 06 09:58:12 crc kubenswrapper[4824]: E1006 09:58:12.860895 4824 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.50:6443: connect: connection refused" node="crc" Oct 06 09:58:13 crc kubenswrapper[4824]: E1006 09:58:13.037625 4824 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.50:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.186bde791e4ae506 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-10-06 09:58:11.198944518 +0000 UTC m=+0.563367409,LastTimestamp:2025-10-06 09:58:11.198944518 +0000 UTC m=+0.563367409,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Oct 06 09:58:13 crc kubenswrapper[4824]: I1006 09:58:13.204076 4824 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.50:6443: connect: connection refused Oct 06 09:58:13 crc kubenswrapper[4824]: I1006 09:58:13.287233 4824 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="d6f4a5101060399a25f1d98b0495745ab0a679b88cc6bee5ea1bf0caff7b6767" exitCode=0 Oct 06 09:58:13 crc kubenswrapper[4824]: I1006 09:58:13.287283 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"d6f4a5101060399a25f1d98b0495745ab0a679b88cc6bee5ea1bf0caff7b6767"} Oct 06 09:58:13 crc kubenswrapper[4824]: I1006 09:58:13.287430 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 06 09:58:13 crc kubenswrapper[4824]: I1006 09:58:13.289099 4824 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="507540c1b777046fb6b90bb1de444f904c92fbd7b6e598217bb844b25495c683" exitCode=0 Oct 06 09:58:13 crc kubenswrapper[4824]: I1006 09:58:13.289148 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"507540c1b777046fb6b90bb1de444f904c92fbd7b6e598217bb844b25495c683"} Oct 06 09:58:13 crc kubenswrapper[4824]: I1006 09:58:13.289235 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:13 crc kubenswrapper[4824]: I1006 09:58:13.289251 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:13 crc kubenswrapper[4824]: I1006 09:58:13.289261 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:13 crc kubenswrapper[4824]: I1006 09:58:13.289475 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 06 09:58:13 crc kubenswrapper[4824]: I1006 09:58:13.291938 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:13 crc kubenswrapper[4824]: I1006 09:58:13.292041 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:13 crc kubenswrapper[4824]: I1006 09:58:13.292067 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:13 crc kubenswrapper[4824]: I1006 09:58:13.296257 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"7fedc40f0217626c56d155fc7e78e0592556fef9464db2aaf09c5facc486aad8"} Oct 06 09:58:13 crc kubenswrapper[4824]: I1006 09:58:13.296309 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"464e279c5903770c34d80f31ed231b194d5628b39055c4058908039d10690059"} Oct 06 09:58:13 crc kubenswrapper[4824]: I1006 09:58:13.296328 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"efa1b62728466ef7aac0e745230348ce920eb9ec9b07fb2cd5a276150dac13fc"} Oct 06 09:58:13 crc kubenswrapper[4824]: I1006 09:58:13.298840 4824 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="e37e8c7400f7fd14a5c5390a5985e3d8d277e9c5ca60d5e9e9757d07d1778570" exitCode=0 Oct 06 09:58:13 crc kubenswrapper[4824]: I1006 09:58:13.298953 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"e37e8c7400f7fd14a5c5390a5985e3d8d277e9c5ca60d5e9e9757d07d1778570"} Oct 06 09:58:13 crc kubenswrapper[4824]: I1006 09:58:13.299042 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 06 09:58:13 crc kubenswrapper[4824]: I1006 09:58:13.300905 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:13 crc kubenswrapper[4824]: I1006 09:58:13.300950 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:13 crc kubenswrapper[4824]: I1006 09:58:13.300966 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:13 crc kubenswrapper[4824]: I1006 09:58:13.304169 4824 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="3af57ca08bb3265d9e4cccc260a5345ee33b50665274179e0fbf969a7a52abff" exitCode=0 Oct 06 09:58:13 crc kubenswrapper[4824]: I1006 09:58:13.304198 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"3af57ca08bb3265d9e4cccc260a5345ee33b50665274179e0fbf969a7a52abff"} Oct 06 09:58:13 crc kubenswrapper[4824]: I1006 09:58:13.304263 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 06 09:58:13 crc kubenswrapper[4824]: I1006 09:58:13.304299 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 06 09:58:13 crc kubenswrapper[4824]: I1006 09:58:13.305910 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:13 crc kubenswrapper[4824]: I1006 09:58:13.305959 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:13 crc kubenswrapper[4824]: I1006 09:58:13.306004 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:13 crc kubenswrapper[4824]: I1006 09:58:13.306010 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:13 crc kubenswrapper[4824]: I1006 09:58:13.306061 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:13 crc kubenswrapper[4824]: I1006 09:58:13.306079 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:13 crc kubenswrapper[4824]: W1006 09:58:13.630024 4824 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.50:6443: connect: connection refused Oct 06 09:58:13 crc kubenswrapper[4824]: E1006 09:58:13.630119 4824 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.50:6443: connect: connection refused" logger="UnhandledError" Oct 06 09:58:14 crc kubenswrapper[4824]: I1006 09:58:14.203663 4824 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.50:6443: connect: connection refused Oct 06 09:58:14 crc kubenswrapper[4824]: E1006 09:58:14.209363 4824 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.50:6443: connect: connection refused" interval="3.2s" Oct 06 09:58:14 crc kubenswrapper[4824]: I1006 09:58:14.317627 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"268983a8eccc2af9242e8f5ebea3722e30cd66086272ec6f681ff96ed9a0e7fb"} Oct 06 09:58:14 crc kubenswrapper[4824]: I1006 09:58:14.317694 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"a0658bdad936b07337e1eb58a4d62c1d7fc98fb6b3b115d9be4c1d5d79ecbc8d"} Oct 06 09:58:14 crc kubenswrapper[4824]: I1006 09:58:14.317713 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"8b5bb30b2f9b1a35bf1dfd2274c7e3d280757902687d9ae79b710fbd5f07f398"} Oct 06 09:58:14 crc kubenswrapper[4824]: I1006 09:58:14.317845 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 06 09:58:14 crc kubenswrapper[4824]: I1006 09:58:14.319306 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:14 crc kubenswrapper[4824]: I1006 09:58:14.319349 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:14 crc kubenswrapper[4824]: I1006 09:58:14.319365 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:14 crc kubenswrapper[4824]: I1006 09:58:14.321855 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"79b92b42226cd9bba974d119a8e7f3d1d822e8a21cf1e6cdfff48dcbe7a5a898"} Oct 06 09:58:14 crc kubenswrapper[4824]: I1006 09:58:14.322037 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 06 09:58:14 crc kubenswrapper[4824]: I1006 09:58:14.323184 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:14 crc kubenswrapper[4824]: I1006 09:58:14.323222 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:14 crc kubenswrapper[4824]: I1006 09:58:14.323238 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:14 crc kubenswrapper[4824]: I1006 09:58:14.326719 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"78fac300cbe2ac52f5788306b8fc0011f6ba3a22c3546974e70998f2d394987a"} Oct 06 09:58:14 crc kubenswrapper[4824]: I1006 09:58:14.326821 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 06 09:58:14 crc kubenswrapper[4824]: I1006 09:58:14.327719 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:14 crc kubenswrapper[4824]: I1006 09:58:14.327752 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:14 crc kubenswrapper[4824]: I1006 09:58:14.327766 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:14 crc kubenswrapper[4824]: I1006 09:58:14.330694 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"413822c99ba2866521177cce4f1894aef17c160aebb2662da91824a04827d510"} Oct 06 09:58:14 crc kubenswrapper[4824]: I1006 09:58:14.330724 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"3b2a920dbc506df2426232990e31fbb60d255c00be54941826a0f094af4a92f0"} Oct 06 09:58:14 crc kubenswrapper[4824]: I1006 09:58:14.330748 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"d516f278f362d7c45e7149de82fa835997ac198f1165c04352910d51fb674d69"} Oct 06 09:58:14 crc kubenswrapper[4824]: I1006 09:58:14.330761 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"9235c742d8c45bf667f32979c2891fece50f4ee1bd0c7cf6a753ed2d2b9ab30e"} Oct 06 09:58:14 crc kubenswrapper[4824]: I1006 09:58:14.336295 4824 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="2cc8b694ed919ecda5e6e91dad516f351472982121075d4cc10017fa73aaffa1" exitCode=0 Oct 06 09:58:14 crc kubenswrapper[4824]: I1006 09:58:14.336343 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"2cc8b694ed919ecda5e6e91dad516f351472982121075d4cc10017fa73aaffa1"} Oct 06 09:58:14 crc kubenswrapper[4824]: I1006 09:58:14.336476 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 06 09:58:14 crc kubenswrapper[4824]: I1006 09:58:14.337490 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:14 crc kubenswrapper[4824]: I1006 09:58:14.337554 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:14 crc kubenswrapper[4824]: I1006 09:58:14.337570 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:14 crc kubenswrapper[4824]: I1006 09:58:14.461541 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 06 09:58:14 crc kubenswrapper[4824]: I1006 09:58:14.463588 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:14 crc kubenswrapper[4824]: I1006 09:58:14.463648 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:14 crc kubenswrapper[4824]: I1006 09:58:14.463664 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:14 crc kubenswrapper[4824]: I1006 09:58:14.463700 4824 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 06 09:58:14 crc kubenswrapper[4824]: E1006 09:58:14.464371 4824 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.50:6443: connect: connection refused" node="crc" Oct 06 09:58:14 crc kubenswrapper[4824]: I1006 09:58:14.488272 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 06 09:58:15 crc kubenswrapper[4824]: I1006 09:58:15.344476 4824 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="e99fcc109df84830c025b58f648b5d57ce7101feffa7edb7fd8e232b7fcefe58" exitCode=0 Oct 06 09:58:15 crc kubenswrapper[4824]: I1006 09:58:15.344563 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"e99fcc109df84830c025b58f648b5d57ce7101feffa7edb7fd8e232b7fcefe58"} Oct 06 09:58:15 crc kubenswrapper[4824]: I1006 09:58:15.344654 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 06 09:58:15 crc kubenswrapper[4824]: I1006 09:58:15.346077 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:15 crc kubenswrapper[4824]: I1006 09:58:15.346128 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:15 crc kubenswrapper[4824]: I1006 09:58:15.346152 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:15 crc kubenswrapper[4824]: I1006 09:58:15.352007 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"f226ce5a85d5a413bd4881ccece290f6d46b2e64fbdadbbef4ffaabd8633236c"} Oct 06 09:58:15 crc kubenswrapper[4824]: I1006 09:58:15.352030 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 06 09:58:15 crc kubenswrapper[4824]: I1006 09:58:15.352118 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 06 09:58:15 crc kubenswrapper[4824]: I1006 09:58:15.352142 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 06 09:58:15 crc kubenswrapper[4824]: I1006 09:58:15.352177 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 06 09:58:15 crc kubenswrapper[4824]: I1006 09:58:15.352129 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 06 09:58:15 crc kubenswrapper[4824]: I1006 09:58:15.354255 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:15 crc kubenswrapper[4824]: I1006 09:58:15.354322 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:15 crc kubenswrapper[4824]: I1006 09:58:15.354323 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:15 crc kubenswrapper[4824]: I1006 09:58:15.354377 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:15 crc kubenswrapper[4824]: I1006 09:58:15.354394 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:15 crc kubenswrapper[4824]: I1006 09:58:15.354348 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:15 crc kubenswrapper[4824]: I1006 09:58:15.354273 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:15 crc kubenswrapper[4824]: I1006 09:58:15.354444 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:15 crc kubenswrapper[4824]: I1006 09:58:15.354475 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:15 crc kubenswrapper[4824]: I1006 09:58:15.354495 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:15 crc kubenswrapper[4824]: I1006 09:58:15.354455 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:15 crc kubenswrapper[4824]: I1006 09:58:15.354556 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:15 crc kubenswrapper[4824]: I1006 09:58:15.461803 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 06 09:58:16 crc kubenswrapper[4824]: I1006 09:58:16.359752 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"a103bbb4ced5e0e90537c55bd5ce141275325daad0de8d953ad15b39d91cc48b"} Oct 06 09:58:16 crc kubenswrapper[4824]: I1006 09:58:16.359829 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"eeec5b1669e021c7d3364f80eae9e725deed89c4b4824c4cc3ec77bac907233f"} Oct 06 09:58:16 crc kubenswrapper[4824]: I1006 09:58:16.359833 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 06 09:58:16 crc kubenswrapper[4824]: I1006 09:58:16.359859 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 06 09:58:16 crc kubenswrapper[4824]: I1006 09:58:16.359849 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"3f3ab59fa7cb60b936b7853635a3b9b99e6a85236da890346f0d0eaaf177b2dc"} Oct 06 09:58:16 crc kubenswrapper[4824]: I1006 09:58:16.360019 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 06 09:58:16 crc kubenswrapper[4824]: I1006 09:58:16.360099 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 06 09:58:16 crc kubenswrapper[4824]: I1006 09:58:16.361667 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:16 crc kubenswrapper[4824]: I1006 09:58:16.361718 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:16 crc kubenswrapper[4824]: I1006 09:58:16.361735 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:16 crc kubenswrapper[4824]: I1006 09:58:16.362213 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:16 crc kubenswrapper[4824]: I1006 09:58:16.362237 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:16 crc kubenswrapper[4824]: I1006 09:58:16.362289 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:16 crc kubenswrapper[4824]: I1006 09:58:16.362301 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:16 crc kubenswrapper[4824]: I1006 09:58:16.362261 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:16 crc kubenswrapper[4824]: I1006 09:58:16.362384 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:17 crc kubenswrapper[4824]: I1006 09:58:17.345472 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 06 09:58:17 crc kubenswrapper[4824]: I1006 09:58:17.369382 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"8a23c075d2625ea626184f1b014e74a59bdb3cbfea2eac01dfcf338a449cf6e1"} Oct 06 09:58:17 crc kubenswrapper[4824]: I1006 09:58:17.369425 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"99531a736e62df57a553190b2b2bd7fddbcb768633825529dfc7d829cb0bcfeb"} Oct 06 09:58:17 crc kubenswrapper[4824]: I1006 09:58:17.369498 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 06 09:58:17 crc kubenswrapper[4824]: I1006 09:58:17.369547 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 06 09:58:17 crc kubenswrapper[4824]: I1006 09:58:17.370697 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:17 crc kubenswrapper[4824]: I1006 09:58:17.370737 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:17 crc kubenswrapper[4824]: I1006 09:58:17.370750 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:17 crc kubenswrapper[4824]: I1006 09:58:17.371275 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:17 crc kubenswrapper[4824]: I1006 09:58:17.371319 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:17 crc kubenswrapper[4824]: I1006 09:58:17.371332 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:17 crc kubenswrapper[4824]: I1006 09:58:17.665175 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 06 09:58:17 crc kubenswrapper[4824]: I1006 09:58:17.667208 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:17 crc kubenswrapper[4824]: I1006 09:58:17.667278 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:17 crc kubenswrapper[4824]: I1006 09:58:17.667294 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:17 crc kubenswrapper[4824]: I1006 09:58:17.667331 4824 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 06 09:58:18 crc kubenswrapper[4824]: I1006 09:58:18.372479 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 06 09:58:18 crc kubenswrapper[4824]: I1006 09:58:18.372604 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 06 09:58:18 crc kubenswrapper[4824]: I1006 09:58:18.373966 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:18 crc kubenswrapper[4824]: I1006 09:58:18.374016 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:18 crc kubenswrapper[4824]: I1006 09:58:18.374028 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:18 crc kubenswrapper[4824]: I1006 09:58:18.374089 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:18 crc kubenswrapper[4824]: I1006 09:58:18.374125 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:18 crc kubenswrapper[4824]: I1006 09:58:18.374143 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:18 crc kubenswrapper[4824]: I1006 09:58:18.806967 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 06 09:58:18 crc kubenswrapper[4824]: I1006 09:58:18.807213 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 06 09:58:18 crc kubenswrapper[4824]: I1006 09:58:18.808751 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:18 crc kubenswrapper[4824]: I1006 09:58:18.808821 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:18 crc kubenswrapper[4824]: I1006 09:58:18.808832 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:19 crc kubenswrapper[4824]: I1006 09:58:19.242909 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Oct 06 09:58:19 crc kubenswrapper[4824]: I1006 09:58:19.375465 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 06 09:58:19 crc kubenswrapper[4824]: I1006 09:58:19.376386 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:19 crc kubenswrapper[4824]: I1006 09:58:19.376413 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:19 crc kubenswrapper[4824]: I1006 09:58:19.376421 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:19 crc kubenswrapper[4824]: I1006 09:58:19.616738 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 06 09:58:19 crc kubenswrapper[4824]: I1006 09:58:19.616943 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 06 09:58:19 crc kubenswrapper[4824]: I1006 09:58:19.618464 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:19 crc kubenswrapper[4824]: I1006 09:58:19.618501 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:19 crc kubenswrapper[4824]: I1006 09:58:19.618512 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:21 crc kubenswrapper[4824]: I1006 09:58:21.136169 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 06 09:58:21 crc kubenswrapper[4824]: I1006 09:58:21.136396 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 06 09:58:21 crc kubenswrapper[4824]: I1006 09:58:21.138175 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:21 crc kubenswrapper[4824]: I1006 09:58:21.138230 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:21 crc kubenswrapper[4824]: I1006 09:58:21.138250 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:21 crc kubenswrapper[4824]: I1006 09:58:21.140742 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 06 09:58:21 crc kubenswrapper[4824]: E1006 09:58:21.362967 4824 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Oct 06 09:58:21 crc kubenswrapper[4824]: I1006 09:58:21.380632 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 06 09:58:21 crc kubenswrapper[4824]: I1006 09:58:21.381501 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:21 crc kubenswrapper[4824]: I1006 09:58:21.381569 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:21 crc kubenswrapper[4824]: I1006 09:58:21.381592 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:22 crc kubenswrapper[4824]: I1006 09:58:22.616992 4824 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 06 09:58:22 crc kubenswrapper[4824]: I1006 09:58:22.617089 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 06 09:58:25 crc kubenswrapper[4824]: W1006 09:58:25.050881 4824 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": net/http: TLS handshake timeout Oct 06 09:58:25 crc kubenswrapper[4824]: I1006 09:58:25.051549 4824 trace.go:236] Trace[1323612511]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (06-Oct-2025 09:58:15.046) (total time: 10004ms): Oct 06 09:58:25 crc kubenswrapper[4824]: Trace[1323612511]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": net/http: TLS handshake timeout 10004ms (09:58:25.050) Oct 06 09:58:25 crc kubenswrapper[4824]: Trace[1323612511]: [10.004923562s] [10.004923562s] END Oct 06 09:58:25 crc kubenswrapper[4824]: E1006 09:58:25.051584 4824 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Oct 06 09:58:25 crc kubenswrapper[4824]: I1006 09:58:25.177920 4824 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Oct 06 09:58:25 crc kubenswrapper[4824]: I1006 09:58:25.178599 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Oct 06 09:58:25 crc kubenswrapper[4824]: I1006 09:58:25.186216 4824 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Oct 06 09:58:25 crc kubenswrapper[4824]: I1006 09:58:25.186299 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Oct 06 09:58:25 crc kubenswrapper[4824]: I1006 09:58:25.394606 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Oct 06 09:58:25 crc kubenswrapper[4824]: I1006 09:58:25.396823 4824 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="f226ce5a85d5a413bd4881ccece290f6d46b2e64fbdadbbef4ffaabd8633236c" exitCode=255 Oct 06 09:58:25 crc kubenswrapper[4824]: I1006 09:58:25.396866 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"f226ce5a85d5a413bd4881ccece290f6d46b2e64fbdadbbef4ffaabd8633236c"} Oct 06 09:58:25 crc kubenswrapper[4824]: I1006 09:58:25.397041 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 06 09:58:25 crc kubenswrapper[4824]: I1006 09:58:25.398220 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:25 crc kubenswrapper[4824]: I1006 09:58:25.398293 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:25 crc kubenswrapper[4824]: I1006 09:58:25.398317 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:25 crc kubenswrapper[4824]: I1006 09:58:25.399457 4824 scope.go:117] "RemoveContainer" containerID="f226ce5a85d5a413bd4881ccece290f6d46b2e64fbdadbbef4ffaabd8633236c" Oct 06 09:58:25 crc kubenswrapper[4824]: I1006 09:58:25.470313 4824 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Oct 06 09:58:25 crc kubenswrapper[4824]: [+]log ok Oct 06 09:58:25 crc kubenswrapper[4824]: [+]etcd ok Oct 06 09:58:25 crc kubenswrapper[4824]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Oct 06 09:58:25 crc kubenswrapper[4824]: [+]poststarthook/openshift.io-api-request-count-filter ok Oct 06 09:58:25 crc kubenswrapper[4824]: [+]poststarthook/openshift.io-startkubeinformers ok Oct 06 09:58:25 crc kubenswrapper[4824]: [+]poststarthook/openshift.io-openshift-apiserver-reachable ok Oct 06 09:58:25 crc kubenswrapper[4824]: [+]poststarthook/openshift.io-oauth-apiserver-reachable ok Oct 06 09:58:25 crc kubenswrapper[4824]: [+]poststarthook/start-apiserver-admission-initializer ok Oct 06 09:58:25 crc kubenswrapper[4824]: [+]poststarthook/generic-apiserver-start-informers ok Oct 06 09:58:25 crc kubenswrapper[4824]: [+]poststarthook/priority-and-fairness-config-consumer ok Oct 06 09:58:25 crc kubenswrapper[4824]: [+]poststarthook/priority-and-fairness-filter ok Oct 06 09:58:25 crc kubenswrapper[4824]: [+]poststarthook/storage-object-count-tracker-hook ok Oct 06 09:58:25 crc kubenswrapper[4824]: [+]poststarthook/start-apiextensions-informers ok Oct 06 09:58:25 crc kubenswrapper[4824]: [+]poststarthook/start-apiextensions-controllers ok Oct 06 09:58:25 crc kubenswrapper[4824]: [+]poststarthook/crd-informer-synced ok Oct 06 09:58:25 crc kubenswrapper[4824]: [+]poststarthook/start-system-namespaces-controller ok Oct 06 09:58:25 crc kubenswrapper[4824]: [+]poststarthook/start-cluster-authentication-info-controller ok Oct 06 09:58:25 crc kubenswrapper[4824]: [+]poststarthook/start-kube-apiserver-identity-lease-controller ok Oct 06 09:58:25 crc kubenswrapper[4824]: [+]poststarthook/start-kube-apiserver-identity-lease-garbage-collector ok Oct 06 09:58:25 crc kubenswrapper[4824]: [+]poststarthook/start-legacy-token-tracking-controller ok Oct 06 09:58:25 crc kubenswrapper[4824]: [+]poststarthook/start-service-ip-repair-controllers ok Oct 06 09:58:25 crc kubenswrapper[4824]: [-]poststarthook/rbac/bootstrap-roles failed: reason withheld Oct 06 09:58:25 crc kubenswrapper[4824]: [-]poststarthook/scheduling/bootstrap-system-priority-classes failed: reason withheld Oct 06 09:58:25 crc kubenswrapper[4824]: [+]poststarthook/priority-and-fairness-config-producer ok Oct 06 09:58:25 crc kubenswrapper[4824]: [+]poststarthook/bootstrap-controller ok Oct 06 09:58:25 crc kubenswrapper[4824]: [+]poststarthook/aggregator-reload-proxy-client-cert ok Oct 06 09:58:25 crc kubenswrapper[4824]: [+]poststarthook/start-kube-aggregator-informers ok Oct 06 09:58:25 crc kubenswrapper[4824]: [+]poststarthook/apiservice-status-local-available-controller ok Oct 06 09:58:25 crc kubenswrapper[4824]: [+]poststarthook/apiservice-status-remote-available-controller ok Oct 06 09:58:25 crc kubenswrapper[4824]: [+]poststarthook/apiservice-registration-controller ok Oct 06 09:58:25 crc kubenswrapper[4824]: [+]poststarthook/apiservice-wait-for-first-sync ok Oct 06 09:58:25 crc kubenswrapper[4824]: [+]poststarthook/apiservice-discovery-controller ok Oct 06 09:58:25 crc kubenswrapper[4824]: [+]poststarthook/kube-apiserver-autoregistration ok Oct 06 09:58:25 crc kubenswrapper[4824]: [+]autoregister-completion ok Oct 06 09:58:25 crc kubenswrapper[4824]: [+]poststarthook/apiservice-openapi-controller ok Oct 06 09:58:25 crc kubenswrapper[4824]: [+]poststarthook/apiservice-openapiv3-controller ok Oct 06 09:58:25 crc kubenswrapper[4824]: livez check failed Oct 06 09:58:25 crc kubenswrapper[4824]: I1006 09:58:25.471034 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 06 09:58:26 crc kubenswrapper[4824]: I1006 09:58:26.232917 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Oct 06 09:58:26 crc kubenswrapper[4824]: I1006 09:58:26.233654 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 06 09:58:26 crc kubenswrapper[4824]: I1006 09:58:26.235174 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:26 crc kubenswrapper[4824]: I1006 09:58:26.235198 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:26 crc kubenswrapper[4824]: I1006 09:58:26.235207 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:26 crc kubenswrapper[4824]: I1006 09:58:26.281961 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Oct 06 09:58:26 crc kubenswrapper[4824]: I1006 09:58:26.404452 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Oct 06 09:58:26 crc kubenswrapper[4824]: I1006 09:58:26.407542 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"0b8cdf54407cc95783919eba8e92f093bbc09eac5e908c4ba0685629948c71bd"} Oct 06 09:58:26 crc kubenswrapper[4824]: I1006 09:58:26.407631 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 06 09:58:26 crc kubenswrapper[4824]: I1006 09:58:26.407871 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 06 09:58:26 crc kubenswrapper[4824]: I1006 09:58:26.409833 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:26 crc kubenswrapper[4824]: I1006 09:58:26.409895 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:26 crc kubenswrapper[4824]: I1006 09:58:26.409913 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:26 crc kubenswrapper[4824]: I1006 09:58:26.411799 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:26 crc kubenswrapper[4824]: I1006 09:58:26.411856 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:26 crc kubenswrapper[4824]: I1006 09:58:26.411880 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:26 crc kubenswrapper[4824]: I1006 09:58:26.428649 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Oct 06 09:58:27 crc kubenswrapper[4824]: I1006 09:58:27.412746 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Oct 06 09:58:27 crc kubenswrapper[4824]: I1006 09:58:27.413825 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Oct 06 09:58:27 crc kubenswrapper[4824]: I1006 09:58:27.416497 4824 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="0b8cdf54407cc95783919eba8e92f093bbc09eac5e908c4ba0685629948c71bd" exitCode=255 Oct 06 09:58:27 crc kubenswrapper[4824]: I1006 09:58:27.416620 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"0b8cdf54407cc95783919eba8e92f093bbc09eac5e908c4ba0685629948c71bd"} Oct 06 09:58:27 crc kubenswrapper[4824]: I1006 09:58:27.416708 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 06 09:58:27 crc kubenswrapper[4824]: I1006 09:58:27.416713 4824 scope.go:117] "RemoveContainer" containerID="f226ce5a85d5a413bd4881ccece290f6d46b2e64fbdadbbef4ffaabd8633236c" Oct 06 09:58:27 crc kubenswrapper[4824]: I1006 09:58:27.416910 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 06 09:58:27 crc kubenswrapper[4824]: I1006 09:58:27.417940 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:27 crc kubenswrapper[4824]: I1006 09:58:27.418017 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:27 crc kubenswrapper[4824]: I1006 09:58:27.418035 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:27 crc kubenswrapper[4824]: I1006 09:58:27.419162 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:27 crc kubenswrapper[4824]: I1006 09:58:27.419232 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:27 crc kubenswrapper[4824]: I1006 09:58:27.419261 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:27 crc kubenswrapper[4824]: I1006 09:58:27.420262 4824 scope.go:117] "RemoveContainer" containerID="0b8cdf54407cc95783919eba8e92f093bbc09eac5e908c4ba0685629948c71bd" Oct 06 09:58:27 crc kubenswrapper[4824]: E1006 09:58:27.420565 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Oct 06 09:58:28 crc kubenswrapper[4824]: I1006 09:58:28.422303 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Oct 06 09:58:28 crc kubenswrapper[4824]: I1006 09:58:28.812505 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 06 09:58:28 crc kubenswrapper[4824]: I1006 09:58:28.812744 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 06 09:58:28 crc kubenswrapper[4824]: I1006 09:58:28.814936 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:28 crc kubenswrapper[4824]: I1006 09:58:28.815039 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:28 crc kubenswrapper[4824]: I1006 09:58:28.815058 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:30 crc kubenswrapper[4824]: E1006 09:58:30.179505 4824 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="6.4s" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.182135 4824 trace.go:236] Trace[1671878021]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (06-Oct-2025 09:58:17.903) (total time: 12278ms): Oct 06 09:58:30 crc kubenswrapper[4824]: Trace[1671878021]: ---"Objects listed" error: 12278ms (09:58:30.181) Oct 06 09:58:30 crc kubenswrapper[4824]: Trace[1671878021]: [12.27889088s] [12.27889088s] END Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.182186 4824 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Oct 06 09:58:30 crc kubenswrapper[4824]: E1006 09:58:30.182888 4824 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.185046 4824 trace.go:236] Trace[1021732042]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (06-Oct-2025 09:58:15.424) (total time: 14760ms): Oct 06 09:58:30 crc kubenswrapper[4824]: Trace[1021732042]: ---"Objects listed" error: 14760ms (09:58:30.184) Oct 06 09:58:30 crc kubenswrapper[4824]: Trace[1021732042]: [14.760079285s] [14.760079285s] END Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.185080 4824 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.187843 4824 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.189494 4824 trace.go:236] Trace[2135492436]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (06-Oct-2025 09:58:15.710) (total time: 14478ms): Oct 06 09:58:30 crc kubenswrapper[4824]: Trace[2135492436]: ---"Objects listed" error: 14478ms (09:58:30.189) Oct 06 09:58:30 crc kubenswrapper[4824]: Trace[2135492436]: [14.478615586s] [14.478615586s] END Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.189532 4824 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.196367 4824 apiserver.go:52] "Watching apiserver" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.202514 4824 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.202849 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb"] Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.203276 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.203318 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.203437 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.203490 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 09:58:30 crc kubenswrapper[4824]: E1006 09:58:30.203644 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.203771 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.203865 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 06 09:58:30 crc kubenswrapper[4824]: E1006 09:58:30.204001 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 06 09:58:30 crc kubenswrapper[4824]: E1006 09:58:30.203774 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.205743 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.207389 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.207800 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.208041 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.208113 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.208149 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.208183 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.210935 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.211591 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.259029 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.260116 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.269496 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.275259 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.286860 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.299628 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.302761 4824 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.315593 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.327854 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.339728 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.352094 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.353834 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.365719 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.378551 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.390184 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.390244 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.390272 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.390322 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.391159 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.391459 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.391506 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.391530 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.391569 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.391592 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.391619 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.391648 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.391674 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.391696 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.391755 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.391795 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.391766 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.391880 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.391924 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.391956 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.392011 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.392016 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.392042 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.392070 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.392098 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.392126 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.392155 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.392164 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.392184 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.392215 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.392216 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.392240 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.392302 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.392334 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.392365 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.392406 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.392442 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.392470 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.392601 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.392607 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.392630 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.392659 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.392665 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.392696 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.392751 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.392781 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.392810 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.392839 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.392864 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.392867 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.392893 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.392923 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.392948 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.393008 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.393044 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.393047 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.393041 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.393100 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.393127 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.393152 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.393181 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.393204 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.393231 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.393234 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.393241 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.393254 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.393284 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.393340 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.393366 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.393391 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.393416 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.393425 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.393455 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.393483 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.393511 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.393538 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.393567 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.393597 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.393679 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.393707 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.393736 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.393767 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.393792 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.393823 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.393854 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.393884 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.393912 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.393939 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.393966 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.394022 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.394051 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.394081 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.394107 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.394134 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.394158 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.394185 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.394214 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.394239 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.394266 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.394294 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.394335 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.394362 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.394388 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.393638 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.393717 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.393814 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.396523 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.393827 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.393854 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.393956 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.394049 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.393946 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.394103 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.394156 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.394229 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.394256 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.394304 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.394415 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: E1006 09:58:30.394424 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-06 09:58:30.894390477 +0000 UTC m=+20.258813348 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.394468 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.394502 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.394581 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.394813 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.395058 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.395138 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.396076 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.396141 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.396314 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.396437 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.396677 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.397022 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.397080 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.397264 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.397287 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.397365 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.397497 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.397603 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.397604 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.397807 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.397821 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.398319 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.398349 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.398377 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.398615 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.398735 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.398793 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.398871 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.399013 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.399050 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.399411 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.399557 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.399671 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.399766 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.399887 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.400397 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.400505 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.400672 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.400710 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.400738 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.400725 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.400768 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.400792 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.400816 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.400841 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.400865 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.400888 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.400914 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.401053 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.401086 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.401112 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.401138 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.401168 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.401194 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.401221 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.401244 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.401268 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.401294 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.401320 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.401344 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.401365 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.401388 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.401415 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.401437 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.401472 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.401497 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.401521 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.401545 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.401567 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.401608 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.401630 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.401651 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.401676 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.401706 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.401732 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.401755 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.401779 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.401801 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.401825 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.401850 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.401876 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.401901 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.401923 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.401942 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.401961 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.402021 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.402044 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.402064 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.402088 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.402113 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.402138 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.402161 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.402186 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.402207 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.402227 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.402252 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.402274 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.402303 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.402325 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.402355 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.402376 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.402400 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.402427 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.402454 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.402484 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.402520 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.402542 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.402567 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.402588 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.402609 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.402632 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.402654 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.402677 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.402696 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.402716 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.402840 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.402871 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.402900 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.402924 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.402956 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.403009 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.403037 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.403059 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.403084 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.403112 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.403133 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.403156 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.403181 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.403202 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.403235 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.403261 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.403284 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.403306 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.403328 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.403352 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.403377 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.403407 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.403434 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.403470 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.403493 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.403516 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.403539 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.403566 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.403592 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.403618 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.403675 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.403705 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.403735 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.403758 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.403784 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.403806 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.403833 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.403861 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.403875 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.403883 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.403899 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.403997 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.404066 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.404339 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.404432 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.404492 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.404819 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.404864 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.405061 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.405353 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.405360 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.405647 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.405935 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.406165 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.406254 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.406419 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.406726 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.406691 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.407000 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.406967 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.407165 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.407439 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.407461 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.407958 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.407277 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.407723 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.408408 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.409786 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.411168 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.411374 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.411389 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.411499 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.411509 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.411520 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.411969 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.412091 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.412385 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.412415 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.412772 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.412971 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.413385 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.413465 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.413489 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.413473 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.413547 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.413620 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.413721 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.413911 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.413950 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.414011 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.414029 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.414512 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.414730 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.414745 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.414913 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.415087 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.415151 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.415158 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.415219 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.415370 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.415518 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.415542 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.415554 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.415870 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.415890 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.416044 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.416258 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.416589 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: E1006 09:58:30.416659 4824 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 06 09:58:30 crc kubenswrapper[4824]: E1006 09:58:30.418021 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-06 09:58:30.917974612 +0000 UTC m=+20.282397473 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.416896 4824 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.418294 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.417453 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.418392 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.418439 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.418488 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.418540 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.418700 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.418726 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.418749 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.418769 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.418789 4824 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.418810 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.418832 4824 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.418851 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.418871 4824 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.418890 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.418909 4824 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.418932 4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.418953 4824 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.418972 4824 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.419026 4824 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.419045 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.419065 4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.419086 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.419855 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.420090 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.420198 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.420280 4824 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.420317 4824 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.420341 4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.420361 4824 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.420384 4824 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.420405 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.420424 4824 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.420445 4824 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.420457 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.420479 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.420688 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.420707 4824 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.420722 4824 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.420714 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.420737 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.420754 4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: E1006 09:58:30.416737 4824 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.420769 4824 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.420785 4824 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.420800 4824 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.420814 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.420828 4824 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.420841 4824 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.420854 4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.420873 4824 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.420887 4824 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.420892 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.420901 4824 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.420908 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.420941 4824 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.420969 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.421014 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.421039 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.421022 4824 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.421103 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.421122 4824 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.417119 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.417344 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: E1006 09:58:30.421158 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-06 09:58:30.921127469 +0000 UTC m=+20.285550530 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.421213 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.421249 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.421270 4824 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.421292 4824 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.421312 4824 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.421329 4824 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.421347 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.421347 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.421366 4824 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.421392 4824 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.421412 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.421433 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.421450 4824 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.421468 4824 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.421487 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.421505 4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.421525 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.421543 4824 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.421559 4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.421575 4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.421692 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.421714 4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.421776 4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.421797 4824 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.421693 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.421817 4824 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.422379 4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.422402 4824 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.422466 4824 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.422488 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.422542 4824 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.422592 4824 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.422790 4824 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.422811 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.422864 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.422886 4824 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.422904 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.422952 4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.422973 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.423036 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.423058 4824 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.423116 4824 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.423802 4824 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.423824 4824 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.423839 4824 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.423858 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.423874 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.423888 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.423902 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.423915 4824 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.423930 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.423944 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.423959 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.423973 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.424012 4824 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.424024 4824 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.424037 4824 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.424051 4824 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.424064 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.424077 4824 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.422410 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.424092 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.424105 4824 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.424119 4824 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.424154 4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.424166 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.422688 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.424182 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.424195 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.424374 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.424396 4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.424413 4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.424431 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.424445 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.424458 4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.424471 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.424484 4824 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.424496 4824 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.424509 4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.424522 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.424536 4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.424550 4824 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.424564 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.424577 4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.424591 4824 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.424604 4824 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.424617 4824 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.424632 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.424645 4824 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.424658 4824 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.424670 4824 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.424682 4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.424722 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.424737 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.425905 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.426763 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.427115 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.428142 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.429646 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.430290 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.431188 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 06 09:58:30 crc kubenswrapper[4824]: E1006 09:58:30.434207 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 06 09:58:30 crc kubenswrapper[4824]: E1006 09:58:30.434237 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 06 09:58:30 crc kubenswrapper[4824]: E1006 09:58:30.434252 4824 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 06 09:58:30 crc kubenswrapper[4824]: E1006 09:58:30.434325 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-06 09:58:30.934303771 +0000 UTC m=+20.298726632 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.441417 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.441622 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.442119 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.442843 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: E1006 09:58:30.444690 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 06 09:58:30 crc kubenswrapper[4824]: E1006 09:58:30.444777 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 06 09:58:30 crc kubenswrapper[4824]: E1006 09:58:30.444825 4824 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 06 09:58:30 crc kubenswrapper[4824]: E1006 09:58:30.444945 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-06 09:58:30.944914779 +0000 UTC m=+20.309337660 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.447048 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.448594 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.448875 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.449460 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.449709 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.449711 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.450065 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.450318 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.450939 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.451115 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.451382 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.451868 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.451965 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.452850 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.453384 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.453488 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.453790 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.453853 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.454612 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.457254 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.457297 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.457493 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.457844 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.458113 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: E1006 09:58:30.458151 4824 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"kube-controller-manager-crc\" already exists" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.458786 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.459094 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.459219 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.459591 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.460336 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.464475 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.465042 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.465328 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.466666 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.466896 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.478284 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.488138 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.488491 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.490201 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.493495 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.500131 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.503196 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.505855 4824 scope.go:117] "RemoveContainer" containerID="0b8cdf54407cc95783919eba8e92f093bbc09eac5e908c4ba0685629948c71bd" Oct 06 09:58:30 crc kubenswrapper[4824]: E1006 09:58:30.506104 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.506450 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.521778 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.525648 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.525797 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.526000 4824 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.526014 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.526070 4824 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.526236 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.526252 4824 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.526267 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.526781 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.526799 4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.526810 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.526821 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.526846 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.526894 4824 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.526990 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.526718 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.527009 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.527023 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.527039 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.527061 4824 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.527072 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.526572 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.527084 4824 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.527109 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.527124 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.527134 4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.527173 4824 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.527183 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.527193 4824 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.527201 4824 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.527209 4824 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.527218 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.527225 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.527251 4824 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.527329 4824 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.527344 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.527361 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.527376 4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.527388 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.527399 4824 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.527411 4824 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.527420 4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.527430 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.527441 4824 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.527451 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.527461 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.527471 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.527480 4824 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.527490 4824 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.527500 4824 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.527509 4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.527519 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.527529 4824 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.527539 4824 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.527549 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.527558 4824 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.527568 4824 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.527577 4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.527587 4824 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.527597 4824 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.527605 4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.531955 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.539523 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.540901 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.546897 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.551256 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 06 09:58:30 crc kubenswrapper[4824]: E1006 09:58:30.552561 4824 kuberuntime_manager.go:1274] "Unhandled Error" err=< Oct 06 09:58:30 crc kubenswrapper[4824]: container &Container{Name:network-operator,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b,Command:[/bin/bash -c #!/bin/bash Oct 06 09:58:30 crc kubenswrapper[4824]: set -o allexport Oct 06 09:58:30 crc kubenswrapper[4824]: if [[ -f /etc/kubernetes/apiserver-url.env ]]; then Oct 06 09:58:30 crc kubenswrapper[4824]: source /etc/kubernetes/apiserver-url.env Oct 06 09:58:30 crc kubenswrapper[4824]: else Oct 06 09:58:30 crc kubenswrapper[4824]: echo "Error: /etc/kubernetes/apiserver-url.env is missing" Oct 06 09:58:30 crc kubenswrapper[4824]: exit 1 Oct 06 09:58:30 crc kubenswrapper[4824]: fi Oct 06 09:58:30 crc kubenswrapper[4824]: exec /usr/bin/cluster-network-operator start --listen=0.0.0.0:9104 Oct 06 09:58:30 crc kubenswrapper[4824]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:cno,HostPort:9104,ContainerPort:9104,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:RELEASE_VERSION,Value:4.18.1,ValueFrom:nil,},EnvVar{Name:KUBE_PROXY_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b97554198294bf544fbc116c94a0a1fb2ec8a4de0e926bf9d9e320135f0bee6f,ValueFrom:nil,},EnvVar{Name:KUBE_RBAC_PROXY_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09,ValueFrom:nil,},EnvVar{Name:MULTUS_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26,ValueFrom:nil,},EnvVar{Name:MULTUS_ADMISSION_CONTROLLER_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317,ValueFrom:nil,},EnvVar{Name:CNI_PLUGINS_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc,ValueFrom:nil,},EnvVar{Name:BOND_CNI_PLUGIN_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78,ValueFrom:nil,},EnvVar{Name:WHEREABOUTS_CNI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4,ValueFrom:nil,},EnvVar{Name:ROUTE_OVERRRIDE_CNI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa,ValueFrom:nil,},EnvVar{Name:MULTUS_NETWORKPOLICY_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:23f833d3738d68706eb2f2868bd76bd71cee016cffa6faf5f045a60cc8c6eddd,ValueFrom:nil,},EnvVar{Name:OVN_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2,ValueFrom:nil,},EnvVar{Name:OVN_NB_RAFT_ELECTION_TIMER,Value:10,ValueFrom:nil,},EnvVar{Name:OVN_SB_RAFT_ELECTION_TIMER,Value:16,ValueFrom:nil,},EnvVar{Name:OVN_NORTHD_PROBE_INTERVAL,Value:10000,ValueFrom:nil,},EnvVar{Name:OVN_CONTROLLER_INACTIVITY_PROBE,Value:180000,ValueFrom:nil,},EnvVar{Name:OVN_NB_INACTIVITY_PROBE,Value:60000,ValueFrom:nil,},EnvVar{Name:EGRESS_ROUTER_CNI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c,ValueFrom:nil,},EnvVar{Name:NETWORK_METRICS_DAEMON_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d,ValueFrom:nil,},EnvVar{Name:NETWORK_CHECK_SOURCE_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b,ValueFrom:nil,},EnvVar{Name:NETWORK_CHECK_TARGET_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b,ValueFrom:nil,},EnvVar{Name:NETWORK_OPERATOR_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b,ValueFrom:nil,},EnvVar{Name:CLOUD_NETWORK_CONFIG_CONTROLLER_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8048f1cb0be521f09749c0a489503cd56d85b68c6ca93380e082cfd693cd97a8,ValueFrom:nil,},EnvVar{Name:CLI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2,ValueFrom:nil,},EnvVar{Name:FRR_K8S_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5dbf844e49bb46b78586930149e5e5f5dc121014c8afd10fe36f3651967cc256,ValueFrom:nil,},EnvVar{Name:NETWORKING_CONSOLE_PLUGIN_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd,ValueFrom:nil,},EnvVar{Name:POD_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.name,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:host-etc-kube,ReadOnly:true,MountPath:/etc/kubernetes,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:metrics-tls,ReadOnly:false,MountPath:/var/run/secrets/serving-cert,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rdwmf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod network-operator-58b4c7f79c-55gtf_openshift-network-operator(37a5e44f-9a88-4405-be8a-b645485e7312): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Oct 06 09:58:30 crc kubenswrapper[4824]: > logger="UnhandledError" Oct 06 09:58:30 crc kubenswrapper[4824]: E1006 09:58:30.553774 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"network-operator\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" podUID="37a5e44f-9a88-4405-be8a-b645485e7312" Oct 06 09:58:30 crc kubenswrapper[4824]: W1006 09:58:30.554919 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef543e1b_8068_4ea3_b32a_61027b32e95d.slice/crio-773785d1d0ea8666b53a5eb0e854f9c83b73d5be1a421323fc35cdbefb62fbc6 WatchSource:0}: Error finding container 773785d1d0ea8666b53a5eb0e854f9c83b73d5be1a421323fc35cdbefb62fbc6: Status 404 returned error can't find the container with id 773785d1d0ea8666b53a5eb0e854f9c83b73d5be1a421323fc35cdbefb62fbc6 Oct 06 09:58:30 crc kubenswrapper[4824]: E1006 09:58:30.557440 4824 kuberuntime_manager.go:1274] "Unhandled Error" err=< Oct 06 09:58:30 crc kubenswrapper[4824]: container &Container{Name:webhook,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2,Command:[/bin/bash -c set -xe Oct 06 09:58:30 crc kubenswrapper[4824]: if [[ -f "/env/_master" ]]; then Oct 06 09:58:30 crc kubenswrapper[4824]: set -o allexport Oct 06 09:58:30 crc kubenswrapper[4824]: source "/env/_master" Oct 06 09:58:30 crc kubenswrapper[4824]: set +o allexport Oct 06 09:58:30 crc kubenswrapper[4824]: fi Oct 06 09:58:30 crc kubenswrapper[4824]: # OVN-K will try to remove hybrid overlay node annotations even when the hybrid overlay is not enabled. Oct 06 09:58:30 crc kubenswrapper[4824]: # https://github.com/ovn-org/ovn-kubernetes/blob/ac6820df0b338a246f10f412cd5ec903bd234694/go-controller/pkg/ovn/master.go#L791 Oct 06 09:58:30 crc kubenswrapper[4824]: ho_enable="--enable-hybrid-overlay" Oct 06 09:58:30 crc kubenswrapper[4824]: echo "I$(date "+%m%d %H:%M:%S.%N") - network-node-identity - start webhook" Oct 06 09:58:30 crc kubenswrapper[4824]: # extra-allowed-user: service account `ovn-kubernetes-control-plane` Oct 06 09:58:30 crc kubenswrapper[4824]: # sets pod annotations in multi-homing layer3 network controller (cluster-manager) Oct 06 09:58:30 crc kubenswrapper[4824]: exec /usr/bin/ovnkube-identity --k8s-apiserver=https://api-int.crc.testing:6443 \ Oct 06 09:58:30 crc kubenswrapper[4824]: --webhook-cert-dir="/etc/webhook-cert" \ Oct 06 09:58:30 crc kubenswrapper[4824]: --webhook-host=127.0.0.1 \ Oct 06 09:58:30 crc kubenswrapper[4824]: --webhook-port=9743 \ Oct 06 09:58:30 crc kubenswrapper[4824]: ${ho_enable} \ Oct 06 09:58:30 crc kubenswrapper[4824]: --enable-interconnect \ Oct 06 09:58:30 crc kubenswrapper[4824]: --disable-approver \ Oct 06 09:58:30 crc kubenswrapper[4824]: --extra-allowed-user="system:serviceaccount:openshift-ovn-kubernetes:ovn-kubernetes-control-plane" \ Oct 06 09:58:30 crc kubenswrapper[4824]: --wait-for-kubernetes-api=200s \ Oct 06 09:58:30 crc kubenswrapper[4824]: --pod-admission-conditions="/var/run/ovnkube-identity-config/additional-pod-admission-cond.json" \ Oct 06 09:58:30 crc kubenswrapper[4824]: --loglevel="${LOGLEVEL}" Oct 06 09:58:30 crc kubenswrapper[4824]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOGLEVEL,Value:2,ValueFrom:nil,},EnvVar{Name:KUBERNETES_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:webhook-cert,ReadOnly:false,MountPath:/etc/webhook-cert/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:env-overrides,ReadOnly:false,MountPath:/env,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovnkube-identity-cm,ReadOnly:false,MountPath:/var/run/ovnkube-identity-config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-s2kz5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000470000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod network-node-identity-vrzqb_openshift-network-node-identity(ef543e1b-8068-4ea3-b32a-61027b32e95d): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Oct 06 09:58:30 crc kubenswrapper[4824]: > logger="UnhandledError" Oct 06 09:58:30 crc kubenswrapper[4824]: E1006 09:58:30.560461 4824 kuberuntime_manager.go:1274] "Unhandled Error" err=< Oct 06 09:58:30 crc kubenswrapper[4824]: container &Container{Name:approver,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2,Command:[/bin/bash -c set -xe Oct 06 09:58:30 crc kubenswrapper[4824]: if [[ -f "/env/_master" ]]; then Oct 06 09:58:30 crc kubenswrapper[4824]: set -o allexport Oct 06 09:58:30 crc kubenswrapper[4824]: source "/env/_master" Oct 06 09:58:30 crc kubenswrapper[4824]: set +o allexport Oct 06 09:58:30 crc kubenswrapper[4824]: fi Oct 06 09:58:30 crc kubenswrapper[4824]: Oct 06 09:58:30 crc kubenswrapper[4824]: echo "I$(date "+%m%d %H:%M:%S.%N") - network-node-identity - start approver" Oct 06 09:58:30 crc kubenswrapper[4824]: exec /usr/bin/ovnkube-identity --k8s-apiserver=https://api-int.crc.testing:6443 \ Oct 06 09:58:30 crc kubenswrapper[4824]: --disable-webhook \ Oct 06 09:58:30 crc kubenswrapper[4824]: --csr-acceptance-conditions="/var/run/ovnkube-identity-config/additional-cert-acceptance-cond.json" \ Oct 06 09:58:30 crc kubenswrapper[4824]: --loglevel="${LOGLEVEL}" Oct 06 09:58:30 crc kubenswrapper[4824]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOGLEVEL,Value:4,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:env-overrides,ReadOnly:false,MountPath:/env,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovnkube-identity-cm,ReadOnly:false,MountPath:/var/run/ovnkube-identity-config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-s2kz5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000470000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod network-node-identity-vrzqb_openshift-network-node-identity(ef543e1b-8068-4ea3-b32a-61027b32e95d): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Oct 06 09:58:30 crc kubenswrapper[4824]: > logger="UnhandledError" Oct 06 09:58:30 crc kubenswrapper[4824]: E1006 09:58:30.560851 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:iptables-alerter,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2,Command:[/iptables-alerter/iptables-alerter.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONTAINER_RUNTIME_ENDPOINT,Value:unix:///run/crio/crio.sock,ValueFrom:nil,},EnvVar{Name:ALERTER_POD_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.name,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{68157440 0} {} 65Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:iptables-alerter-script,ReadOnly:false,MountPath:/iptables-alerter,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host-slash,ReadOnly:true,MountPath:/host,SubPath:,MountPropagation:*HostToContainer,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rczfb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:*true,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod iptables-alerter-4ln5h_openshift-network-operator(d75a4c96-2883-4a0b-bab2-0fab2b6c0b49): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars" logger="UnhandledError" Oct 06 09:58:30 crc kubenswrapper[4824]: E1006 09:58:30.561963 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"iptables-alerter\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-network-operator/iptables-alerter-4ln5h" podUID="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" Oct 06 09:58:30 crc kubenswrapper[4824]: E1006 09:58:30.562037 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"webhook\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\", failed to \"StartContainer\" for \"approver\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"]" pod="openshift-network-node-identity/network-node-identity-vrzqb" podUID="ef543e1b-8068-4ea3-b32a-61027b32e95d" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.562919 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.576694 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.589707 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.591343 4824 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.601280 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d739433-0a38-42fe-b884-b8433011967e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://464e279c5903770c34d80f31ed231b194d5628b39055c4058908039d10690059\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://efa1b62728466ef7aac0e745230348ce920eb9ec9b07fb2cd5a276150dac13fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fedc40f0217626c56d155fc7e78e0592556fef9464db2aaf09c5facc486aad8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78fac300cbe2ac52f5788306b8fc0011f6ba3a22c3546974e70998f2d394987a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.612508 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b665cda3-da4e-482e-8248-023abec382f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9235c742d8c45bf667f32979c2891fece50f4ee1bd0c7cf6a753ed2d2b9ab30e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b2a920dbc506df2426232990e31fbb60d255c00be54941826a0f094af4a92f0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d516f278f362d7c45e7149de82fa835997ac198f1165c04352910d51fb674d69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b8cdf54407cc95783919eba8e92f093bbc09eac5e908c4ba0685629948c71bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b8cdf54407cc95783919eba8e92f093bbc09eac5e908c4ba0685629948c71bd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-06T09:58:27Z\\\",\\\"message\\\":\\\"arting RequestHeaderAuthRequestController\\\\nI1006 09:58:27.156944 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1006 09:58:27.156783 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1006 09:58:27.157213 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-432254888/tls.crt::/tmp/serving-cert-432254888/tls.key\\\\\\\"\\\\nI1006 09:58:27.157407 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1006 09:58:27.157443 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1006 09:58:27.157410 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-432254888/tls.crt::/tmp/serving-cert-432254888/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759744705\\\\\\\\\\\\\\\" (2025-10-06 09:58:25 +0000 UTC to 2025-11-05 09:58:26 +0000 UTC (now=2025-10-06 09:58:27.157390076 +0000 UTC))\\\\\\\"\\\\nI1006 09:58:27.157632 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1006 09:58:27.158158 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759744707\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759744707\\\\\\\\\\\\\\\" (2025-10-06 08:58:26 +0000 UTC to 2026-10-06 08:58:26 +0000 UTC (now=2025-10-06 09:58:27.158090663 +0000 UTC))\\\\\\\"\\\\nI1006 09:58:27.158220 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1006 09:58:27.158270 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1006 09:58:27.158310 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nF1006 09:58:27.158617 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://413822c99ba2866521177cce4f1894aef17c160aebb2662da91824a04827d510\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e37e8c7400f7fd14a5c5390a5985e3d8d277e9c5ca60d5e9e9757d07d1778570\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e37e8c7400f7fd14a5c5390a5985e3d8d277e9c5ca60d5e9e9757d07d1778570\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.624903 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d739433-0a38-42fe-b884-b8433011967e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://464e279c5903770c34d80f31ed231b194d5628b39055c4058908039d10690059\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://efa1b62728466ef7aac0e745230348ce920eb9ec9b07fb2cd5a276150dac13fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fedc40f0217626c56d155fc7e78e0592556fef9464db2aaf09c5facc486aad8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78fac300cbe2ac52f5788306b8fc0011f6ba3a22c3546974e70998f2d394987a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.627911 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.627935 4824 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.636762 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.649316 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.659721 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.671274 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.684868 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.697482 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.931302 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.931410 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 09:58:30 crc kubenswrapper[4824]: I1006 09:58:30.931440 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 09:58:30 crc kubenswrapper[4824]: E1006 09:58:30.931526 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-06 09:58:31.93147853 +0000 UTC m=+21.295901431 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 09:58:30 crc kubenswrapper[4824]: E1006 09:58:30.931576 4824 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 06 09:58:30 crc kubenswrapper[4824]: E1006 09:58:30.931649 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-06 09:58:31.931632314 +0000 UTC m=+21.296055255 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 06 09:58:30 crc kubenswrapper[4824]: E1006 09:58:30.931713 4824 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 06 09:58:30 crc kubenswrapper[4824]: E1006 09:58:30.931937 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-06 09:58:31.93189993 +0000 UTC m=+21.296322821 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 06 09:58:31 crc kubenswrapper[4824]: I1006 09:58:31.032293 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 09:58:31 crc kubenswrapper[4824]: I1006 09:58:31.032396 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 09:58:31 crc kubenswrapper[4824]: E1006 09:58:31.032563 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 06 09:58:31 crc kubenswrapper[4824]: E1006 09:58:31.032589 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 06 09:58:31 crc kubenswrapper[4824]: E1006 09:58:31.032610 4824 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 06 09:58:31 crc kubenswrapper[4824]: E1006 09:58:31.032607 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 06 09:58:31 crc kubenswrapper[4824]: E1006 09:58:31.032663 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 06 09:58:31 crc kubenswrapper[4824]: E1006 09:58:31.032688 4824 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 06 09:58:31 crc kubenswrapper[4824]: E1006 09:58:31.032697 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-06 09:58:32.032673349 +0000 UTC m=+21.397096240 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 06 09:58:31 crc kubenswrapper[4824]: E1006 09:58:31.032828 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-06 09:58:32.032794901 +0000 UTC m=+21.397217792 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 06 09:58:31 crc kubenswrapper[4824]: I1006 09:58:31.278203 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Oct 06 09:58:31 crc kubenswrapper[4824]: I1006 09:58:31.279327 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Oct 06 09:58:31 crc kubenswrapper[4824]: I1006 09:58:31.281785 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Oct 06 09:58:31 crc kubenswrapper[4824]: I1006 09:58:31.283076 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Oct 06 09:58:31 crc kubenswrapper[4824]: I1006 09:58:31.285322 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Oct 06 09:58:31 crc kubenswrapper[4824]: I1006 09:58:31.286528 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Oct 06 09:58:31 crc kubenswrapper[4824]: I1006 09:58:31.287726 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Oct 06 09:58:31 crc kubenswrapper[4824]: I1006 09:58:31.289776 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Oct 06 09:58:31 crc kubenswrapper[4824]: I1006 09:58:31.290759 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Oct 06 09:58:31 crc kubenswrapper[4824]: I1006 09:58:31.291760 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Oct 06 09:58:31 crc kubenswrapper[4824]: I1006 09:58:31.292560 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Oct 06 09:58:31 crc kubenswrapper[4824]: I1006 09:58:31.292616 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b665cda3-da4e-482e-8248-023abec382f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9235c742d8c45bf667f32979c2891fece50f4ee1bd0c7cf6a753ed2d2b9ab30e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b2a920dbc506df2426232990e31fbb60d255c00be54941826a0f094af4a92f0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d516f278f362d7c45e7149de82fa835997ac198f1165c04352910d51fb674d69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b8cdf54407cc95783919eba8e92f093bbc09eac5e908c4ba0685629948c71bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b8cdf54407cc95783919eba8e92f093bbc09eac5e908c4ba0685629948c71bd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-06T09:58:27Z\\\",\\\"message\\\":\\\"arting RequestHeaderAuthRequestController\\\\nI1006 09:58:27.156944 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1006 09:58:27.156783 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1006 09:58:27.157213 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-432254888/tls.crt::/tmp/serving-cert-432254888/tls.key\\\\\\\"\\\\nI1006 09:58:27.157407 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1006 09:58:27.157443 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1006 09:58:27.157410 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-432254888/tls.crt::/tmp/serving-cert-432254888/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759744705\\\\\\\\\\\\\\\" (2025-10-06 09:58:25 +0000 UTC to 2025-11-05 09:58:26 +0000 UTC (now=2025-10-06 09:58:27.157390076 +0000 UTC))\\\\\\\"\\\\nI1006 09:58:27.157632 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1006 09:58:27.158158 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759744707\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759744707\\\\\\\\\\\\\\\" (2025-10-06 08:58:26 +0000 UTC to 2026-10-06 08:58:26 +0000 UTC (now=2025-10-06 09:58:27.158090663 +0000 UTC))\\\\\\\"\\\\nI1006 09:58:27.158220 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1006 09:58:27.158270 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1006 09:58:27.158310 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nF1006 09:58:27.158617 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://413822c99ba2866521177cce4f1894aef17c160aebb2662da91824a04827d510\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e37e8c7400f7fd14a5c5390a5985e3d8d277e9c5ca60d5e9e9757d07d1778570\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e37e8c7400f7fd14a5c5390a5985e3d8d277e9c5ca60d5e9e9757d07d1778570\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 06 09:58:31 crc kubenswrapper[4824]: I1006 09:58:31.293664 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Oct 06 09:58:31 crc kubenswrapper[4824]: I1006 09:58:31.294831 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Oct 06 09:58:31 crc kubenswrapper[4824]: I1006 09:58:31.295620 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Oct 06 09:58:31 crc kubenswrapper[4824]: I1006 09:58:31.296442 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Oct 06 09:58:31 crc kubenswrapper[4824]: I1006 09:58:31.297231 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Oct 06 09:58:31 crc kubenswrapper[4824]: I1006 09:58:31.299468 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Oct 06 09:58:31 crc kubenswrapper[4824]: I1006 09:58:31.300424 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Oct 06 09:58:31 crc kubenswrapper[4824]: I1006 09:58:31.301174 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Oct 06 09:58:31 crc kubenswrapper[4824]: I1006 09:58:31.301844 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Oct 06 09:58:31 crc kubenswrapper[4824]: I1006 09:58:31.303027 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Oct 06 09:58:31 crc kubenswrapper[4824]: I1006 09:58:31.303655 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Oct 06 09:58:31 crc kubenswrapper[4824]: I1006 09:58:31.304615 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Oct 06 09:58:31 crc kubenswrapper[4824]: I1006 09:58:31.305384 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Oct 06 09:58:31 crc kubenswrapper[4824]: I1006 09:58:31.306717 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Oct 06 09:58:31 crc kubenswrapper[4824]: I1006 09:58:31.307410 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Oct 06 09:58:31 crc kubenswrapper[4824]: I1006 09:58:31.308524 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Oct 06 09:58:31 crc kubenswrapper[4824]: I1006 09:58:31.309086 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Oct 06 09:58:31 crc kubenswrapper[4824]: I1006 09:58:31.309685 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Oct 06 09:58:31 crc kubenswrapper[4824]: I1006 09:58:31.310583 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Oct 06 09:58:31 crc kubenswrapper[4824]: I1006 09:58:31.311089 4824 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Oct 06 09:58:31 crc kubenswrapper[4824]: I1006 09:58:31.311198 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Oct 06 09:58:31 crc kubenswrapper[4824]: I1006 09:58:31.313330 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Oct 06 09:58:31 crc kubenswrapper[4824]: I1006 09:58:31.313828 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Oct 06 09:58:31 crc kubenswrapper[4824]: I1006 09:58:31.314402 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Oct 06 09:58:31 crc kubenswrapper[4824]: I1006 09:58:31.316410 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Oct 06 09:58:31 crc kubenswrapper[4824]: I1006 09:58:31.316494 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d739433-0a38-42fe-b884-b8433011967e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://464e279c5903770c34d80f31ed231b194d5628b39055c4058908039d10690059\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://efa1b62728466ef7aac0e745230348ce920eb9ec9b07fb2cd5a276150dac13fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fedc40f0217626c56d155fc7e78e0592556fef9464db2aaf09c5facc486aad8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78fac300cbe2ac52f5788306b8fc0011f6ba3a22c3546974e70998f2d394987a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 06 09:58:31 crc kubenswrapper[4824]: I1006 09:58:31.317147 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Oct 06 09:58:31 crc kubenswrapper[4824]: I1006 09:58:31.318129 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Oct 06 09:58:31 crc kubenswrapper[4824]: I1006 09:58:31.318896 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Oct 06 09:58:31 crc kubenswrapper[4824]: I1006 09:58:31.320191 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Oct 06 09:58:31 crc kubenswrapper[4824]: I1006 09:58:31.320823 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Oct 06 09:58:31 crc kubenswrapper[4824]: I1006 09:58:31.321608 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Oct 06 09:58:31 crc kubenswrapper[4824]: I1006 09:58:31.322804 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Oct 06 09:58:31 crc kubenswrapper[4824]: I1006 09:58:31.324150 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Oct 06 09:58:31 crc kubenswrapper[4824]: I1006 09:58:31.324722 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Oct 06 09:58:31 crc kubenswrapper[4824]: I1006 09:58:31.325727 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Oct 06 09:58:31 crc kubenswrapper[4824]: I1006 09:58:31.326284 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Oct 06 09:58:31 crc kubenswrapper[4824]: I1006 09:58:31.327452 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Oct 06 09:58:31 crc kubenswrapper[4824]: I1006 09:58:31.327997 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Oct 06 09:58:31 crc kubenswrapper[4824]: I1006 09:58:31.328838 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Oct 06 09:58:31 crc kubenswrapper[4824]: I1006 09:58:31.329337 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Oct 06 09:58:31 crc kubenswrapper[4824]: I1006 09:58:31.329877 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Oct 06 09:58:31 crc kubenswrapper[4824]: I1006 09:58:31.330022 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 06 09:58:31 crc kubenswrapper[4824]: I1006 09:58:31.330820 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Oct 06 09:58:31 crc kubenswrapper[4824]: I1006 09:58:31.331360 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Oct 06 09:58:31 crc kubenswrapper[4824]: I1006 09:58:31.342338 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 06 09:58:31 crc kubenswrapper[4824]: I1006 09:58:31.355123 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 06 09:58:31 crc kubenswrapper[4824]: I1006 09:58:31.371033 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 06 09:58:31 crc kubenswrapper[4824]: I1006 09:58:31.394156 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 06 09:58:31 crc kubenswrapper[4824]: I1006 09:58:31.411767 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 06 09:58:31 crc kubenswrapper[4824]: I1006 09:58:31.449765 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"773785d1d0ea8666b53a5eb0e854f9c83b73d5be1a421323fc35cdbefb62fbc6"} Oct 06 09:58:31 crc kubenswrapper[4824]: I1006 09:58:31.452589 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"b1cff2d5844d65f5073a258156f1e2de78e5f44b80c1cbb66e9b1cf326e184dd"} Oct 06 09:58:31 crc kubenswrapper[4824]: I1006 09:58:31.454628 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"deca304bad592750c5e145ed984e74fde5a8ba37284dc8a8d2d78c0f37990976"} Oct 06 09:58:31 crc kubenswrapper[4824]: E1006 09:58:31.469230 4824 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"kube-apiserver-crc\" already exists" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 06 09:58:31 crc kubenswrapper[4824]: I1006 09:58:31.469572 4824 scope.go:117] "RemoveContainer" containerID="0b8cdf54407cc95783919eba8e92f093bbc09eac5e908c4ba0685629948c71bd" Oct 06 09:58:31 crc kubenswrapper[4824]: E1006 09:58:31.469805 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Oct 06 09:58:31 crc kubenswrapper[4824]: I1006 09:58:31.483420 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b665cda3-da4e-482e-8248-023abec382f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9235c742d8c45bf667f32979c2891fece50f4ee1bd0c7cf6a753ed2d2b9ab30e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b2a920dbc506df2426232990e31fbb60d255c00be54941826a0f094af4a92f0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d516f278f362d7c45e7149de82fa835997ac198f1165c04352910d51fb674d69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b8cdf54407cc95783919eba8e92f093bbc09eac5e908c4ba0685629948c71bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b8cdf54407cc95783919eba8e92f093bbc09eac5e908c4ba0685629948c71bd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-06T09:58:27Z\\\",\\\"message\\\":\\\"arting RequestHeaderAuthRequestController\\\\nI1006 09:58:27.156944 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1006 09:58:27.156783 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1006 09:58:27.157213 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-432254888/tls.crt::/tmp/serving-cert-432254888/tls.key\\\\\\\"\\\\nI1006 09:58:27.157407 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1006 09:58:27.157443 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1006 09:58:27.157410 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-432254888/tls.crt::/tmp/serving-cert-432254888/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759744705\\\\\\\\\\\\\\\" (2025-10-06 09:58:25 +0000 UTC to 2025-11-05 09:58:26 +0000 UTC (now=2025-10-06 09:58:27.157390076 +0000 UTC))\\\\\\\"\\\\nI1006 09:58:27.157632 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1006 09:58:27.158158 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759744707\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759744707\\\\\\\\\\\\\\\" (2025-10-06 08:58:26 +0000 UTC to 2026-10-06 08:58:26 +0000 UTC (now=2025-10-06 09:58:27.158090663 +0000 UTC))\\\\\\\"\\\\nI1006 09:58:27.158220 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1006 09:58:27.158270 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1006 09:58:27.158310 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nF1006 09:58:27.158617 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://413822c99ba2866521177cce4f1894aef17c160aebb2662da91824a04827d510\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e37e8c7400f7fd14a5c5390a5985e3d8d277e9c5ca60d5e9e9757d07d1778570\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e37e8c7400f7fd14a5c5390a5985e3d8d277e9c5ca60d5e9e9757d07d1778570\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 06 09:58:31 crc kubenswrapper[4824]: I1006 09:58:31.497834 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d739433-0a38-42fe-b884-b8433011967e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://464e279c5903770c34d80f31ed231b194d5628b39055c4058908039d10690059\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://efa1b62728466ef7aac0e745230348ce920eb9ec9b07fb2cd5a276150dac13fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fedc40f0217626c56d155fc7e78e0592556fef9464db2aaf09c5facc486aad8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78fac300cbe2ac52f5788306b8fc0011f6ba3a22c3546974e70998f2d394987a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 06 09:58:31 crc kubenswrapper[4824]: I1006 09:58:31.513276 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 06 09:58:31 crc kubenswrapper[4824]: I1006 09:58:31.528323 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 06 09:58:31 crc kubenswrapper[4824]: I1006 09:58:31.555276 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 06 09:58:31 crc kubenswrapper[4824]: I1006 09:58:31.573578 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 06 09:58:31 crc kubenswrapper[4824]: I1006 09:58:31.588596 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 06 09:58:31 crc kubenswrapper[4824]: I1006 09:58:31.602027 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 06 09:58:31 crc kubenswrapper[4824]: I1006 09:58:31.615286 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 06 09:58:31 crc kubenswrapper[4824]: I1006 09:58:31.629181 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 06 09:58:31 crc kubenswrapper[4824]: I1006 09:58:31.638941 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 06 09:58:31 crc kubenswrapper[4824]: I1006 09:58:31.656557 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 06 09:58:31 crc kubenswrapper[4824]: I1006 09:58:31.667320 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d739433-0a38-42fe-b884-b8433011967e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://464e279c5903770c34d80f31ed231b194d5628b39055c4058908039d10690059\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://efa1b62728466ef7aac0e745230348ce920eb9ec9b07fb2cd5a276150dac13fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fedc40f0217626c56d155fc7e78e0592556fef9464db2aaf09c5facc486aad8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78fac300cbe2ac52f5788306b8fc0011f6ba3a22c3546974e70998f2d394987a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 06 09:58:31 crc kubenswrapper[4824]: I1006 09:58:31.678052 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 06 09:58:31 crc kubenswrapper[4824]: I1006 09:58:31.695677 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 06 09:58:31 crc kubenswrapper[4824]: I1006 09:58:31.710400 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b665cda3-da4e-482e-8248-023abec382f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9235c742d8c45bf667f32979c2891fece50f4ee1bd0c7cf6a753ed2d2b9ab30e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b2a920dbc506df2426232990e31fbb60d255c00be54941826a0f094af4a92f0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d516f278f362d7c45e7149de82fa835997ac198f1165c04352910d51fb674d69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b8cdf54407cc95783919eba8e92f093bbc09eac5e908c4ba0685629948c71bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b8cdf54407cc95783919eba8e92f093bbc09eac5e908c4ba0685629948c71bd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-06T09:58:27Z\\\",\\\"message\\\":\\\"arting RequestHeaderAuthRequestController\\\\nI1006 09:58:27.156944 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1006 09:58:27.156783 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1006 09:58:27.157213 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-432254888/tls.crt::/tmp/serving-cert-432254888/tls.key\\\\\\\"\\\\nI1006 09:58:27.157407 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1006 09:58:27.157443 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1006 09:58:27.157410 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-432254888/tls.crt::/tmp/serving-cert-432254888/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759744705\\\\\\\\\\\\\\\" (2025-10-06 09:58:25 +0000 UTC to 2025-11-05 09:58:26 +0000 UTC (now=2025-10-06 09:58:27.157390076 +0000 UTC))\\\\\\\"\\\\nI1006 09:58:27.157632 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1006 09:58:27.158158 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759744707\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759744707\\\\\\\\\\\\\\\" (2025-10-06 08:58:26 +0000 UTC to 2026-10-06 08:58:26 +0000 UTC (now=2025-10-06 09:58:27.158090663 +0000 UTC))\\\\\\\"\\\\nI1006 09:58:27.158220 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1006 09:58:27.158270 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1006 09:58:27.158310 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nF1006 09:58:27.158617 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://413822c99ba2866521177cce4f1894aef17c160aebb2662da91824a04827d510\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e37e8c7400f7fd14a5c5390a5985e3d8d277e9c5ca60d5e9e9757d07d1778570\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e37e8c7400f7fd14a5c5390a5985e3d8d277e9c5ca60d5e9e9757d07d1778570\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 06 09:58:31 crc kubenswrapper[4824]: I1006 09:58:31.802483 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-xcgwc"] Oct 06 09:58:31 crc kubenswrapper[4824]: I1006 09:58:31.802805 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-xcgwc" Oct 06 09:58:31 crc kubenswrapper[4824]: I1006 09:58:31.805293 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Oct 06 09:58:31 crc kubenswrapper[4824]: I1006 09:58:31.805441 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Oct 06 09:58:31 crc kubenswrapper[4824]: I1006 09:58:31.805452 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Oct 06 09:58:31 crc kubenswrapper[4824]: I1006 09:58:31.817711 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xcgwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74317f75-9ebc-4327-b88c-598c2241240f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v6sg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xcgwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 06 09:58:31 crc kubenswrapper[4824]: I1006 09:58:31.830607 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b665cda3-da4e-482e-8248-023abec382f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9235c742d8c45bf667f32979c2891fece50f4ee1bd0c7cf6a753ed2d2b9ab30e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b2a920dbc506df2426232990e31fbb60d255c00be54941826a0f094af4a92f0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d516f278f362d7c45e7149de82fa835997ac198f1165c04352910d51fb674d69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b8cdf54407cc95783919eba8e92f093bbc09eac5e908c4ba0685629948c71bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b8cdf54407cc95783919eba8e92f093bbc09eac5e908c4ba0685629948c71bd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-06T09:58:27Z\\\",\\\"message\\\":\\\"arting RequestHeaderAuthRequestController\\\\nI1006 09:58:27.156944 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1006 09:58:27.156783 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1006 09:58:27.157213 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-432254888/tls.crt::/tmp/serving-cert-432254888/tls.key\\\\\\\"\\\\nI1006 09:58:27.157407 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1006 09:58:27.157443 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1006 09:58:27.157410 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-432254888/tls.crt::/tmp/serving-cert-432254888/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759744705\\\\\\\\\\\\\\\" (2025-10-06 09:58:25 +0000 UTC to 2025-11-05 09:58:26 +0000 UTC (now=2025-10-06 09:58:27.157390076 +0000 UTC))\\\\\\\"\\\\nI1006 09:58:27.157632 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1006 09:58:27.158158 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759744707\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759744707\\\\\\\\\\\\\\\" (2025-10-06 08:58:26 +0000 UTC to 2026-10-06 08:58:26 +0000 UTC (now=2025-10-06 09:58:27.158090663 +0000 UTC))\\\\\\\"\\\\nI1006 09:58:27.158220 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1006 09:58:27.158270 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1006 09:58:27.158310 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nF1006 09:58:27.158617 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://413822c99ba2866521177cce4f1894aef17c160aebb2662da91824a04827d510\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e37e8c7400f7fd14a5c5390a5985e3d8d277e9c5ca60d5e9e9757d07d1778570\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e37e8c7400f7fd14a5c5390a5985e3d8d277e9c5ca60d5e9e9757d07d1778570\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 06 09:58:31 crc kubenswrapper[4824]: I1006 09:58:31.839123 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/74317f75-9ebc-4327-b88c-598c2241240f-hosts-file\") pod \"node-resolver-xcgwc\" (UID: \"74317f75-9ebc-4327-b88c-598c2241240f\") " pod="openshift-dns/node-resolver-xcgwc" Oct 06 09:58:31 crc kubenswrapper[4824]: I1006 09:58:31.839159 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7v6sg\" (UniqueName: \"kubernetes.io/projected/74317f75-9ebc-4327-b88c-598c2241240f-kube-api-access-7v6sg\") pod \"node-resolver-xcgwc\" (UID: \"74317f75-9ebc-4327-b88c-598c2241240f\") " pod="openshift-dns/node-resolver-xcgwc" Oct 06 09:58:31 crc kubenswrapper[4824]: I1006 09:58:31.840360 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d739433-0a38-42fe-b884-b8433011967e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://464e279c5903770c34d80f31ed231b194d5628b39055c4058908039d10690059\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://efa1b62728466ef7aac0e745230348ce920eb9ec9b07fb2cd5a276150dac13fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fedc40f0217626c56d155fc7e78e0592556fef9464db2aaf09c5facc486aad8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78fac300cbe2ac52f5788306b8fc0011f6ba3a22c3546974e70998f2d394987a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 06 09:58:31 crc kubenswrapper[4824]: I1006 09:58:31.850201 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 06 09:58:31 crc kubenswrapper[4824]: I1006 09:58:31.860463 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 06 09:58:31 crc kubenswrapper[4824]: I1006 09:58:31.870111 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 06 09:58:31 crc kubenswrapper[4824]: I1006 09:58:31.884180 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 06 09:58:31 crc kubenswrapper[4824]: I1006 09:58:31.918116 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 06 09:58:31 crc kubenswrapper[4824]: I1006 09:58:31.939640 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 09:58:31 crc kubenswrapper[4824]: I1006 09:58:31.939714 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/74317f75-9ebc-4327-b88c-598c2241240f-hosts-file\") pod \"node-resolver-xcgwc\" (UID: \"74317f75-9ebc-4327-b88c-598c2241240f\") " pod="openshift-dns/node-resolver-xcgwc" Oct 06 09:58:31 crc kubenswrapper[4824]: I1006 09:58:31.939742 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7v6sg\" (UniqueName: \"kubernetes.io/projected/74317f75-9ebc-4327-b88c-598c2241240f-kube-api-access-7v6sg\") pod \"node-resolver-xcgwc\" (UID: \"74317f75-9ebc-4327-b88c-598c2241240f\") " pod="openshift-dns/node-resolver-xcgwc" Oct 06 09:58:31 crc kubenswrapper[4824]: I1006 09:58:31.939772 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 09:58:31 crc kubenswrapper[4824]: I1006 09:58:31.939789 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 09:58:31 crc kubenswrapper[4824]: E1006 09:58:31.939911 4824 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 06 09:58:31 crc kubenswrapper[4824]: E1006 09:58:31.939935 4824 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 06 09:58:31 crc kubenswrapper[4824]: E1006 09:58:31.940013 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-06 09:58:33.939924732 +0000 UTC m=+23.304347593 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 09:58:31 crc kubenswrapper[4824]: E1006 09:58:31.940080 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-06 09:58:33.940051045 +0000 UTC m=+23.304474126 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 06 09:58:31 crc kubenswrapper[4824]: E1006 09:58:31.940101 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-06 09:58:33.940090686 +0000 UTC m=+23.304513757 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 06 09:58:31 crc kubenswrapper[4824]: I1006 09:58:31.940367 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/74317f75-9ebc-4327-b88c-598c2241240f-hosts-file\") pod \"node-resolver-xcgwc\" (UID: \"74317f75-9ebc-4327-b88c-598c2241240f\") " pod="openshift-dns/node-resolver-xcgwc" Oct 06 09:58:31 crc kubenswrapper[4824]: I1006 09:58:31.947032 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 06 09:58:31 crc kubenswrapper[4824]: I1006 09:58:31.961215 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7v6sg\" (UniqueName: \"kubernetes.io/projected/74317f75-9ebc-4327-b88c-598c2241240f-kube-api-access-7v6sg\") pod \"node-resolver-xcgwc\" (UID: \"74317f75-9ebc-4327-b88c-598c2241240f\") " pod="openshift-dns/node-resolver-xcgwc" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.040521 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.040611 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 09:58:32 crc kubenswrapper[4824]: E1006 09:58:32.040756 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 06 09:58:32 crc kubenswrapper[4824]: E1006 09:58:32.040797 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 06 09:58:32 crc kubenswrapper[4824]: E1006 09:58:32.040809 4824 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 06 09:58:32 crc kubenswrapper[4824]: E1006 09:58:32.040865 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 06 09:58:32 crc kubenswrapper[4824]: E1006 09:58:32.040919 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 06 09:58:32 crc kubenswrapper[4824]: E1006 09:58:32.040935 4824 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 06 09:58:32 crc kubenswrapper[4824]: E1006 09:58:32.040887 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-06 09:58:34.040865155 +0000 UTC m=+23.405288016 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 06 09:58:32 crc kubenswrapper[4824]: E1006 09:58:32.041076 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-06 09:58:34.041029279 +0000 UTC m=+23.405452150 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.116335 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-xcgwc" Oct 06 09:58:32 crc kubenswrapper[4824]: W1006 09:58:32.127889 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod74317f75_9ebc_4327_b88c_598c2241240f.slice/crio-bf05486bc610482fa7ea1ccb39a39b55a38904af509ea2b474fa6159fd35b9f7 WatchSource:0}: Error finding container bf05486bc610482fa7ea1ccb39a39b55a38904af509ea2b474fa6159fd35b9f7: Status 404 returned error can't find the container with id bf05486bc610482fa7ea1ccb39a39b55a38904af509ea2b474fa6159fd35b9f7 Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.197913 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-khgzw"] Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.198318 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-szn8q"] Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.198550 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-7c22q"] Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.198609 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.198565 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-szn8q" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.199441 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-7c22q" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.203080 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.203429 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.203800 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.204018 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.204539 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.204753 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.204783 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.204822 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.204943 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.205047 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.205557 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.209455 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.225176 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xcgwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74317f75-9ebc-4327-b88c-598c2241240f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v6sg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xcgwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:32Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.239398 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1140c8c-93f1-4fce-8c06-a69561ff8a6d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gq49w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gq49w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-khgzw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:32Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.241944 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/b65abf9f-4e88-4571-960d-3ca997d9c344-system-cni-dir\") pod \"multus-szn8q\" (UID: \"b65abf9f-4e88-4571-960d-3ca997d9c344\") " pod="openshift-multus/multus-szn8q" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.242018 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/b65abf9f-4e88-4571-960d-3ca997d9c344-host-var-lib-cni-multus\") pod \"multus-szn8q\" (UID: \"b65abf9f-4e88-4571-960d-3ca997d9c344\") " pod="openshift-multus/multus-szn8q" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.242046 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/b65abf9f-4e88-4571-960d-3ca997d9c344-host-var-lib-kubelet\") pod \"multus-szn8q\" (UID: \"b65abf9f-4e88-4571-960d-3ca997d9c344\") " pod="openshift-multus/multus-szn8q" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.242090 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z7pmj\" (UniqueName: \"kubernetes.io/projected/b7eaaa68-d971-4a1c-b6c5-551ded8ff27f-kube-api-access-z7pmj\") pod \"multus-additional-cni-plugins-7c22q\" (UID: \"b7eaaa68-d971-4a1c-b6c5-551ded8ff27f\") " pod="openshift-multus/multus-additional-cni-plugins-7c22q" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.242125 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b65abf9f-4e88-4571-960d-3ca997d9c344-etc-kubernetes\") pod \"multus-szn8q\" (UID: \"b65abf9f-4e88-4571-960d-3ca997d9c344\") " pod="openshift-multus/multus-szn8q" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.242149 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/b7eaaa68-d971-4a1c-b6c5-551ded8ff27f-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-7c22q\" (UID: \"b7eaaa68-d971-4a1c-b6c5-551ded8ff27f\") " pod="openshift-multus/multus-additional-cni-plugins-7c22q" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.242174 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/b7eaaa68-d971-4a1c-b6c5-551ded8ff27f-cni-binary-copy\") pod \"multus-additional-cni-plugins-7c22q\" (UID: \"b7eaaa68-d971-4a1c-b6c5-551ded8ff27f\") " pod="openshift-multus/multus-additional-cni-plugins-7c22q" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.242251 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/b7eaaa68-d971-4a1c-b6c5-551ded8ff27f-os-release\") pod \"multus-additional-cni-plugins-7c22q\" (UID: \"b7eaaa68-d971-4a1c-b6c5-551ded8ff27f\") " pod="openshift-multus/multus-additional-cni-plugins-7c22q" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.242329 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/b65abf9f-4e88-4571-960d-3ca997d9c344-cnibin\") pod \"multus-szn8q\" (UID: \"b65abf9f-4e88-4571-960d-3ca997d9c344\") " pod="openshift-multus/multus-szn8q" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.242361 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/b65abf9f-4e88-4571-960d-3ca997d9c344-multus-socket-dir-parent\") pod \"multus-szn8q\" (UID: \"b65abf9f-4e88-4571-960d-3ca997d9c344\") " pod="openshift-multus/multus-szn8q" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.242381 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/b65abf9f-4e88-4571-960d-3ca997d9c344-host-var-lib-cni-bin\") pod \"multus-szn8q\" (UID: \"b65abf9f-4e88-4571-960d-3ca997d9c344\") " pod="openshift-multus/multus-szn8q" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.242398 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/b65abf9f-4e88-4571-960d-3ca997d9c344-hostroot\") pod \"multus-szn8q\" (UID: \"b65abf9f-4e88-4571-960d-3ca997d9c344\") " pod="openshift-multus/multus-szn8q" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.242447 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/b65abf9f-4e88-4571-960d-3ca997d9c344-multus-conf-dir\") pod \"multus-szn8q\" (UID: \"b65abf9f-4e88-4571-960d-3ca997d9c344\") " pod="openshift-multus/multus-szn8q" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.242486 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-49s5w\" (UniqueName: \"kubernetes.io/projected/b65abf9f-4e88-4571-960d-3ca997d9c344-kube-api-access-49s5w\") pod \"multus-szn8q\" (UID: \"b65abf9f-4e88-4571-960d-3ca997d9c344\") " pod="openshift-multus/multus-szn8q" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.242523 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/b65abf9f-4e88-4571-960d-3ca997d9c344-multus-cni-dir\") pod \"multus-szn8q\" (UID: \"b65abf9f-4e88-4571-960d-3ca997d9c344\") " pod="openshift-multus/multus-szn8q" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.242544 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gq49w\" (UniqueName: \"kubernetes.io/projected/e1140c8c-93f1-4fce-8c06-a69561ff8a6d-kube-api-access-gq49w\") pod \"machine-config-daemon-khgzw\" (UID: \"e1140c8c-93f1-4fce-8c06-a69561ff8a6d\") " pod="openshift-machine-config-operator/machine-config-daemon-khgzw" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.242563 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/e1140c8c-93f1-4fce-8c06-a69561ff8a6d-proxy-tls\") pod \"machine-config-daemon-khgzw\" (UID: \"e1140c8c-93f1-4fce-8c06-a69561ff8a6d\") " pod="openshift-machine-config-operator/machine-config-daemon-khgzw" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.242580 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e1140c8c-93f1-4fce-8c06-a69561ff8a6d-mcd-auth-proxy-config\") pod \"machine-config-daemon-khgzw\" (UID: \"e1140c8c-93f1-4fce-8c06-a69561ff8a6d\") " pod="openshift-machine-config-operator/machine-config-daemon-khgzw" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.242603 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/b65abf9f-4e88-4571-960d-3ca997d9c344-host-run-netns\") pod \"multus-szn8q\" (UID: \"b65abf9f-4e88-4571-960d-3ca997d9c344\") " pod="openshift-multus/multus-szn8q" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.242618 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/b7eaaa68-d971-4a1c-b6c5-551ded8ff27f-tuning-conf-dir\") pod \"multus-additional-cni-plugins-7c22q\" (UID: \"b7eaaa68-d971-4a1c-b6c5-551ded8ff27f\") " pod="openshift-multus/multus-additional-cni-plugins-7c22q" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.242635 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/e1140c8c-93f1-4fce-8c06-a69561ff8a6d-rootfs\") pod \"machine-config-daemon-khgzw\" (UID: \"e1140c8c-93f1-4fce-8c06-a69561ff8a6d\") " pod="openshift-machine-config-operator/machine-config-daemon-khgzw" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.242652 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/b65abf9f-4e88-4571-960d-3ca997d9c344-cni-binary-copy\") pod \"multus-szn8q\" (UID: \"b65abf9f-4e88-4571-960d-3ca997d9c344\") " pod="openshift-multus/multus-szn8q" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.242669 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/b65abf9f-4e88-4571-960d-3ca997d9c344-multus-daemon-config\") pod \"multus-szn8q\" (UID: \"b65abf9f-4e88-4571-960d-3ca997d9c344\") " pod="openshift-multus/multus-szn8q" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.242686 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/b65abf9f-4e88-4571-960d-3ca997d9c344-os-release\") pod \"multus-szn8q\" (UID: \"b65abf9f-4e88-4571-960d-3ca997d9c344\") " pod="openshift-multus/multus-szn8q" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.242701 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/b65abf9f-4e88-4571-960d-3ca997d9c344-host-run-multus-certs\") pod \"multus-szn8q\" (UID: \"b65abf9f-4e88-4571-960d-3ca997d9c344\") " pod="openshift-multus/multus-szn8q" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.242750 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/b65abf9f-4e88-4571-960d-3ca997d9c344-host-run-k8s-cni-cncf-io\") pod \"multus-szn8q\" (UID: \"b65abf9f-4e88-4571-960d-3ca997d9c344\") " pod="openshift-multus/multus-szn8q" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.242777 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/b7eaaa68-d971-4a1c-b6c5-551ded8ff27f-cnibin\") pod \"multus-additional-cni-plugins-7c22q\" (UID: \"b7eaaa68-d971-4a1c-b6c5-551ded8ff27f\") " pod="openshift-multus/multus-additional-cni-plugins-7c22q" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.242800 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/b7eaaa68-d971-4a1c-b6c5-551ded8ff27f-system-cni-dir\") pod \"multus-additional-cni-plugins-7c22q\" (UID: \"b7eaaa68-d971-4a1c-b6c5-551ded8ff27f\") " pod="openshift-multus/multus-additional-cni-plugins-7c22q" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.254150 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:32Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.267105 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:32Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.273307 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.273354 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.273425 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 09:58:32 crc kubenswrapper[4824]: E1006 09:58:32.273457 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 06 09:58:32 crc kubenswrapper[4824]: E1006 09:58:32.273620 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 06 09:58:32 crc kubenswrapper[4824]: E1006 09:58:32.273732 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.283872 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:32Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.300477 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:32Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.315451 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:32Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.333314 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d739433-0a38-42fe-b884-b8433011967e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://464e279c5903770c34d80f31ed231b194d5628b39055c4058908039d10690059\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://efa1b62728466ef7aac0e745230348ce920eb9ec9b07fb2cd5a276150dac13fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fedc40f0217626c56d155fc7e78e0592556fef9464db2aaf09c5facc486aad8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78fac300cbe2ac52f5788306b8fc0011f6ba3a22c3546974e70998f2d394987a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:32Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.343906 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/b65abf9f-4e88-4571-960d-3ca997d9c344-hostroot\") pod \"multus-szn8q\" (UID: \"b65abf9f-4e88-4571-960d-3ca997d9c344\") " pod="openshift-multus/multus-szn8q" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.344084 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/b65abf9f-4e88-4571-960d-3ca997d9c344-hostroot\") pod \"multus-szn8q\" (UID: \"b65abf9f-4e88-4571-960d-3ca997d9c344\") " pod="openshift-multus/multus-szn8q" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.344101 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/b65abf9f-4e88-4571-960d-3ca997d9c344-multus-conf-dir\") pod \"multus-szn8q\" (UID: \"b65abf9f-4e88-4571-960d-3ca997d9c344\") " pod="openshift-multus/multus-szn8q" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.344219 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-49s5w\" (UniqueName: \"kubernetes.io/projected/b65abf9f-4e88-4571-960d-3ca997d9c344-kube-api-access-49s5w\") pod \"multus-szn8q\" (UID: \"b65abf9f-4e88-4571-960d-3ca997d9c344\") " pod="openshift-multus/multus-szn8q" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.344272 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/b65abf9f-4e88-4571-960d-3ca997d9c344-multus-cni-dir\") pod \"multus-szn8q\" (UID: \"b65abf9f-4e88-4571-960d-3ca997d9c344\") " pod="openshift-multus/multus-szn8q" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.344301 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/b65abf9f-4e88-4571-960d-3ca997d9c344-multus-socket-dir-parent\") pod \"multus-szn8q\" (UID: \"b65abf9f-4e88-4571-960d-3ca997d9c344\") " pod="openshift-multus/multus-szn8q" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.344338 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/b65abf9f-4e88-4571-960d-3ca997d9c344-host-var-lib-cni-bin\") pod \"multus-szn8q\" (UID: \"b65abf9f-4e88-4571-960d-3ca997d9c344\") " pod="openshift-multus/multus-szn8q" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.344366 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gq49w\" (UniqueName: \"kubernetes.io/projected/e1140c8c-93f1-4fce-8c06-a69561ff8a6d-kube-api-access-gq49w\") pod \"machine-config-daemon-khgzw\" (UID: \"e1140c8c-93f1-4fce-8c06-a69561ff8a6d\") " pod="openshift-machine-config-operator/machine-config-daemon-khgzw" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.344412 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/b65abf9f-4e88-4571-960d-3ca997d9c344-host-run-netns\") pod \"multus-szn8q\" (UID: \"b65abf9f-4e88-4571-960d-3ca997d9c344\") " pod="openshift-multus/multus-szn8q" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.344439 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/b7eaaa68-d971-4a1c-b6c5-551ded8ff27f-tuning-conf-dir\") pod \"multus-additional-cni-plugins-7c22q\" (UID: \"b7eaaa68-d971-4a1c-b6c5-551ded8ff27f\") " pod="openshift-multus/multus-additional-cni-plugins-7c22q" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.344468 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/e1140c8c-93f1-4fce-8c06-a69561ff8a6d-proxy-tls\") pod \"machine-config-daemon-khgzw\" (UID: \"e1140c8c-93f1-4fce-8c06-a69561ff8a6d\") " pod="openshift-machine-config-operator/machine-config-daemon-khgzw" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.344494 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e1140c8c-93f1-4fce-8c06-a69561ff8a6d-mcd-auth-proxy-config\") pod \"machine-config-daemon-khgzw\" (UID: \"e1140c8c-93f1-4fce-8c06-a69561ff8a6d\") " pod="openshift-machine-config-operator/machine-config-daemon-khgzw" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.344526 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/b65abf9f-4e88-4571-960d-3ca997d9c344-cni-binary-copy\") pod \"multus-szn8q\" (UID: \"b65abf9f-4e88-4571-960d-3ca997d9c344\") " pod="openshift-multus/multus-szn8q" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.344554 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/b65abf9f-4e88-4571-960d-3ca997d9c344-multus-daemon-config\") pod \"multus-szn8q\" (UID: \"b65abf9f-4e88-4571-960d-3ca997d9c344\") " pod="openshift-multus/multus-szn8q" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.344580 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/e1140c8c-93f1-4fce-8c06-a69561ff8a6d-rootfs\") pod \"machine-config-daemon-khgzw\" (UID: \"e1140c8c-93f1-4fce-8c06-a69561ff8a6d\") " pod="openshift-machine-config-operator/machine-config-daemon-khgzw" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.344616 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/b65abf9f-4e88-4571-960d-3ca997d9c344-os-release\") pod \"multus-szn8q\" (UID: \"b65abf9f-4e88-4571-960d-3ca997d9c344\") " pod="openshift-multus/multus-szn8q" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.344645 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/b65abf9f-4e88-4571-960d-3ca997d9c344-host-run-multus-certs\") pod \"multus-szn8q\" (UID: \"b65abf9f-4e88-4571-960d-3ca997d9c344\") " pod="openshift-multus/multus-szn8q" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.344679 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/b65abf9f-4e88-4571-960d-3ca997d9c344-host-run-k8s-cni-cncf-io\") pod \"multus-szn8q\" (UID: \"b65abf9f-4e88-4571-960d-3ca997d9c344\") " pod="openshift-multus/multus-szn8q" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.344713 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/b65abf9f-4e88-4571-960d-3ca997d9c344-host-var-lib-cni-bin\") pod \"multus-szn8q\" (UID: \"b65abf9f-4e88-4571-960d-3ca997d9c344\") " pod="openshift-multus/multus-szn8q" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.344757 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/b7eaaa68-d971-4a1c-b6c5-551ded8ff27f-system-cni-dir\") pod \"multus-additional-cni-plugins-7c22q\" (UID: \"b7eaaa68-d971-4a1c-b6c5-551ded8ff27f\") " pod="openshift-multus/multus-additional-cni-plugins-7c22q" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.344799 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/b65abf9f-4e88-4571-960d-3ca997d9c344-host-run-netns\") pod \"multus-szn8q\" (UID: \"b65abf9f-4e88-4571-960d-3ca997d9c344\") " pod="openshift-multus/multus-szn8q" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.344873 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/e1140c8c-93f1-4fce-8c06-a69561ff8a6d-rootfs\") pod \"machine-config-daemon-khgzw\" (UID: \"e1140c8c-93f1-4fce-8c06-a69561ff8a6d\") " pod="openshift-machine-config-operator/machine-config-daemon-khgzw" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.344722 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/b7eaaa68-d971-4a1c-b6c5-551ded8ff27f-system-cni-dir\") pod \"multus-additional-cni-plugins-7c22q\" (UID: \"b7eaaa68-d971-4a1c-b6c5-551ded8ff27f\") " pod="openshift-multus/multus-additional-cni-plugins-7c22q" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.345187 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/b65abf9f-4e88-4571-960d-3ca997d9c344-multus-socket-dir-parent\") pod \"multus-szn8q\" (UID: \"b65abf9f-4e88-4571-960d-3ca997d9c344\") " pod="openshift-multus/multus-szn8q" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.345304 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/b65abf9f-4e88-4571-960d-3ca997d9c344-host-run-k8s-cni-cncf-io\") pod \"multus-szn8q\" (UID: \"b65abf9f-4e88-4571-960d-3ca997d9c344\") " pod="openshift-multus/multus-szn8q" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.345321 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/b7eaaa68-d971-4a1c-b6c5-551ded8ff27f-cnibin\") pod \"multus-additional-cni-plugins-7c22q\" (UID: \"b7eaaa68-d971-4a1c-b6c5-551ded8ff27f\") " pod="openshift-multus/multus-additional-cni-plugins-7c22q" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.345263 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/b7eaaa68-d971-4a1c-b6c5-551ded8ff27f-tuning-conf-dir\") pod \"multus-additional-cni-plugins-7c22q\" (UID: \"b7eaaa68-d971-4a1c-b6c5-551ded8ff27f\") " pod="openshift-multus/multus-additional-cni-plugins-7c22q" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.345263 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/b65abf9f-4e88-4571-960d-3ca997d9c344-host-run-multus-certs\") pod \"multus-szn8q\" (UID: \"b65abf9f-4e88-4571-960d-3ca997d9c344\") " pod="openshift-multus/multus-szn8q" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.344670 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/b65abf9f-4e88-4571-960d-3ca997d9c344-multus-cni-dir\") pod \"multus-szn8q\" (UID: \"b65abf9f-4e88-4571-960d-3ca997d9c344\") " pod="openshift-multus/multus-szn8q" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.345209 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/b7eaaa68-d971-4a1c-b6c5-551ded8ff27f-cnibin\") pod \"multus-additional-cni-plugins-7c22q\" (UID: \"b7eaaa68-d971-4a1c-b6c5-551ded8ff27f\") " pod="openshift-multus/multus-additional-cni-plugins-7c22q" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.345515 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/b65abf9f-4e88-4571-960d-3ca997d9c344-system-cni-dir\") pod \"multus-szn8q\" (UID: \"b65abf9f-4e88-4571-960d-3ca997d9c344\") " pod="openshift-multus/multus-szn8q" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.345512 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/b65abf9f-4e88-4571-960d-3ca997d9c344-os-release\") pod \"multus-szn8q\" (UID: \"b65abf9f-4e88-4571-960d-3ca997d9c344\") " pod="openshift-multus/multus-szn8q" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.345607 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/b65abf9f-4e88-4571-960d-3ca997d9c344-host-var-lib-cni-multus\") pod \"multus-szn8q\" (UID: \"b65abf9f-4e88-4571-960d-3ca997d9c344\") " pod="openshift-multus/multus-szn8q" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.345653 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/b65abf9f-4e88-4571-960d-3ca997d9c344-system-cni-dir\") pod \"multus-szn8q\" (UID: \"b65abf9f-4e88-4571-960d-3ca997d9c344\") " pod="openshift-multus/multus-szn8q" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.345689 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/b65abf9f-4e88-4571-960d-3ca997d9c344-host-var-lib-cni-multus\") pod \"multus-szn8q\" (UID: \"b65abf9f-4e88-4571-960d-3ca997d9c344\") " pod="openshift-multus/multus-szn8q" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.345652 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/b65abf9f-4e88-4571-960d-3ca997d9c344-host-var-lib-kubelet\") pod \"multus-szn8q\" (UID: \"b65abf9f-4e88-4571-960d-3ca997d9c344\") " pod="openshift-multus/multus-szn8q" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.345691 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/b65abf9f-4e88-4571-960d-3ca997d9c344-host-var-lib-kubelet\") pod \"multus-szn8q\" (UID: \"b65abf9f-4e88-4571-960d-3ca997d9c344\") " pod="openshift-multus/multus-szn8q" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.345759 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z7pmj\" (UniqueName: \"kubernetes.io/projected/b7eaaa68-d971-4a1c-b6c5-551ded8ff27f-kube-api-access-z7pmj\") pod \"multus-additional-cni-plugins-7c22q\" (UID: \"b7eaaa68-d971-4a1c-b6c5-551ded8ff27f\") " pod="openshift-multus/multus-additional-cni-plugins-7c22q" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.345808 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b65abf9f-4e88-4571-960d-3ca997d9c344-etc-kubernetes\") pod \"multus-szn8q\" (UID: \"b65abf9f-4e88-4571-960d-3ca997d9c344\") " pod="openshift-multus/multus-szn8q" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.345847 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/b7eaaa68-d971-4a1c-b6c5-551ded8ff27f-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-7c22q\" (UID: \"b7eaaa68-d971-4a1c-b6c5-551ded8ff27f\") " pod="openshift-multus/multus-additional-cni-plugins-7c22q" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.345860 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/b65abf9f-4e88-4571-960d-3ca997d9c344-cni-binary-copy\") pod \"multus-szn8q\" (UID: \"b65abf9f-4e88-4571-960d-3ca997d9c344\") " pod="openshift-multus/multus-szn8q" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.345891 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/b7eaaa68-d971-4a1c-b6c5-551ded8ff27f-cni-binary-copy\") pod \"multus-additional-cni-plugins-7c22q\" (UID: \"b7eaaa68-d971-4a1c-b6c5-551ded8ff27f\") " pod="openshift-multus/multus-additional-cni-plugins-7c22q" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.345906 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b65abf9f-4e88-4571-960d-3ca997d9c344-etc-kubernetes\") pod \"multus-szn8q\" (UID: \"b65abf9f-4e88-4571-960d-3ca997d9c344\") " pod="openshift-multus/multus-szn8q" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.345920 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e1140c8c-93f1-4fce-8c06-a69561ff8a6d-mcd-auth-proxy-config\") pod \"machine-config-daemon-khgzw\" (UID: \"e1140c8c-93f1-4fce-8c06-a69561ff8a6d\") " pod="openshift-machine-config-operator/machine-config-daemon-khgzw" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.345949 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/b65abf9f-4e88-4571-960d-3ca997d9c344-cnibin\") pod \"multus-szn8q\" (UID: \"b65abf9f-4e88-4571-960d-3ca997d9c344\") " pod="openshift-multus/multus-szn8q" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.346015 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/b7eaaa68-d971-4a1c-b6c5-551ded8ff27f-os-release\") pod \"multus-additional-cni-plugins-7c22q\" (UID: \"b7eaaa68-d971-4a1c-b6c5-551ded8ff27f\") " pod="openshift-multus/multus-additional-cni-plugins-7c22q" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.346128 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/b7eaaa68-d971-4a1c-b6c5-551ded8ff27f-os-release\") pod \"multus-additional-cni-plugins-7c22q\" (UID: \"b7eaaa68-d971-4a1c-b6c5-551ded8ff27f\") " pod="openshift-multus/multus-additional-cni-plugins-7c22q" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.346188 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/b65abf9f-4e88-4571-960d-3ca997d9c344-cnibin\") pod \"multus-szn8q\" (UID: \"b65abf9f-4e88-4571-960d-3ca997d9c344\") " pod="openshift-multus/multus-szn8q" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.346297 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/b65abf9f-4e88-4571-960d-3ca997d9c344-multus-daemon-config\") pod \"multus-szn8q\" (UID: \"b65abf9f-4e88-4571-960d-3ca997d9c344\") " pod="openshift-multus/multus-szn8q" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.346418 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/b65abf9f-4e88-4571-960d-3ca997d9c344-multus-conf-dir\") pod \"multus-szn8q\" (UID: \"b65abf9f-4e88-4571-960d-3ca997d9c344\") " pod="openshift-multus/multus-szn8q" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.346538 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/b7eaaa68-d971-4a1c-b6c5-551ded8ff27f-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-7c22q\" (UID: \"b7eaaa68-d971-4a1c-b6c5-551ded8ff27f\") " pod="openshift-multus/multus-additional-cni-plugins-7c22q" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.348284 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/b7eaaa68-d971-4a1c-b6c5-551ded8ff27f-cni-binary-copy\") pod \"multus-additional-cni-plugins-7c22q\" (UID: \"b7eaaa68-d971-4a1c-b6c5-551ded8ff27f\") " pod="openshift-multus/multus-additional-cni-plugins-7c22q" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.348829 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:32Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.350636 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/e1140c8c-93f1-4fce-8c06-a69561ff8a6d-proxy-tls\") pod \"machine-config-daemon-khgzw\" (UID: \"e1140c8c-93f1-4fce-8c06-a69561ff8a6d\") " pod="openshift-machine-config-operator/machine-config-daemon-khgzw" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.362069 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gq49w\" (UniqueName: \"kubernetes.io/projected/e1140c8c-93f1-4fce-8c06-a69561ff8a6d-kube-api-access-gq49w\") pod \"machine-config-daemon-khgzw\" (UID: \"e1140c8c-93f1-4fce-8c06-a69561ff8a6d\") " pod="openshift-machine-config-operator/machine-config-daemon-khgzw" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.365182 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-49s5w\" (UniqueName: \"kubernetes.io/projected/b65abf9f-4e88-4571-960d-3ca997d9c344-kube-api-access-49s5w\") pod \"multus-szn8q\" (UID: \"b65abf9f-4e88-4571-960d-3ca997d9c344\") " pod="openshift-multus/multus-szn8q" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.366366 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z7pmj\" (UniqueName: \"kubernetes.io/projected/b7eaaa68-d971-4a1c-b6c5-551ded8ff27f-kube-api-access-z7pmj\") pod \"multus-additional-cni-plugins-7c22q\" (UID: \"b7eaaa68-d971-4a1c-b6c5-551ded8ff27f\") " pod="openshift-multus/multus-additional-cni-plugins-7c22q" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.366688 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-szn8q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b65abf9f-4e88-4571-960d-3ca997d9c344\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49s5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-szn8q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:32Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.388545 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b665cda3-da4e-482e-8248-023abec382f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9235c742d8c45bf667f32979c2891fece50f4ee1bd0c7cf6a753ed2d2b9ab30e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b2a920dbc506df2426232990e31fbb60d255c00be54941826a0f094af4a92f0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d516f278f362d7c45e7149de82fa835997ac198f1165c04352910d51fb674d69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b8cdf54407cc95783919eba8e92f093bbc09eac5e908c4ba0685629948c71bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b8cdf54407cc95783919eba8e92f093bbc09eac5e908c4ba0685629948c71bd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-06T09:58:27Z\\\",\\\"message\\\":\\\"arting RequestHeaderAuthRequestController\\\\nI1006 09:58:27.156944 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1006 09:58:27.156783 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1006 09:58:27.157213 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-432254888/tls.crt::/tmp/serving-cert-432254888/tls.key\\\\\\\"\\\\nI1006 09:58:27.157407 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1006 09:58:27.157443 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1006 09:58:27.157410 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-432254888/tls.crt::/tmp/serving-cert-432254888/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759744705\\\\\\\\\\\\\\\" (2025-10-06 09:58:25 +0000 UTC to 2025-11-05 09:58:26 +0000 UTC (now=2025-10-06 09:58:27.157390076 +0000 UTC))\\\\\\\"\\\\nI1006 09:58:27.157632 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1006 09:58:27.158158 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759744707\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759744707\\\\\\\\\\\\\\\" (2025-10-06 08:58:26 +0000 UTC to 2026-10-06 08:58:26 +0000 UTC (now=2025-10-06 09:58:27.158090663 +0000 UTC))\\\\\\\"\\\\nI1006 09:58:27.158220 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1006 09:58:27.158270 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1006 09:58:27.158310 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nF1006 09:58:27.158617 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://413822c99ba2866521177cce4f1894aef17c160aebb2662da91824a04827d510\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e37e8c7400f7fd14a5c5390a5985e3d8d277e9c5ca60d5e9e9757d07d1778570\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e37e8c7400f7fd14a5c5390a5985e3d8d277e9c5ca60d5e9e9757d07d1778570\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:32Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.405422 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b665cda3-da4e-482e-8248-023abec382f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9235c742d8c45bf667f32979c2891fece50f4ee1bd0c7cf6a753ed2d2b9ab30e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b2a920dbc506df2426232990e31fbb60d255c00be54941826a0f094af4a92f0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d516f278f362d7c45e7149de82fa835997ac198f1165c04352910d51fb674d69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b8cdf54407cc95783919eba8e92f093bbc09eac5e908c4ba0685629948c71bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b8cdf54407cc95783919eba8e92f093bbc09eac5e908c4ba0685629948c71bd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-06T09:58:27Z\\\",\\\"message\\\":\\\"arting RequestHeaderAuthRequestController\\\\nI1006 09:58:27.156944 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1006 09:58:27.156783 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1006 09:58:27.157213 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-432254888/tls.crt::/tmp/serving-cert-432254888/tls.key\\\\\\\"\\\\nI1006 09:58:27.157407 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1006 09:58:27.157443 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1006 09:58:27.157410 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-432254888/tls.crt::/tmp/serving-cert-432254888/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759744705\\\\\\\\\\\\\\\" (2025-10-06 09:58:25 +0000 UTC to 2025-11-05 09:58:26 +0000 UTC (now=2025-10-06 09:58:27.157390076 +0000 UTC))\\\\\\\"\\\\nI1006 09:58:27.157632 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1006 09:58:27.158158 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759744707\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759744707\\\\\\\\\\\\\\\" (2025-10-06 08:58:26 +0000 UTC to 2026-10-06 08:58:26 +0000 UTC (now=2025-10-06 09:58:27.158090663 +0000 UTC))\\\\\\\"\\\\nI1006 09:58:27.158220 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1006 09:58:27.158270 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1006 09:58:27.158310 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nF1006 09:58:27.158617 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://413822c99ba2866521177cce4f1894aef17c160aebb2662da91824a04827d510\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e37e8c7400f7fd14a5c5390a5985e3d8d277e9c5ca60d5e9e9757d07d1778570\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e37e8c7400f7fd14a5c5390a5985e3d8d277e9c5ca60d5e9e9757d07d1778570\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:32Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.430149 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d739433-0a38-42fe-b884-b8433011967e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://464e279c5903770c34d80f31ed231b194d5628b39055c4058908039d10690059\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://efa1b62728466ef7aac0e745230348ce920eb9ec9b07fb2cd5a276150dac13fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fedc40f0217626c56d155fc7e78e0592556fef9464db2aaf09c5facc486aad8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78fac300cbe2ac52f5788306b8fc0011f6ba3a22c3546974e70998f2d394987a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:32Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.458191 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"2035278cf272827b39a45aa41ecd5b1e41aafc1d9eb8bccd853660fb6e587159"} Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.458249 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"ff3152fceadf36a55578f587623f37d748f90f2afeda5702a05504b372d399be"} Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.459552 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"2f9011f6673310dffea3acd7bdc2a56f2af91a87641631490fdc63da610f047a"} Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.461121 4824 scope.go:117] "RemoveContainer" containerID="0b8cdf54407cc95783919eba8e92f093bbc09eac5e908c4ba0685629948c71bd" Oct 06 09:58:32 crc kubenswrapper[4824]: E1006 09:58:32.461242 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.461391 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-xcgwc" event={"ID":"74317f75-9ebc-4327-b88c-598c2241240f","Type":"ContainerStarted","Data":"8def09a126f9bb910affad9b7fb32303b960add9fbfd5bb709d59b4aeba7d245"} Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.461422 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-xcgwc" event={"ID":"74317f75-9ebc-4327-b88c-598c2241240f","Type":"ContainerStarted","Data":"bf05486bc610482fa7ea1ccb39a39b55a38904af509ea2b474fa6159fd35b9f7"} Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.462169 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:32Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.480952 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-szn8q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b65abf9f-4e88-4571-960d-3ca997d9c344\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49s5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-szn8q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:32Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.500123 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-7c22q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7eaaa68-d971-4a1c-b6c5-551ded8ff27f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-7c22q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:32Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.512873 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:32Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.512929 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.529903 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-szn8q" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.538188 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-7c22q" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.547916 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xcgwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74317f75-9ebc-4327-b88c-598c2241240f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v6sg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xcgwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:32Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.552096 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 06 09:58:32 crc kubenswrapper[4824]: W1006 09:58:32.561061 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb65abf9f_4e88_4571_960d_3ca997d9c344.slice/crio-850c316503dae833f8ec3e9794734f84d88e54a512b3197aa9404ec594035b8e WatchSource:0}: Error finding container 850c316503dae833f8ec3e9794734f84d88e54a512b3197aa9404ec594035b8e: Status 404 returned error can't find the container with id 850c316503dae833f8ec3e9794734f84d88e54a512b3197aa9404ec594035b8e Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.570743 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1140c8c-93f1-4fce-8c06-a69561ff8a6d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gq49w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gq49w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-khgzw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:32Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.588109 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-4vwms"] Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.589907 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:32Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.591928 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.599130 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.600690 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.600734 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.600883 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.611911 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:32Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.612170 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.612282 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.612765 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.635350 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:32Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.653151 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/2f511ece-25eb-465f-b85c-9e5f3f886c21-ovnkube-config\") pod \"ovnkube-node-4vwms\" (UID: \"2f511ece-25eb-465f-b85c-9e5f3f886c21\") " pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.653234 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/2f511ece-25eb-465f-b85c-9e5f3f886c21-host-kubelet\") pod \"ovnkube-node-4vwms\" (UID: \"2f511ece-25eb-465f-b85c-9e5f3f886c21\") " pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.653305 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/2f511ece-25eb-465f-b85c-9e5f3f886c21-host-slash\") pod \"ovnkube-node-4vwms\" (UID: \"2f511ece-25eb-465f-b85c-9e5f3f886c21\") " pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.653332 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/2f511ece-25eb-465f-b85c-9e5f3f886c21-run-ovn\") pod \"ovnkube-node-4vwms\" (UID: \"2f511ece-25eb-465f-b85c-9e5f3f886c21\") " pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.653356 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2f511ece-25eb-465f-b85c-9e5f3f886c21-host-run-ovn-kubernetes\") pod \"ovnkube-node-4vwms\" (UID: \"2f511ece-25eb-465f-b85c-9e5f3f886c21\") " pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.653379 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/2f511ece-25eb-465f-b85c-9e5f3f886c21-systemd-units\") pod \"ovnkube-node-4vwms\" (UID: \"2f511ece-25eb-465f-b85c-9e5f3f886c21\") " pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.653400 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/2f511ece-25eb-465f-b85c-9e5f3f886c21-run-systemd\") pod \"ovnkube-node-4vwms\" (UID: \"2f511ece-25eb-465f-b85c-9e5f3f886c21\") " pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.653516 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2f511ece-25eb-465f-b85c-9e5f3f886c21-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-4vwms\" (UID: \"2f511ece-25eb-465f-b85c-9e5f3f886c21\") " pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.653613 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/2f511ece-25eb-465f-b85c-9e5f3f886c21-ovnkube-script-lib\") pod \"ovnkube-node-4vwms\" (UID: \"2f511ece-25eb-465f-b85c-9e5f3f886c21\") " pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.653660 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2f511ece-25eb-465f-b85c-9e5f3f886c21-run-openvswitch\") pod \"ovnkube-node-4vwms\" (UID: \"2f511ece-25eb-465f-b85c-9e5f3f886c21\") " pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.653705 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/2f511ece-25eb-465f-b85c-9e5f3f886c21-host-cni-bin\") pod \"ovnkube-node-4vwms\" (UID: \"2f511ece-25eb-465f-b85c-9e5f3f886c21\") " pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.653737 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mc4k6\" (UniqueName: \"kubernetes.io/projected/2f511ece-25eb-465f-b85c-9e5f3f886c21-kube-api-access-mc4k6\") pod \"ovnkube-node-4vwms\" (UID: \"2f511ece-25eb-465f-b85c-9e5f3f886c21\") " pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.653786 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/2f511ece-25eb-465f-b85c-9e5f3f886c21-host-run-netns\") pod \"ovnkube-node-4vwms\" (UID: \"2f511ece-25eb-465f-b85c-9e5f3f886c21\") " pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.653812 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/2f511ece-25eb-465f-b85c-9e5f3f886c21-ovn-node-metrics-cert\") pod \"ovnkube-node-4vwms\" (UID: \"2f511ece-25eb-465f-b85c-9e5f3f886c21\") " pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.653843 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2f511ece-25eb-465f-b85c-9e5f3f886c21-etc-openvswitch\") pod \"ovnkube-node-4vwms\" (UID: \"2f511ece-25eb-465f-b85c-9e5f3f886c21\") " pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.653873 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/2f511ece-25eb-465f-b85c-9e5f3f886c21-node-log\") pod \"ovnkube-node-4vwms\" (UID: \"2f511ece-25eb-465f-b85c-9e5f3f886c21\") " pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.653899 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/2f511ece-25eb-465f-b85c-9e5f3f886c21-log-socket\") pod \"ovnkube-node-4vwms\" (UID: \"2f511ece-25eb-465f-b85c-9e5f3f886c21\") " pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.653927 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/2f511ece-25eb-465f-b85c-9e5f3f886c21-host-cni-netd\") pod \"ovnkube-node-4vwms\" (UID: \"2f511ece-25eb-465f-b85c-9e5f3f886c21\") " pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.653954 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/2f511ece-25eb-465f-b85c-9e5f3f886c21-env-overrides\") pod \"ovnkube-node-4vwms\" (UID: \"2f511ece-25eb-465f-b85c-9e5f3f886c21\") " pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.654001 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2f511ece-25eb-465f-b85c-9e5f3f886c21-var-lib-openvswitch\") pod \"ovnkube-node-4vwms\" (UID: \"2f511ece-25eb-465f-b85c-9e5f3f886c21\") " pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.657655 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:32Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.674254 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2035278cf272827b39a45aa41ecd5b1e41aafc1d9eb8bccd853660fb6e587159\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff3152fceadf36a55578f587623f37d748f90f2afeda5702a05504b372d399be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:32Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.701229 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xcgwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74317f75-9ebc-4327-b88c-598c2241240f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8def09a126f9bb910affad9b7fb32303b960add9fbfd5bb709d59b4aeba7d245\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v6sg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xcgwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:32Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.728431 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1140c8c-93f1-4fce-8c06-a69561ff8a6d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gq49w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gq49w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-khgzw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:32Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.754472 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/2f511ece-25eb-465f-b85c-9e5f3f886c21-ovnkube-script-lib\") pod \"ovnkube-node-4vwms\" (UID: \"2f511ece-25eb-465f-b85c-9e5f3f886c21\") " pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.754515 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2f511ece-25eb-465f-b85c-9e5f3f886c21-run-openvswitch\") pod \"ovnkube-node-4vwms\" (UID: \"2f511ece-25eb-465f-b85c-9e5f3f886c21\") " pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.754543 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/2f511ece-25eb-465f-b85c-9e5f3f886c21-host-cni-bin\") pod \"ovnkube-node-4vwms\" (UID: \"2f511ece-25eb-465f-b85c-9e5f3f886c21\") " pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.754562 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mc4k6\" (UniqueName: \"kubernetes.io/projected/2f511ece-25eb-465f-b85c-9e5f3f886c21-kube-api-access-mc4k6\") pod \"ovnkube-node-4vwms\" (UID: \"2f511ece-25eb-465f-b85c-9e5f3f886c21\") " pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.754586 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/2f511ece-25eb-465f-b85c-9e5f3f886c21-host-run-netns\") pod \"ovnkube-node-4vwms\" (UID: \"2f511ece-25eb-465f-b85c-9e5f3f886c21\") " pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.754604 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/2f511ece-25eb-465f-b85c-9e5f3f886c21-ovn-node-metrics-cert\") pod \"ovnkube-node-4vwms\" (UID: \"2f511ece-25eb-465f-b85c-9e5f3f886c21\") " pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.754638 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2f511ece-25eb-465f-b85c-9e5f3f886c21-etc-openvswitch\") pod \"ovnkube-node-4vwms\" (UID: \"2f511ece-25eb-465f-b85c-9e5f3f886c21\") " pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.754657 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/2f511ece-25eb-465f-b85c-9e5f3f886c21-node-log\") pod \"ovnkube-node-4vwms\" (UID: \"2f511ece-25eb-465f-b85c-9e5f3f886c21\") " pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.754660 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2f511ece-25eb-465f-b85c-9e5f3f886c21-run-openvswitch\") pod \"ovnkube-node-4vwms\" (UID: \"2f511ece-25eb-465f-b85c-9e5f3f886c21\") " pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.754731 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2f511ece-25eb-465f-b85c-9e5f3f886c21-etc-openvswitch\") pod \"ovnkube-node-4vwms\" (UID: \"2f511ece-25eb-465f-b85c-9e5f3f886c21\") " pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.754678 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/2f511ece-25eb-465f-b85c-9e5f3f886c21-log-socket\") pod \"ovnkube-node-4vwms\" (UID: \"2f511ece-25eb-465f-b85c-9e5f3f886c21\") " pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.754773 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/2f511ece-25eb-465f-b85c-9e5f3f886c21-host-cni-netd\") pod \"ovnkube-node-4vwms\" (UID: \"2f511ece-25eb-465f-b85c-9e5f3f886c21\") " pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.754814 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/2f511ece-25eb-465f-b85c-9e5f3f886c21-node-log\") pod \"ovnkube-node-4vwms\" (UID: \"2f511ece-25eb-465f-b85c-9e5f3f886c21\") " pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.754833 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/2f511ece-25eb-465f-b85c-9e5f3f886c21-env-overrides\") pod \"ovnkube-node-4vwms\" (UID: \"2f511ece-25eb-465f-b85c-9e5f3f886c21\") " pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.754856 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2f511ece-25eb-465f-b85c-9e5f3f886c21-var-lib-openvswitch\") pod \"ovnkube-node-4vwms\" (UID: \"2f511ece-25eb-465f-b85c-9e5f3f886c21\") " pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.754876 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/2f511ece-25eb-465f-b85c-9e5f3f886c21-ovnkube-config\") pod \"ovnkube-node-4vwms\" (UID: \"2f511ece-25eb-465f-b85c-9e5f3f886c21\") " pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.754882 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/2f511ece-25eb-465f-b85c-9e5f3f886c21-host-run-netns\") pod \"ovnkube-node-4vwms\" (UID: \"2f511ece-25eb-465f-b85c-9e5f3f886c21\") " pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.754912 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/2f511ece-25eb-465f-b85c-9e5f3f886c21-host-kubelet\") pod \"ovnkube-node-4vwms\" (UID: \"2f511ece-25eb-465f-b85c-9e5f3f886c21\") " pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.754893 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/2f511ece-25eb-465f-b85c-9e5f3f886c21-host-kubelet\") pod \"ovnkube-node-4vwms\" (UID: \"2f511ece-25eb-465f-b85c-9e5f3f886c21\") " pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.754669 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/2f511ece-25eb-465f-b85c-9e5f3f886c21-host-cni-bin\") pod \"ovnkube-node-4vwms\" (UID: \"2f511ece-25eb-465f-b85c-9e5f3f886c21\") " pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.754964 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/2f511ece-25eb-465f-b85c-9e5f3f886c21-host-cni-netd\") pod \"ovnkube-node-4vwms\" (UID: \"2f511ece-25eb-465f-b85c-9e5f3f886c21\") " pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.755314 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/2f511ece-25eb-465f-b85c-9e5f3f886c21-log-socket\") pod \"ovnkube-node-4vwms\" (UID: \"2f511ece-25eb-465f-b85c-9e5f3f886c21\") " pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.755438 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2f511ece-25eb-465f-b85c-9e5f3f886c21-var-lib-openvswitch\") pod \"ovnkube-node-4vwms\" (UID: \"2f511ece-25eb-465f-b85c-9e5f3f886c21\") " pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.755447 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2f511ece-25eb-465f-b85c-9e5f3f886c21-host-run-ovn-kubernetes\") pod \"ovnkube-node-4vwms\" (UID: \"2f511ece-25eb-465f-b85c-9e5f3f886c21\") " pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.755502 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/2f511ece-25eb-465f-b85c-9e5f3f886c21-ovnkube-script-lib\") pod \"ovnkube-node-4vwms\" (UID: \"2f511ece-25eb-465f-b85c-9e5f3f886c21\") " pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.755521 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/2f511ece-25eb-465f-b85c-9e5f3f886c21-env-overrides\") pod \"ovnkube-node-4vwms\" (UID: \"2f511ece-25eb-465f-b85c-9e5f3f886c21\") " pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.755544 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2f511ece-25eb-465f-b85c-9e5f3f886c21-host-run-ovn-kubernetes\") pod \"ovnkube-node-4vwms\" (UID: \"2f511ece-25eb-465f-b85c-9e5f3f886c21\") " pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.755527 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/2f511ece-25eb-465f-b85c-9e5f3f886c21-ovnkube-config\") pod \"ovnkube-node-4vwms\" (UID: \"2f511ece-25eb-465f-b85c-9e5f3f886c21\") " pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.755526 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/2f511ece-25eb-465f-b85c-9e5f3f886c21-host-slash\") pod \"ovnkube-node-4vwms\" (UID: \"2f511ece-25eb-465f-b85c-9e5f3f886c21\") " pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.755604 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/2f511ece-25eb-465f-b85c-9e5f3f886c21-run-ovn\") pod \"ovnkube-node-4vwms\" (UID: \"2f511ece-25eb-465f-b85c-9e5f3f886c21\") " pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.755566 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/2f511ece-25eb-465f-b85c-9e5f3f886c21-host-slash\") pod \"ovnkube-node-4vwms\" (UID: \"2f511ece-25eb-465f-b85c-9e5f3f886c21\") " pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.755625 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/2f511ece-25eb-465f-b85c-9e5f3f886c21-systemd-units\") pod \"ovnkube-node-4vwms\" (UID: \"2f511ece-25eb-465f-b85c-9e5f3f886c21\") " pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.755644 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/2f511ece-25eb-465f-b85c-9e5f3f886c21-run-systemd\") pod \"ovnkube-node-4vwms\" (UID: \"2f511ece-25eb-465f-b85c-9e5f3f886c21\") " pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.755664 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2f511ece-25eb-465f-b85c-9e5f3f886c21-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-4vwms\" (UID: \"2f511ece-25eb-465f-b85c-9e5f3f886c21\") " pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.755668 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/2f511ece-25eb-465f-b85c-9e5f3f886c21-run-ovn\") pod \"ovnkube-node-4vwms\" (UID: \"2f511ece-25eb-465f-b85c-9e5f3f886c21\") " pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.755712 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/2f511ece-25eb-465f-b85c-9e5f3f886c21-systemd-units\") pod \"ovnkube-node-4vwms\" (UID: \"2f511ece-25eb-465f-b85c-9e5f3f886c21\") " pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.755709 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/2f511ece-25eb-465f-b85c-9e5f3f886c21-run-systemd\") pod \"ovnkube-node-4vwms\" (UID: \"2f511ece-25eb-465f-b85c-9e5f3f886c21\") " pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.755758 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2f511ece-25eb-465f-b85c-9e5f3f886c21-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-4vwms\" (UID: \"2f511ece-25eb-465f-b85c-9e5f3f886c21\") " pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.765496 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/2f511ece-25eb-465f-b85c-9e5f3f886c21-ovn-node-metrics-cert\") pod \"ovnkube-node-4vwms\" (UID: \"2f511ece-25eb-465f-b85c-9e5f3f886c21\") " pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.774639 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f9011f6673310dffea3acd7bdc2a56f2af91a87641631490fdc63da610f047a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:32Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.797745 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mc4k6\" (UniqueName: \"kubernetes.io/projected/2f511ece-25eb-465f-b85c-9e5f3f886c21-kube-api-access-mc4k6\") pod \"ovnkube-node-4vwms\" (UID: \"2f511ece-25eb-465f-b85c-9e5f3f886c21\") " pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.826720 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:32Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.867474 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:32Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.913899 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:32Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.927211 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" Oct 06 09:58:32 crc kubenswrapper[4824]: W1006 09:58:32.948125 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2f511ece_25eb_465f_b85c_9e5f3f886c21.slice/crio-b7982e1a18ea2e6f50d59ed4fcc3bb8dd2242517d06b86feab7d9b463fca89cb WatchSource:0}: Error finding container b7982e1a18ea2e6f50d59ed4fcc3bb8dd2242517d06b86feab7d9b463fca89cb: Status 404 returned error can't find the container with id b7982e1a18ea2e6f50d59ed4fcc3bb8dd2242517d06b86feab7d9b463fca89cb Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.964099 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b665cda3-da4e-482e-8248-023abec382f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9235c742d8c45bf667f32979c2891fece50f4ee1bd0c7cf6a753ed2d2b9ab30e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b2a920dbc506df2426232990e31fbb60d255c00be54941826a0f094af4a92f0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d516f278f362d7c45e7149de82fa835997ac198f1165c04352910d51fb674d69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b8cdf54407cc95783919eba8e92f093bbc09eac5e908c4ba0685629948c71bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b8cdf54407cc95783919eba8e92f093bbc09eac5e908c4ba0685629948c71bd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-06T09:58:27Z\\\",\\\"message\\\":\\\"arting RequestHeaderAuthRequestController\\\\nI1006 09:58:27.156944 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1006 09:58:27.156783 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1006 09:58:27.157213 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-432254888/tls.crt::/tmp/serving-cert-432254888/tls.key\\\\\\\"\\\\nI1006 09:58:27.157407 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1006 09:58:27.157443 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1006 09:58:27.157410 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-432254888/tls.crt::/tmp/serving-cert-432254888/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759744705\\\\\\\\\\\\\\\" (2025-10-06 09:58:25 +0000 UTC to 2025-11-05 09:58:26 +0000 UTC (now=2025-10-06 09:58:27.157390076 +0000 UTC))\\\\\\\"\\\\nI1006 09:58:27.157632 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1006 09:58:27.158158 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759744707\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759744707\\\\\\\\\\\\\\\" (2025-10-06 08:58:26 +0000 UTC to 2026-10-06 08:58:26 +0000 UTC (now=2025-10-06 09:58:27.158090663 +0000 UTC))\\\\\\\"\\\\nI1006 09:58:27.158220 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1006 09:58:27.158270 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1006 09:58:27.158310 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nF1006 09:58:27.158617 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://413822c99ba2866521177cce4f1894aef17c160aebb2662da91824a04827d510\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e37e8c7400f7fd14a5c5390a5985e3d8d277e9c5ca60d5e9e9757d07d1778570\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e37e8c7400f7fd14a5c5390a5985e3d8d277e9c5ca60d5e9e9757d07d1778570\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:32Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:32 crc kubenswrapper[4824]: I1006 09:58:32.993427 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d739433-0a38-42fe-b884-b8433011967e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://464e279c5903770c34d80f31ed231b194d5628b39055c4058908039d10690059\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://efa1b62728466ef7aac0e745230348ce920eb9ec9b07fb2cd5a276150dac13fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fedc40f0217626c56d155fc7e78e0592556fef9464db2aaf09c5facc486aad8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78fac300cbe2ac52f5788306b8fc0011f6ba3a22c3546974e70998f2d394987a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:32Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:33 crc kubenswrapper[4824]: I1006 09:58:33.024097 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:33Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:33 crc kubenswrapper[4824]: I1006 09:58:33.065282 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-szn8q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b65abf9f-4e88-4571-960d-3ca997d9c344\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49s5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-szn8q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:33Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:33 crc kubenswrapper[4824]: I1006 09:58:33.116219 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f511ece-25eb-465f-b85c-9e5f3f886c21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4vwms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:33Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:33 crc kubenswrapper[4824]: I1006 09:58:33.151971 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-7c22q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7eaaa68-d971-4a1c-b6c5-551ded8ff27f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-7c22q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:33Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:33 crc kubenswrapper[4824]: I1006 09:58:33.466534 4824 generic.go:334] "Generic (PLEG): container finished" podID="b7eaaa68-d971-4a1c-b6c5-551ded8ff27f" containerID="53336c7d3c509b4e022708998aedc5dee9fdaf7136ae3ee69b4181394c3a22fb" exitCode=0 Oct 06 09:58:33 crc kubenswrapper[4824]: I1006 09:58:33.466611 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-7c22q" event={"ID":"b7eaaa68-d971-4a1c-b6c5-551ded8ff27f","Type":"ContainerDied","Data":"53336c7d3c509b4e022708998aedc5dee9fdaf7136ae3ee69b4181394c3a22fb"} Oct 06 09:58:33 crc kubenswrapper[4824]: I1006 09:58:33.466638 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-7c22q" event={"ID":"b7eaaa68-d971-4a1c-b6c5-551ded8ff27f","Type":"ContainerStarted","Data":"304b4ad7d517472db6949799578707ef6995a39093810c3f8f6d94067595da12"} Oct 06 09:58:33 crc kubenswrapper[4824]: I1006 09:58:33.470553 4824 generic.go:334] "Generic (PLEG): container finished" podID="2f511ece-25eb-465f-b85c-9e5f3f886c21" containerID="6da4c07be4d507a6f20bd8b42be6e6ea4712a956f57ef164036cfded7142a87e" exitCode=0 Oct 06 09:58:33 crc kubenswrapper[4824]: I1006 09:58:33.470645 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" event={"ID":"2f511ece-25eb-465f-b85c-9e5f3f886c21","Type":"ContainerDied","Data":"6da4c07be4d507a6f20bd8b42be6e6ea4712a956f57ef164036cfded7142a87e"} Oct 06 09:58:33 crc kubenswrapper[4824]: I1006 09:58:33.470669 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" event={"ID":"2f511ece-25eb-465f-b85c-9e5f3f886c21","Type":"ContainerStarted","Data":"b7982e1a18ea2e6f50d59ed4fcc3bb8dd2242517d06b86feab7d9b463fca89cb"} Oct 06 09:58:33 crc kubenswrapper[4824]: I1006 09:58:33.475143 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-szn8q" event={"ID":"b65abf9f-4e88-4571-960d-3ca997d9c344","Type":"ContainerStarted","Data":"664bf42e1324a9c9529f36f08d61e26999b3386bdd1fac77851fb7e7411e8934"} Oct 06 09:58:33 crc kubenswrapper[4824]: I1006 09:58:33.475281 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-szn8q" event={"ID":"b65abf9f-4e88-4571-960d-3ca997d9c344","Type":"ContainerStarted","Data":"850c316503dae833f8ec3e9794734f84d88e54a512b3197aa9404ec594035b8e"} Oct 06 09:58:33 crc kubenswrapper[4824]: I1006 09:58:33.483918 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" event={"ID":"e1140c8c-93f1-4fce-8c06-a69561ff8a6d","Type":"ContainerStarted","Data":"5462cb77b7d61aed7d4b4be2fbc6ba1451d267a66afd6dede2f259c15230e7c6"} Oct 06 09:58:33 crc kubenswrapper[4824]: I1006 09:58:33.484049 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" event={"ID":"e1140c8c-93f1-4fce-8c06-a69561ff8a6d","Type":"ContainerStarted","Data":"5512be36013baeca611ee92645a1d339da237b70d63aac43334beafe24513d44"} Oct 06 09:58:33 crc kubenswrapper[4824]: I1006 09:58:33.484069 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" event={"ID":"e1140c8c-93f1-4fce-8c06-a69561ff8a6d","Type":"ContainerStarted","Data":"aa95a4bac0bde57ebd2b79568d25851f9dcb2b481e5afde839f66e69bdb6f980"} Oct 06 09:58:33 crc kubenswrapper[4824]: I1006 09:58:33.484651 4824 scope.go:117] "RemoveContainer" containerID="0b8cdf54407cc95783919eba8e92f093bbc09eac5e908c4ba0685629948c71bd" Oct 06 09:58:33 crc kubenswrapper[4824]: E1006 09:58:33.484869 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Oct 06 09:58:33 crc kubenswrapper[4824]: I1006 09:58:33.496357 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:33Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:33 crc kubenswrapper[4824]: I1006 09:58:33.518105 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:33Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:33 crc kubenswrapper[4824]: I1006 09:58:33.534678 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:33Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:33 crc kubenswrapper[4824]: I1006 09:58:33.550489 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f9011f6673310dffea3acd7bdc2a56f2af91a87641631490fdc63da610f047a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:33Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:33 crc kubenswrapper[4824]: I1006 09:58:33.568302 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-szn8q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b65abf9f-4e88-4571-960d-3ca997d9c344\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49s5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-szn8q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:33Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:33 crc kubenswrapper[4824]: I1006 09:58:33.589246 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f511ece-25eb-465f-b85c-9e5f3f886c21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4vwms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:33Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:33 crc kubenswrapper[4824]: I1006 09:58:33.606360 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b665cda3-da4e-482e-8248-023abec382f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9235c742d8c45bf667f32979c2891fece50f4ee1bd0c7cf6a753ed2d2b9ab30e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b2a920dbc506df2426232990e31fbb60d255c00be54941826a0f094af4a92f0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d516f278f362d7c45e7149de82fa835997ac198f1165c04352910d51fb674d69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b8cdf54407cc95783919eba8e92f093bbc09eac5e908c4ba0685629948c71bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b8cdf54407cc95783919eba8e92f093bbc09eac5e908c4ba0685629948c71bd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-06T09:58:27Z\\\",\\\"message\\\":\\\"arting RequestHeaderAuthRequestController\\\\nI1006 09:58:27.156944 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1006 09:58:27.156783 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1006 09:58:27.157213 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-432254888/tls.crt::/tmp/serving-cert-432254888/tls.key\\\\\\\"\\\\nI1006 09:58:27.157407 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1006 09:58:27.157443 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1006 09:58:27.157410 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-432254888/tls.crt::/tmp/serving-cert-432254888/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759744705\\\\\\\\\\\\\\\" (2025-10-06 09:58:25 +0000 UTC to 2025-11-05 09:58:26 +0000 UTC (now=2025-10-06 09:58:27.157390076 +0000 UTC))\\\\\\\"\\\\nI1006 09:58:27.157632 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1006 09:58:27.158158 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759744707\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759744707\\\\\\\\\\\\\\\" (2025-10-06 08:58:26 +0000 UTC to 2026-10-06 08:58:26 +0000 UTC (now=2025-10-06 09:58:27.158090663 +0000 UTC))\\\\\\\"\\\\nI1006 09:58:27.158220 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1006 09:58:27.158270 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1006 09:58:27.158310 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nF1006 09:58:27.158617 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://413822c99ba2866521177cce4f1894aef17c160aebb2662da91824a04827d510\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e37e8c7400f7fd14a5c5390a5985e3d8d277e9c5ca60d5e9e9757d07d1778570\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e37e8c7400f7fd14a5c5390a5985e3d8d277e9c5ca60d5e9e9757d07d1778570\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:33Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:33 crc kubenswrapper[4824]: I1006 09:58:33.619425 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d739433-0a38-42fe-b884-b8433011967e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://464e279c5903770c34d80f31ed231b194d5628b39055c4058908039d10690059\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://efa1b62728466ef7aac0e745230348ce920eb9ec9b07fb2cd5a276150dac13fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fedc40f0217626c56d155fc7e78e0592556fef9464db2aaf09c5facc486aad8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78fac300cbe2ac52f5788306b8fc0011f6ba3a22c3546974e70998f2d394987a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:33Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:33 crc kubenswrapper[4824]: I1006 09:58:33.633216 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:33Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:33 crc kubenswrapper[4824]: I1006 09:58:33.661850 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-7c22q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7eaaa68-d971-4a1c-b6c5-551ded8ff27f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53336c7d3c509b4e022708998aedc5dee9fdaf7136ae3ee69b4181394c3a22fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53336c7d3c509b4e022708998aedc5dee9fdaf7136ae3ee69b4181394c3a22fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-7c22q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:33Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:33 crc kubenswrapper[4824]: I1006 09:58:33.675934 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2035278cf272827b39a45aa41ecd5b1e41aafc1d9eb8bccd853660fb6e587159\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff3152fceadf36a55578f587623f37d748f90f2afeda5702a05504b372d399be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:33Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:33 crc kubenswrapper[4824]: I1006 09:58:33.687412 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xcgwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74317f75-9ebc-4327-b88c-598c2241240f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8def09a126f9bb910affad9b7fb32303b960add9fbfd5bb709d59b4aeba7d245\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v6sg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xcgwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:33Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:33 crc kubenswrapper[4824]: I1006 09:58:33.700931 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1140c8c-93f1-4fce-8c06-a69561ff8a6d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gq49w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gq49w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-khgzw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:33Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:33 crc kubenswrapper[4824]: I1006 09:58:33.736216 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-7c22q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7eaaa68-d971-4a1c-b6c5-551ded8ff27f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53336c7d3c509b4e022708998aedc5dee9fdaf7136ae3ee69b4181394c3a22fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53336c7d3c509b4e022708998aedc5dee9fdaf7136ae3ee69b4181394c3a22fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-7c22q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:33Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:33 crc kubenswrapper[4824]: I1006 09:58:33.754141 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2035278cf272827b39a45aa41ecd5b1e41aafc1d9eb8bccd853660fb6e587159\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff3152fceadf36a55578f587623f37d748f90f2afeda5702a05504b372d399be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:33Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:33 crc kubenswrapper[4824]: I1006 09:58:33.782164 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xcgwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74317f75-9ebc-4327-b88c-598c2241240f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8def09a126f9bb910affad9b7fb32303b960add9fbfd5bb709d59b4aeba7d245\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v6sg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xcgwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:33Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:33 crc kubenswrapper[4824]: I1006 09:58:33.825577 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1140c8c-93f1-4fce-8c06-a69561ff8a6d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5462cb77b7d61aed7d4b4be2fbc6ba1451d267a66afd6dede2f259c15230e7c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gq49w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5512be36013baeca611ee92645a1d339da237b70d63aac43334beafe24513d44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gq49w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-khgzw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:33Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:33 crc kubenswrapper[4824]: I1006 09:58:33.864358 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f9011f6673310dffea3acd7bdc2a56f2af91a87641631490fdc63da610f047a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:33Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:33 crc kubenswrapper[4824]: I1006 09:58:33.906942 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:33Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:33 crc kubenswrapper[4824]: I1006 09:58:33.953575 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:33Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:33 crc kubenswrapper[4824]: I1006 09:58:33.968109 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 09:58:33 crc kubenswrapper[4824]: I1006 09:58:33.968283 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 09:58:33 crc kubenswrapper[4824]: E1006 09:58:33.968315 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-06 09:58:37.968272327 +0000 UTC m=+27.332695218 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 09:58:33 crc kubenswrapper[4824]: I1006 09:58:33.968377 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 09:58:33 crc kubenswrapper[4824]: E1006 09:58:33.968461 4824 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 06 09:58:33 crc kubenswrapper[4824]: E1006 09:58:33.968571 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-06 09:58:37.968553063 +0000 UTC m=+27.332975954 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 06 09:58:33 crc kubenswrapper[4824]: E1006 09:58:33.968608 4824 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 06 09:58:33 crc kubenswrapper[4824]: E1006 09:58:33.968734 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-06 09:58:37.968703897 +0000 UTC m=+27.333126768 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 06 09:58:33 crc kubenswrapper[4824]: I1006 09:58:33.992852 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:33Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:34 crc kubenswrapper[4824]: I1006 09:58:34.053505 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b665cda3-da4e-482e-8248-023abec382f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9235c742d8c45bf667f32979c2891fece50f4ee1bd0c7cf6a753ed2d2b9ab30e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b2a920dbc506df2426232990e31fbb60d255c00be54941826a0f094af4a92f0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d516f278f362d7c45e7149de82fa835997ac198f1165c04352910d51fb674d69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b8cdf54407cc95783919eba8e92f093bbc09eac5e908c4ba0685629948c71bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b8cdf54407cc95783919eba8e92f093bbc09eac5e908c4ba0685629948c71bd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-06T09:58:27Z\\\",\\\"message\\\":\\\"arting RequestHeaderAuthRequestController\\\\nI1006 09:58:27.156944 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1006 09:58:27.156783 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1006 09:58:27.157213 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-432254888/tls.crt::/tmp/serving-cert-432254888/tls.key\\\\\\\"\\\\nI1006 09:58:27.157407 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1006 09:58:27.157443 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1006 09:58:27.157410 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-432254888/tls.crt::/tmp/serving-cert-432254888/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759744705\\\\\\\\\\\\\\\" (2025-10-06 09:58:25 +0000 UTC to 2025-11-05 09:58:26 +0000 UTC (now=2025-10-06 09:58:27.157390076 +0000 UTC))\\\\\\\"\\\\nI1006 09:58:27.157632 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1006 09:58:27.158158 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759744707\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759744707\\\\\\\\\\\\\\\" (2025-10-06 08:58:26 +0000 UTC to 2026-10-06 08:58:26 +0000 UTC (now=2025-10-06 09:58:27.158090663 +0000 UTC))\\\\\\\"\\\\nI1006 09:58:27.158220 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1006 09:58:27.158270 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1006 09:58:27.158310 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nF1006 09:58:27.158617 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://413822c99ba2866521177cce4f1894aef17c160aebb2662da91824a04827d510\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e37e8c7400f7fd14a5c5390a5985e3d8d277e9c5ca60d5e9e9757d07d1778570\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e37e8c7400f7fd14a5c5390a5985e3d8d277e9c5ca60d5e9e9757d07d1778570\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:34Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:34 crc kubenswrapper[4824]: I1006 09:58:34.069723 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 09:58:34 crc kubenswrapper[4824]: I1006 09:58:34.069808 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 09:58:34 crc kubenswrapper[4824]: E1006 09:58:34.069950 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 06 09:58:34 crc kubenswrapper[4824]: E1006 09:58:34.069997 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 06 09:58:34 crc kubenswrapper[4824]: E1006 09:58:34.069951 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 06 09:58:34 crc kubenswrapper[4824]: E1006 09:58:34.070057 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 06 09:58:34 crc kubenswrapper[4824]: E1006 09:58:34.070075 4824 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 06 09:58:34 crc kubenswrapper[4824]: E1006 09:58:34.070014 4824 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 06 09:58:34 crc kubenswrapper[4824]: E1006 09:58:34.070157 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-06 09:58:38.070130901 +0000 UTC m=+27.434553772 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 06 09:58:34 crc kubenswrapper[4824]: E1006 09:58:34.070185 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-06 09:58:38.070175132 +0000 UTC m=+27.434598013 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 06 09:58:34 crc kubenswrapper[4824]: I1006 09:58:34.096304 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d739433-0a38-42fe-b884-b8433011967e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://464e279c5903770c34d80f31ed231b194d5628b39055c4058908039d10690059\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://efa1b62728466ef7aac0e745230348ce920eb9ec9b07fb2cd5a276150dac13fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fedc40f0217626c56d155fc7e78e0592556fef9464db2aaf09c5facc486aad8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78fac300cbe2ac52f5788306b8fc0011f6ba3a22c3546974e70998f2d394987a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:34Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:34 crc kubenswrapper[4824]: I1006 09:58:34.133358 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:34Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:34 crc kubenswrapper[4824]: I1006 09:58:34.157936 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-szn8q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b65abf9f-4e88-4571-960d-3ca997d9c344\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://664bf42e1324a9c9529f36f08d61e26999b3386bdd1fac77851fb7e7411e8934\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49s5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-szn8q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:34Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:34 crc kubenswrapper[4824]: I1006 09:58:34.190513 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f511ece-25eb-465f-b85c-9e5f3f886c21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6da4c07be4d507a6f20bd8b42be6e6ea4712a956f57ef164036cfded7142a87e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6da4c07be4d507a6f20bd8b42be6e6ea4712a956f57ef164036cfded7142a87e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4vwms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:34Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:34 crc kubenswrapper[4824]: I1006 09:58:34.273475 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 09:58:34 crc kubenswrapper[4824]: I1006 09:58:34.273484 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 09:58:34 crc kubenswrapper[4824]: E1006 09:58:34.273609 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 06 09:58:34 crc kubenswrapper[4824]: I1006 09:58:34.273502 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 09:58:34 crc kubenswrapper[4824]: E1006 09:58:34.273707 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 06 09:58:34 crc kubenswrapper[4824]: E1006 09:58:34.273819 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 06 09:58:34 crc kubenswrapper[4824]: I1006 09:58:34.497812 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" event={"ID":"2f511ece-25eb-465f-b85c-9e5f3f886c21","Type":"ContainerStarted","Data":"43199f86ad4a1f02bc57a926cd96b2c3c0e2ea46fb36a3f495668706be4c8041"} Oct 06 09:58:34 crc kubenswrapper[4824]: I1006 09:58:34.498621 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" event={"ID":"2f511ece-25eb-465f-b85c-9e5f3f886c21","Type":"ContainerStarted","Data":"10bfed3e523b7422ce1e981f98166eca78bd12dd4a7cf0634915e1f4ecc3de2b"} Oct 06 09:58:34 crc kubenswrapper[4824]: I1006 09:58:34.498668 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" event={"ID":"2f511ece-25eb-465f-b85c-9e5f3f886c21","Type":"ContainerStarted","Data":"aa6d59fac987f60a4340ab35d84aa059f2b7085a06362e7207a79702d17eee6d"} Oct 06 09:58:34 crc kubenswrapper[4824]: I1006 09:58:34.498685 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" event={"ID":"2f511ece-25eb-465f-b85c-9e5f3f886c21","Type":"ContainerStarted","Data":"09638fbc9d163ad0b63fa57d24ea5298207ad1f55bab8671be311960be2627a8"} Oct 06 09:58:34 crc kubenswrapper[4824]: I1006 09:58:34.500211 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"ab36ed34d6ca02b59043c46e151019620ce31733209d072c21394a79fe36be4c"} Oct 06 09:58:34 crc kubenswrapper[4824]: I1006 09:58:34.505177 4824 generic.go:334] "Generic (PLEG): container finished" podID="b7eaaa68-d971-4a1c-b6c5-551ded8ff27f" containerID="652ab0a13a2bb421872edda3f7188dce8e296247052550c937f8447fea206756" exitCode=0 Oct 06 09:58:34 crc kubenswrapper[4824]: I1006 09:58:34.505232 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-7c22q" event={"ID":"b7eaaa68-d971-4a1c-b6c5-551ded8ff27f","Type":"ContainerDied","Data":"652ab0a13a2bb421872edda3f7188dce8e296247052550c937f8447fea206756"} Oct 06 09:58:34 crc kubenswrapper[4824]: I1006 09:58:34.520411 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-7c22q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7eaaa68-d971-4a1c-b6c5-551ded8ff27f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53336c7d3c509b4e022708998aedc5dee9fdaf7136ae3ee69b4181394c3a22fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53336c7d3c509b4e022708998aedc5dee9fdaf7136ae3ee69b4181394c3a22fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-7c22q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:34Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:34 crc kubenswrapper[4824]: I1006 09:58:34.534739 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2035278cf272827b39a45aa41ecd5b1e41aafc1d9eb8bccd853660fb6e587159\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff3152fceadf36a55578f587623f37d748f90f2afeda5702a05504b372d399be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:34Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:34 crc kubenswrapper[4824]: I1006 09:58:34.553279 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xcgwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74317f75-9ebc-4327-b88c-598c2241240f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8def09a126f9bb910affad9b7fb32303b960add9fbfd5bb709d59b4aeba7d245\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v6sg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xcgwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:34Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:34 crc kubenswrapper[4824]: I1006 09:58:34.573427 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1140c8c-93f1-4fce-8c06-a69561ff8a6d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5462cb77b7d61aed7d4b4be2fbc6ba1451d267a66afd6dede2f259c15230e7c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gq49w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5512be36013baeca611ee92645a1d339da237b70d63aac43334beafe24513d44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gq49w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-khgzw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:34Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:34 crc kubenswrapper[4824]: I1006 09:58:34.606249 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f9011f6673310dffea3acd7bdc2a56f2af91a87641631490fdc63da610f047a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:34Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:34 crc kubenswrapper[4824]: I1006 09:58:34.626479 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:34Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:34 crc kubenswrapper[4824]: I1006 09:58:34.649576 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:34Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:34 crc kubenswrapper[4824]: I1006 09:58:34.664385 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab36ed34d6ca02b59043c46e151019620ce31733209d072c21394a79fe36be4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:34Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:34 crc kubenswrapper[4824]: I1006 09:58:34.681533 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b665cda3-da4e-482e-8248-023abec382f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9235c742d8c45bf667f32979c2891fece50f4ee1bd0c7cf6a753ed2d2b9ab30e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b2a920dbc506df2426232990e31fbb60d255c00be54941826a0f094af4a92f0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d516f278f362d7c45e7149de82fa835997ac198f1165c04352910d51fb674d69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b8cdf54407cc95783919eba8e92f093bbc09eac5e908c4ba0685629948c71bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b8cdf54407cc95783919eba8e92f093bbc09eac5e908c4ba0685629948c71bd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-06T09:58:27Z\\\",\\\"message\\\":\\\"arting RequestHeaderAuthRequestController\\\\nI1006 09:58:27.156944 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1006 09:58:27.156783 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1006 09:58:27.157213 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-432254888/tls.crt::/tmp/serving-cert-432254888/tls.key\\\\\\\"\\\\nI1006 09:58:27.157407 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1006 09:58:27.157443 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1006 09:58:27.157410 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-432254888/tls.crt::/tmp/serving-cert-432254888/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759744705\\\\\\\\\\\\\\\" (2025-10-06 09:58:25 +0000 UTC to 2025-11-05 09:58:26 +0000 UTC (now=2025-10-06 09:58:27.157390076 +0000 UTC))\\\\\\\"\\\\nI1006 09:58:27.157632 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1006 09:58:27.158158 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759744707\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759744707\\\\\\\\\\\\\\\" (2025-10-06 08:58:26 +0000 UTC to 2026-10-06 08:58:26 +0000 UTC (now=2025-10-06 09:58:27.158090663 +0000 UTC))\\\\\\\"\\\\nI1006 09:58:27.158220 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1006 09:58:27.158270 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1006 09:58:27.158310 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nF1006 09:58:27.158617 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://413822c99ba2866521177cce4f1894aef17c160aebb2662da91824a04827d510\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e37e8c7400f7fd14a5c5390a5985e3d8d277e9c5ca60d5e9e9757d07d1778570\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e37e8c7400f7fd14a5c5390a5985e3d8d277e9c5ca60d5e9e9757d07d1778570\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:34Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:34 crc kubenswrapper[4824]: I1006 09:58:34.696102 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d739433-0a38-42fe-b884-b8433011967e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://464e279c5903770c34d80f31ed231b194d5628b39055c4058908039d10690059\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://efa1b62728466ef7aac0e745230348ce920eb9ec9b07fb2cd5a276150dac13fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fedc40f0217626c56d155fc7e78e0592556fef9464db2aaf09c5facc486aad8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78fac300cbe2ac52f5788306b8fc0011f6ba3a22c3546974e70998f2d394987a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:34Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:34 crc kubenswrapper[4824]: I1006 09:58:34.714098 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:34Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:34 crc kubenswrapper[4824]: I1006 09:58:34.727098 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-szn8q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b65abf9f-4e88-4571-960d-3ca997d9c344\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://664bf42e1324a9c9529f36f08d61e26999b3386bdd1fac77851fb7e7411e8934\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49s5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-szn8q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:34Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:34 crc kubenswrapper[4824]: I1006 09:58:34.748939 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f511ece-25eb-465f-b85c-9e5f3f886c21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6da4c07be4d507a6f20bd8b42be6e6ea4712a956f57ef164036cfded7142a87e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6da4c07be4d507a6f20bd8b42be6e6ea4712a956f57ef164036cfded7142a87e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4vwms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:34Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:34 crc kubenswrapper[4824]: I1006 09:58:34.766352 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-7c22q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7eaaa68-d971-4a1c-b6c5-551ded8ff27f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53336c7d3c509b4e022708998aedc5dee9fdaf7136ae3ee69b4181394c3a22fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53336c7d3c509b4e022708998aedc5dee9fdaf7136ae3ee69b4181394c3a22fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://652ab0a13a2bb421872edda3f7188dce8e296247052550c937f8447fea206756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://652ab0a13a2bb421872edda3f7188dce8e296247052550c937f8447fea206756\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-7c22q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:34Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:34 crc kubenswrapper[4824]: I1006 09:58:34.786102 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2035278cf272827b39a45aa41ecd5b1e41aafc1d9eb8bccd853660fb6e587159\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff3152fceadf36a55578f587623f37d748f90f2afeda5702a05504b372d399be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:34Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:34 crc kubenswrapper[4824]: I1006 09:58:34.846688 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xcgwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74317f75-9ebc-4327-b88c-598c2241240f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8def09a126f9bb910affad9b7fb32303b960add9fbfd5bb709d59b4aeba7d245\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v6sg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xcgwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:34Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:34 crc kubenswrapper[4824]: I1006 09:58:34.865304 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1140c8c-93f1-4fce-8c06-a69561ff8a6d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5462cb77b7d61aed7d4b4be2fbc6ba1451d267a66afd6dede2f259c15230e7c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gq49w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5512be36013baeca611ee92645a1d339da237b70d63aac43334beafe24513d44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gq49w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-khgzw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:34Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:34 crc kubenswrapper[4824]: I1006 09:58:34.910991 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f9011f6673310dffea3acd7bdc2a56f2af91a87641631490fdc63da610f047a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:34Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:34 crc kubenswrapper[4824]: I1006 09:58:34.950969 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:34Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:34 crc kubenswrapper[4824]: I1006 09:58:34.991088 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:34Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:35 crc kubenswrapper[4824]: I1006 09:58:35.028713 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab36ed34d6ca02b59043c46e151019620ce31733209d072c21394a79fe36be4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:35Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:35 crc kubenswrapper[4824]: I1006 09:58:35.071122 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b665cda3-da4e-482e-8248-023abec382f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9235c742d8c45bf667f32979c2891fece50f4ee1bd0c7cf6a753ed2d2b9ab30e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b2a920dbc506df2426232990e31fbb60d255c00be54941826a0f094af4a92f0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d516f278f362d7c45e7149de82fa835997ac198f1165c04352910d51fb674d69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b8cdf54407cc95783919eba8e92f093bbc09eac5e908c4ba0685629948c71bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b8cdf54407cc95783919eba8e92f093bbc09eac5e908c4ba0685629948c71bd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-06T09:58:27Z\\\",\\\"message\\\":\\\"arting RequestHeaderAuthRequestController\\\\nI1006 09:58:27.156944 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1006 09:58:27.156783 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1006 09:58:27.157213 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-432254888/tls.crt::/tmp/serving-cert-432254888/tls.key\\\\\\\"\\\\nI1006 09:58:27.157407 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1006 09:58:27.157443 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1006 09:58:27.157410 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-432254888/tls.crt::/tmp/serving-cert-432254888/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759744705\\\\\\\\\\\\\\\" (2025-10-06 09:58:25 +0000 UTC to 2025-11-05 09:58:26 +0000 UTC (now=2025-10-06 09:58:27.157390076 +0000 UTC))\\\\\\\"\\\\nI1006 09:58:27.157632 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1006 09:58:27.158158 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759744707\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759744707\\\\\\\\\\\\\\\" (2025-10-06 08:58:26 +0000 UTC to 2026-10-06 08:58:26 +0000 UTC (now=2025-10-06 09:58:27.158090663 +0000 UTC))\\\\\\\"\\\\nI1006 09:58:27.158220 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1006 09:58:27.158270 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1006 09:58:27.158310 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nF1006 09:58:27.158617 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://413822c99ba2866521177cce4f1894aef17c160aebb2662da91824a04827d510\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e37e8c7400f7fd14a5c5390a5985e3d8d277e9c5ca60d5e9e9757d07d1778570\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e37e8c7400f7fd14a5c5390a5985e3d8d277e9c5ca60d5e9e9757d07d1778570\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:35Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:35 crc kubenswrapper[4824]: I1006 09:58:35.110619 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d739433-0a38-42fe-b884-b8433011967e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://464e279c5903770c34d80f31ed231b194d5628b39055c4058908039d10690059\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://efa1b62728466ef7aac0e745230348ce920eb9ec9b07fb2cd5a276150dac13fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fedc40f0217626c56d155fc7e78e0592556fef9464db2aaf09c5facc486aad8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78fac300cbe2ac52f5788306b8fc0011f6ba3a22c3546974e70998f2d394987a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:35Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:35 crc kubenswrapper[4824]: I1006 09:58:35.151492 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:35Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:35 crc kubenswrapper[4824]: I1006 09:58:35.192140 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-szn8q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b65abf9f-4e88-4571-960d-3ca997d9c344\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://664bf42e1324a9c9529f36f08d61e26999b3386bdd1fac77851fb7e7411e8934\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49s5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-szn8q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:35Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:35 crc kubenswrapper[4824]: I1006 09:58:35.234395 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f511ece-25eb-465f-b85c-9e5f3f886c21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6da4c07be4d507a6f20bd8b42be6e6ea4712a956f57ef164036cfded7142a87e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6da4c07be4d507a6f20bd8b42be6e6ea4712a956f57ef164036cfded7142a87e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4vwms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:35Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:35 crc kubenswrapper[4824]: I1006 09:58:35.517706 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" event={"ID":"2f511ece-25eb-465f-b85c-9e5f3f886c21","Type":"ContainerStarted","Data":"9d6064ad3bcdafebb4c66e7d91f183cc787e9ce13660545c71a722b3ff04ce75"} Oct 06 09:58:35 crc kubenswrapper[4824]: I1006 09:58:35.517798 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" event={"ID":"2f511ece-25eb-465f-b85c-9e5f3f886c21","Type":"ContainerStarted","Data":"6213164d366243c2139d5971041d6f076cf38fd0cce4c0ca5495111963487ec8"} Oct 06 09:58:35 crc kubenswrapper[4824]: I1006 09:58:35.522394 4824 generic.go:334] "Generic (PLEG): container finished" podID="b7eaaa68-d971-4a1c-b6c5-551ded8ff27f" containerID="3514dec0386bb8c016a7f555d7191ea420d5058b0ee98f97d965dd857c5376e6" exitCode=0 Oct 06 09:58:35 crc kubenswrapper[4824]: I1006 09:58:35.522473 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-7c22q" event={"ID":"b7eaaa68-d971-4a1c-b6c5-551ded8ff27f","Type":"ContainerDied","Data":"3514dec0386bb8c016a7f555d7191ea420d5058b0ee98f97d965dd857c5376e6"} Oct 06 09:58:35 crc kubenswrapper[4824]: I1006 09:58:35.553921 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f511ece-25eb-465f-b85c-9e5f3f886c21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6da4c07be4d507a6f20bd8b42be6e6ea4712a956f57ef164036cfded7142a87e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6da4c07be4d507a6f20bd8b42be6e6ea4712a956f57ef164036cfded7142a87e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4vwms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:35Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:35 crc kubenswrapper[4824]: I1006 09:58:35.578122 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b665cda3-da4e-482e-8248-023abec382f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9235c742d8c45bf667f32979c2891fece50f4ee1bd0c7cf6a753ed2d2b9ab30e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b2a920dbc506df2426232990e31fbb60d255c00be54941826a0f094af4a92f0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d516f278f362d7c45e7149de82fa835997ac198f1165c04352910d51fb674d69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b8cdf54407cc95783919eba8e92f093bbc09eac5e908c4ba0685629948c71bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b8cdf54407cc95783919eba8e92f093bbc09eac5e908c4ba0685629948c71bd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-06T09:58:27Z\\\",\\\"message\\\":\\\"arting RequestHeaderAuthRequestController\\\\nI1006 09:58:27.156944 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1006 09:58:27.156783 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1006 09:58:27.157213 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-432254888/tls.crt::/tmp/serving-cert-432254888/tls.key\\\\\\\"\\\\nI1006 09:58:27.157407 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1006 09:58:27.157443 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1006 09:58:27.157410 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-432254888/tls.crt::/tmp/serving-cert-432254888/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759744705\\\\\\\\\\\\\\\" (2025-10-06 09:58:25 +0000 UTC to 2025-11-05 09:58:26 +0000 UTC (now=2025-10-06 09:58:27.157390076 +0000 UTC))\\\\\\\"\\\\nI1006 09:58:27.157632 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1006 09:58:27.158158 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759744707\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759744707\\\\\\\\\\\\\\\" (2025-10-06 08:58:26 +0000 UTC to 2026-10-06 08:58:26 +0000 UTC (now=2025-10-06 09:58:27.158090663 +0000 UTC))\\\\\\\"\\\\nI1006 09:58:27.158220 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1006 09:58:27.158270 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1006 09:58:27.158310 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nF1006 09:58:27.158617 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://413822c99ba2866521177cce4f1894aef17c160aebb2662da91824a04827d510\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e37e8c7400f7fd14a5c5390a5985e3d8d277e9c5ca60d5e9e9757d07d1778570\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e37e8c7400f7fd14a5c5390a5985e3d8d277e9c5ca60d5e9e9757d07d1778570\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:35Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:35 crc kubenswrapper[4824]: I1006 09:58:35.599451 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d739433-0a38-42fe-b884-b8433011967e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://464e279c5903770c34d80f31ed231b194d5628b39055c4058908039d10690059\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://efa1b62728466ef7aac0e745230348ce920eb9ec9b07fb2cd5a276150dac13fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fedc40f0217626c56d155fc7e78e0592556fef9464db2aaf09c5facc486aad8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78fac300cbe2ac52f5788306b8fc0011f6ba3a22c3546974e70998f2d394987a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:35Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:35 crc kubenswrapper[4824]: I1006 09:58:35.619090 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:35Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:35 crc kubenswrapper[4824]: I1006 09:58:35.647643 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-szn8q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b65abf9f-4e88-4571-960d-3ca997d9c344\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://664bf42e1324a9c9529f36f08d61e26999b3386bdd1fac77851fb7e7411e8934\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49s5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-szn8q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:35Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:35 crc kubenswrapper[4824]: I1006 09:58:35.667894 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-7c22q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7eaaa68-d971-4a1c-b6c5-551ded8ff27f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53336c7d3c509b4e022708998aedc5dee9fdaf7136ae3ee69b4181394c3a22fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53336c7d3c509b4e022708998aedc5dee9fdaf7136ae3ee69b4181394c3a22fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://652ab0a13a2bb421872edda3f7188dce8e296247052550c937f8447fea206756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://652ab0a13a2bb421872edda3f7188dce8e296247052550c937f8447fea206756\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3514dec0386bb8c016a7f555d7191ea420d5058b0ee98f97d965dd857c5376e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3514dec0386bb8c016a7f555d7191ea420d5058b0ee98f97d965dd857c5376e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-7c22q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:35Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:35 crc kubenswrapper[4824]: I1006 09:58:35.687194 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2035278cf272827b39a45aa41ecd5b1e41aafc1d9eb8bccd853660fb6e587159\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff3152fceadf36a55578f587623f37d748f90f2afeda5702a05504b372d399be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:35Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:35 crc kubenswrapper[4824]: I1006 09:58:35.702645 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xcgwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74317f75-9ebc-4327-b88c-598c2241240f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8def09a126f9bb910affad9b7fb32303b960add9fbfd5bb709d59b4aeba7d245\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v6sg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xcgwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:35Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:35 crc kubenswrapper[4824]: I1006 09:58:35.715040 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1140c8c-93f1-4fce-8c06-a69561ff8a6d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5462cb77b7d61aed7d4b4be2fbc6ba1451d267a66afd6dede2f259c15230e7c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gq49w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5512be36013baeca611ee92645a1d339da237b70d63aac43334beafe24513d44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gq49w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-khgzw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:35Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:35 crc kubenswrapper[4824]: I1006 09:58:35.730511 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:35Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:35 crc kubenswrapper[4824]: I1006 09:58:35.744936 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab36ed34d6ca02b59043c46e151019620ce31733209d072c21394a79fe36be4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:35Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:35 crc kubenswrapper[4824]: I1006 09:58:35.769433 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f9011f6673310dffea3acd7bdc2a56f2af91a87641631490fdc63da610f047a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:35Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:35 crc kubenswrapper[4824]: I1006 09:58:35.787487 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:35Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:36 crc kubenswrapper[4824]: I1006 09:58:36.273379 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 09:58:36 crc kubenswrapper[4824]: I1006 09:58:36.273485 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 09:58:36 crc kubenswrapper[4824]: I1006 09:58:36.273561 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 09:58:36 crc kubenswrapper[4824]: E1006 09:58:36.273788 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 06 09:58:36 crc kubenswrapper[4824]: E1006 09:58:36.274346 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 06 09:58:36 crc kubenswrapper[4824]: E1006 09:58:36.274650 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 06 09:58:36 crc kubenswrapper[4824]: I1006 09:58:36.531271 4824 generic.go:334] "Generic (PLEG): container finished" podID="b7eaaa68-d971-4a1c-b6c5-551ded8ff27f" containerID="8b035e0bf34e3429400ac8942124a3e9834231899487d03d15609fe5624747a7" exitCode=0 Oct 06 09:58:36 crc kubenswrapper[4824]: I1006 09:58:36.531351 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-7c22q" event={"ID":"b7eaaa68-d971-4a1c-b6c5-551ded8ff27f","Type":"ContainerDied","Data":"8b035e0bf34e3429400ac8942124a3e9834231899487d03d15609fe5624747a7"} Oct 06 09:58:36 crc kubenswrapper[4824]: I1006 09:58:36.563089 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b665cda3-da4e-482e-8248-023abec382f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9235c742d8c45bf667f32979c2891fece50f4ee1bd0c7cf6a753ed2d2b9ab30e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b2a920dbc506df2426232990e31fbb60d255c00be54941826a0f094af4a92f0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d516f278f362d7c45e7149de82fa835997ac198f1165c04352910d51fb674d69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b8cdf54407cc95783919eba8e92f093bbc09eac5e908c4ba0685629948c71bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b8cdf54407cc95783919eba8e92f093bbc09eac5e908c4ba0685629948c71bd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-06T09:58:27Z\\\",\\\"message\\\":\\\"arting RequestHeaderAuthRequestController\\\\nI1006 09:58:27.156944 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1006 09:58:27.156783 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1006 09:58:27.157213 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-432254888/tls.crt::/tmp/serving-cert-432254888/tls.key\\\\\\\"\\\\nI1006 09:58:27.157407 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1006 09:58:27.157443 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1006 09:58:27.157410 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-432254888/tls.crt::/tmp/serving-cert-432254888/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759744705\\\\\\\\\\\\\\\" (2025-10-06 09:58:25 +0000 UTC to 2025-11-05 09:58:26 +0000 UTC (now=2025-10-06 09:58:27.157390076 +0000 UTC))\\\\\\\"\\\\nI1006 09:58:27.157632 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1006 09:58:27.158158 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759744707\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759744707\\\\\\\\\\\\\\\" (2025-10-06 08:58:26 +0000 UTC to 2026-10-06 08:58:26 +0000 UTC (now=2025-10-06 09:58:27.158090663 +0000 UTC))\\\\\\\"\\\\nI1006 09:58:27.158220 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1006 09:58:27.158270 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1006 09:58:27.158310 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nF1006 09:58:27.158617 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://413822c99ba2866521177cce4f1894aef17c160aebb2662da91824a04827d510\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e37e8c7400f7fd14a5c5390a5985e3d8d277e9c5ca60d5e9e9757d07d1778570\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e37e8c7400f7fd14a5c5390a5985e3d8d277e9c5ca60d5e9e9757d07d1778570\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:36Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:36 crc kubenswrapper[4824]: I1006 09:58:36.583354 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 06 09:58:36 crc kubenswrapper[4824]: I1006 09:58:36.587031 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:36 crc kubenswrapper[4824]: I1006 09:58:36.587112 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:36 crc kubenswrapper[4824]: I1006 09:58:36.587135 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:36 crc kubenswrapper[4824]: I1006 09:58:36.587317 4824 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 06 09:58:36 crc kubenswrapper[4824]: I1006 09:58:36.588245 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d739433-0a38-42fe-b884-b8433011967e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://464e279c5903770c34d80f31ed231b194d5628b39055c4058908039d10690059\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://efa1b62728466ef7aac0e745230348ce920eb9ec9b07fb2cd5a276150dac13fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fedc40f0217626c56d155fc7e78e0592556fef9464db2aaf09c5facc486aad8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78fac300cbe2ac52f5788306b8fc0011f6ba3a22c3546974e70998f2d394987a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:36Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:36 crc kubenswrapper[4824]: I1006 09:58:36.599229 4824 kubelet_node_status.go:115] "Node was previously registered" node="crc" Oct 06 09:58:36 crc kubenswrapper[4824]: I1006 09:58:36.599616 4824 kubelet_node_status.go:79] "Successfully registered node" node="crc" Oct 06 09:58:36 crc kubenswrapper[4824]: I1006 09:58:36.601236 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:36 crc kubenswrapper[4824]: I1006 09:58:36.601301 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:36 crc kubenswrapper[4824]: I1006 09:58:36.601316 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:36 crc kubenswrapper[4824]: I1006 09:58:36.601340 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:36 crc kubenswrapper[4824]: I1006 09:58:36.601354 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:36Z","lastTransitionTime":"2025-10-06T09:58:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:36 crc kubenswrapper[4824]: I1006 09:58:36.611849 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:36Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:36 crc kubenswrapper[4824]: E1006 09:58:36.627724 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T09:58:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T09:58:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T09:58:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T09:58:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"740c2e7e-c94c-411f-b080-eb42c9aecdca\\\",\\\"systemUUID\\\":\\\"77448683-227c-4c6b-a334-eb212350758a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:36Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:36 crc kubenswrapper[4824]: I1006 09:58:36.634597 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:36 crc kubenswrapper[4824]: I1006 09:58:36.634648 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:36 crc kubenswrapper[4824]: I1006 09:58:36.634667 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:36 crc kubenswrapper[4824]: I1006 09:58:36.634745 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:36 crc kubenswrapper[4824]: I1006 09:58:36.634768 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:36Z","lastTransitionTime":"2025-10-06T09:58:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:36 crc kubenswrapper[4824]: I1006 09:58:36.634857 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-szn8q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b65abf9f-4e88-4571-960d-3ca997d9c344\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://664bf42e1324a9c9529f36f08d61e26999b3386bdd1fac77851fb7e7411e8934\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49s5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-szn8q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:36Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:36 crc kubenswrapper[4824]: E1006 09:58:36.656875 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T09:58:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T09:58:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T09:58:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T09:58:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"740c2e7e-c94c-411f-b080-eb42c9aecdca\\\",\\\"systemUUID\\\":\\\"77448683-227c-4c6b-a334-eb212350758a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:36Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:36 crc kubenswrapper[4824]: I1006 09:58:36.663510 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:36 crc kubenswrapper[4824]: I1006 09:58:36.663619 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:36 crc kubenswrapper[4824]: I1006 09:58:36.663645 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:36 crc kubenswrapper[4824]: I1006 09:58:36.663677 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:36 crc kubenswrapper[4824]: I1006 09:58:36.663699 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:36Z","lastTransitionTime":"2025-10-06T09:58:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:36 crc kubenswrapper[4824]: I1006 09:58:36.668522 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f511ece-25eb-465f-b85c-9e5f3f886c21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6da4c07be4d507a6f20bd8b42be6e6ea4712a956f57ef164036cfded7142a87e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6da4c07be4d507a6f20bd8b42be6e6ea4712a956f57ef164036cfded7142a87e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4vwms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:36Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:36 crc kubenswrapper[4824]: E1006 09:58:36.681727 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T09:58:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T09:58:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T09:58:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T09:58:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"740c2e7e-c94c-411f-b080-eb42c9aecdca\\\",\\\"systemUUID\\\":\\\"77448683-227c-4c6b-a334-eb212350758a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:36Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:36 crc kubenswrapper[4824]: I1006 09:58:36.687800 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:36 crc kubenswrapper[4824]: I1006 09:58:36.687885 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:36 crc kubenswrapper[4824]: I1006 09:58:36.687939 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:36 crc kubenswrapper[4824]: I1006 09:58:36.688012 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:36 crc kubenswrapper[4824]: I1006 09:58:36.688041 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:36Z","lastTransitionTime":"2025-10-06T09:58:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:36 crc kubenswrapper[4824]: I1006 09:58:36.691744 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-7c22q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7eaaa68-d971-4a1c-b6c5-551ded8ff27f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53336c7d3c509b4e022708998aedc5dee9fdaf7136ae3ee69b4181394c3a22fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53336c7d3c509b4e022708998aedc5dee9fdaf7136ae3ee69b4181394c3a22fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://652ab0a13a2bb421872edda3f7188dce8e296247052550c937f8447fea206756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://652ab0a13a2bb421872edda3f7188dce8e296247052550c937f8447fea206756\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3514dec0386bb8c016a7f555d7191ea420d5058b0ee98f97d965dd857c5376e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3514dec0386bb8c016a7f555d7191ea420d5058b0ee98f97d965dd857c5376e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b035e0bf34e3429400ac8942124a3e9834231899487d03d15609fe5624747a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b035e0bf34e3429400ac8942124a3e9834231899487d03d15609fe5624747a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-7c22q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:36Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:36 crc kubenswrapper[4824]: E1006 09:58:36.707035 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T09:58:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T09:58:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T09:58:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T09:58:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"740c2e7e-c94c-411f-b080-eb42c9aecdca\\\",\\\"systemUUID\\\":\\\"77448683-227c-4c6b-a334-eb212350758a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:36Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:36 crc kubenswrapper[4824]: I1006 09:58:36.708364 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2035278cf272827b39a45aa41ecd5b1e41aafc1d9eb8bccd853660fb6e587159\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff3152fceadf36a55578f587623f37d748f90f2afeda5702a05504b372d399be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:36Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:36 crc kubenswrapper[4824]: I1006 09:58:36.712577 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:36 crc kubenswrapper[4824]: I1006 09:58:36.712648 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:36 crc kubenswrapper[4824]: I1006 09:58:36.712665 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:36 crc kubenswrapper[4824]: I1006 09:58:36.712690 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:36 crc kubenswrapper[4824]: I1006 09:58:36.712708 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:36Z","lastTransitionTime":"2025-10-06T09:58:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:36 crc kubenswrapper[4824]: I1006 09:58:36.721315 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xcgwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74317f75-9ebc-4327-b88c-598c2241240f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8def09a126f9bb910affad9b7fb32303b960add9fbfd5bb709d59b4aeba7d245\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v6sg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xcgwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:36Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:36 crc kubenswrapper[4824]: E1006 09:58:36.728321 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T09:58:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T09:58:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T09:58:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T09:58:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"740c2e7e-c94c-411f-b080-eb42c9aecdca\\\",\\\"systemUUID\\\":\\\"77448683-227c-4c6b-a334-eb212350758a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:36Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:36 crc kubenswrapper[4824]: E1006 09:58:36.728482 4824 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 06 09:58:36 crc kubenswrapper[4824]: I1006 09:58:36.730431 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:36 crc kubenswrapper[4824]: I1006 09:58:36.730482 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:36 crc kubenswrapper[4824]: I1006 09:58:36.730494 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:36 crc kubenswrapper[4824]: I1006 09:58:36.730514 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:36 crc kubenswrapper[4824]: I1006 09:58:36.730527 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:36Z","lastTransitionTime":"2025-10-06T09:58:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:36 crc kubenswrapper[4824]: I1006 09:58:36.738838 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1140c8c-93f1-4fce-8c06-a69561ff8a6d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5462cb77b7d61aed7d4b4be2fbc6ba1451d267a66afd6dede2f259c15230e7c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gq49w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5512be36013baeca611ee92645a1d339da237b70d63aac43334beafe24513d44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gq49w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-khgzw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:36Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:36 crc kubenswrapper[4824]: I1006 09:58:36.756719 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f9011f6673310dffea3acd7bdc2a56f2af91a87641631490fdc63da610f047a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:36Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:36 crc kubenswrapper[4824]: I1006 09:58:36.774869 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:36Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:36 crc kubenswrapper[4824]: I1006 09:58:36.789609 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:36Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:36 crc kubenswrapper[4824]: I1006 09:58:36.801810 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab36ed34d6ca02b59043c46e151019620ce31733209d072c21394a79fe36be4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:36Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:36 crc kubenswrapper[4824]: I1006 09:58:36.835235 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:36 crc kubenswrapper[4824]: I1006 09:58:36.835282 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:36 crc kubenswrapper[4824]: I1006 09:58:36.835294 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:36 crc kubenswrapper[4824]: I1006 09:58:36.835316 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:36 crc kubenswrapper[4824]: I1006 09:58:36.835335 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:36Z","lastTransitionTime":"2025-10-06T09:58:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:36 crc kubenswrapper[4824]: I1006 09:58:36.939350 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:36 crc kubenswrapper[4824]: I1006 09:58:36.939402 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:36 crc kubenswrapper[4824]: I1006 09:58:36.939413 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:36 crc kubenswrapper[4824]: I1006 09:58:36.939437 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:36 crc kubenswrapper[4824]: I1006 09:58:36.939452 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:36Z","lastTransitionTime":"2025-10-06T09:58:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:37 crc kubenswrapper[4824]: I1006 09:58:37.043650 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:37 crc kubenswrapper[4824]: I1006 09:58:37.043723 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:37 crc kubenswrapper[4824]: I1006 09:58:37.043741 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:37 crc kubenswrapper[4824]: I1006 09:58:37.043769 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:37 crc kubenswrapper[4824]: I1006 09:58:37.043791 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:37Z","lastTransitionTime":"2025-10-06T09:58:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:37 crc kubenswrapper[4824]: I1006 09:58:37.148264 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:37 crc kubenswrapper[4824]: I1006 09:58:37.148341 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:37 crc kubenswrapper[4824]: I1006 09:58:37.148361 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:37 crc kubenswrapper[4824]: I1006 09:58:37.148393 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:37 crc kubenswrapper[4824]: I1006 09:58:37.148414 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:37Z","lastTransitionTime":"2025-10-06T09:58:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:37 crc kubenswrapper[4824]: I1006 09:58:37.251587 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:37 crc kubenswrapper[4824]: I1006 09:58:37.251672 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:37 crc kubenswrapper[4824]: I1006 09:58:37.251698 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:37 crc kubenswrapper[4824]: I1006 09:58:37.251727 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:37 crc kubenswrapper[4824]: I1006 09:58:37.251745 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:37Z","lastTransitionTime":"2025-10-06T09:58:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:37 crc kubenswrapper[4824]: I1006 09:58:37.281270 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 06 09:58:37 crc kubenswrapper[4824]: I1006 09:58:37.282533 4824 scope.go:117] "RemoveContainer" containerID="0b8cdf54407cc95783919eba8e92f093bbc09eac5e908c4ba0685629948c71bd" Oct 06 09:58:37 crc kubenswrapper[4824]: I1006 09:58:37.358117 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:37 crc kubenswrapper[4824]: I1006 09:58:37.358183 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:37 crc kubenswrapper[4824]: I1006 09:58:37.358200 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:37 crc kubenswrapper[4824]: I1006 09:58:37.358228 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:37 crc kubenswrapper[4824]: I1006 09:58:37.358247 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:37Z","lastTransitionTime":"2025-10-06T09:58:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:37 crc kubenswrapper[4824]: I1006 09:58:37.461552 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:37 crc kubenswrapper[4824]: I1006 09:58:37.461636 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:37 crc kubenswrapper[4824]: I1006 09:58:37.461663 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:37 crc kubenswrapper[4824]: I1006 09:58:37.461695 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:37 crc kubenswrapper[4824]: I1006 09:58:37.461723 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:37Z","lastTransitionTime":"2025-10-06T09:58:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:37 crc kubenswrapper[4824]: I1006 09:58:37.542195 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" event={"ID":"2f511ece-25eb-465f-b85c-9e5f3f886c21","Type":"ContainerStarted","Data":"0644c7db28c7d26aa8d69cd5b9fbcff4d4c06b96df113137dea22a72d9fecd85"} Oct 06 09:58:37 crc kubenswrapper[4824]: I1006 09:58:37.546332 4824 generic.go:334] "Generic (PLEG): container finished" podID="b7eaaa68-d971-4a1c-b6c5-551ded8ff27f" containerID="b6743e027a11864ad3f433d110e60451fb36d49791e03b22d437b70f89d056ca" exitCode=0 Oct 06 09:58:37 crc kubenswrapper[4824]: I1006 09:58:37.546409 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-7c22q" event={"ID":"b7eaaa68-d971-4a1c-b6c5-551ded8ff27f","Type":"ContainerDied","Data":"b6743e027a11864ad3f433d110e60451fb36d49791e03b22d437b70f89d056ca"} Oct 06 09:58:37 crc kubenswrapper[4824]: I1006 09:58:37.565109 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:37 crc kubenswrapper[4824]: I1006 09:58:37.565188 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:37 crc kubenswrapper[4824]: I1006 09:58:37.565206 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:37 crc kubenswrapper[4824]: I1006 09:58:37.565233 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:37 crc kubenswrapper[4824]: I1006 09:58:37.565250 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:37Z","lastTransitionTime":"2025-10-06T09:58:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:37 crc kubenswrapper[4824]: I1006 09:58:37.570285 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-7c22q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7eaaa68-d971-4a1c-b6c5-551ded8ff27f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53336c7d3c509b4e022708998aedc5dee9fdaf7136ae3ee69b4181394c3a22fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53336c7d3c509b4e022708998aedc5dee9fdaf7136ae3ee69b4181394c3a22fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://652ab0a13a2bb421872edda3f7188dce8e296247052550c937f8447fea206756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://652ab0a13a2bb421872edda3f7188dce8e296247052550c937f8447fea206756\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3514dec0386bb8c016a7f555d7191ea420d5058b0ee98f97d965dd857c5376e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3514dec0386bb8c016a7f555d7191ea420d5058b0ee98f97d965dd857c5376e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b035e0bf34e3429400ac8942124a3e9834231899487d03d15609fe5624747a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b035e0bf34e3429400ac8942124a3e9834231899487d03d15609fe5624747a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6743e027a11864ad3f433d110e60451fb36d49791e03b22d437b70f89d056ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6743e027a11864ad3f433d110e60451fb36d49791e03b22d437b70f89d056ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-7c22q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:37Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:37 crc kubenswrapper[4824]: I1006 09:58:37.587686 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1140c8c-93f1-4fce-8c06-a69561ff8a6d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5462cb77b7d61aed7d4b4be2fbc6ba1451d267a66afd6dede2f259c15230e7c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gq49w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5512be36013baeca611ee92645a1d339da237b70d63aac43334beafe24513d44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gq49w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-khgzw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:37Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:37 crc kubenswrapper[4824]: I1006 09:58:37.607213 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2035278cf272827b39a45aa41ecd5b1e41aafc1d9eb8bccd853660fb6e587159\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff3152fceadf36a55578f587623f37d748f90f2afeda5702a05504b372d399be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:37Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:37 crc kubenswrapper[4824]: I1006 09:58:37.623282 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xcgwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74317f75-9ebc-4327-b88c-598c2241240f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8def09a126f9bb910affad9b7fb32303b960add9fbfd5bb709d59b4aeba7d245\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v6sg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xcgwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:37Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:37 crc kubenswrapper[4824]: I1006 09:58:37.643243 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f9011f6673310dffea3acd7bdc2a56f2af91a87641631490fdc63da610f047a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:37Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:37 crc kubenswrapper[4824]: I1006 09:58:37.659491 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:37Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:37 crc kubenswrapper[4824]: I1006 09:58:37.669076 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:37 crc kubenswrapper[4824]: I1006 09:58:37.669113 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:37 crc kubenswrapper[4824]: I1006 09:58:37.669125 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:37 crc kubenswrapper[4824]: I1006 09:58:37.669146 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:37 crc kubenswrapper[4824]: I1006 09:58:37.669158 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:37Z","lastTransitionTime":"2025-10-06T09:58:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:37 crc kubenswrapper[4824]: I1006 09:58:37.676679 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:37Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:37 crc kubenswrapper[4824]: I1006 09:58:37.696569 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab36ed34d6ca02b59043c46e151019620ce31733209d072c21394a79fe36be4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:37Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:37 crc kubenswrapper[4824]: I1006 09:58:37.715473 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:37Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:37 crc kubenswrapper[4824]: I1006 09:58:37.734783 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-szn8q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b65abf9f-4e88-4571-960d-3ca997d9c344\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://664bf42e1324a9c9529f36f08d61e26999b3386bdd1fac77851fb7e7411e8934\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49s5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-szn8q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:37Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:37 crc kubenswrapper[4824]: I1006 09:58:37.758124 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f511ece-25eb-465f-b85c-9e5f3f886c21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6da4c07be4d507a6f20bd8b42be6e6ea4712a956f57ef164036cfded7142a87e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6da4c07be4d507a6f20bd8b42be6e6ea4712a956f57ef164036cfded7142a87e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4vwms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:37Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:37 crc kubenswrapper[4824]: I1006 09:58:37.771913 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:37 crc kubenswrapper[4824]: I1006 09:58:37.771964 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:37 crc kubenswrapper[4824]: I1006 09:58:37.772005 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:37 crc kubenswrapper[4824]: I1006 09:58:37.772033 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:37 crc kubenswrapper[4824]: I1006 09:58:37.772053 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:37Z","lastTransitionTime":"2025-10-06T09:58:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:37 crc kubenswrapper[4824]: I1006 09:58:37.778902 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b665cda3-da4e-482e-8248-023abec382f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9235c742d8c45bf667f32979c2891fece50f4ee1bd0c7cf6a753ed2d2b9ab30e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b2a920dbc506df2426232990e31fbb60d255c00be54941826a0f094af4a92f0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d516f278f362d7c45e7149de82fa835997ac198f1165c04352910d51fb674d69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b8cdf54407cc95783919eba8e92f093bbc09eac5e908c4ba0685629948c71bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b8cdf54407cc95783919eba8e92f093bbc09eac5e908c4ba0685629948c71bd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-06T09:58:27Z\\\",\\\"message\\\":\\\"arting RequestHeaderAuthRequestController\\\\nI1006 09:58:27.156944 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1006 09:58:27.156783 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1006 09:58:27.157213 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-432254888/tls.crt::/tmp/serving-cert-432254888/tls.key\\\\\\\"\\\\nI1006 09:58:27.157407 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1006 09:58:27.157443 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1006 09:58:27.157410 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-432254888/tls.crt::/tmp/serving-cert-432254888/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759744705\\\\\\\\\\\\\\\" (2025-10-06 09:58:25 +0000 UTC to 2025-11-05 09:58:26 +0000 UTC (now=2025-10-06 09:58:27.157390076 +0000 UTC))\\\\\\\"\\\\nI1006 09:58:27.157632 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1006 09:58:27.158158 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759744707\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759744707\\\\\\\\\\\\\\\" (2025-10-06 08:58:26 +0000 UTC to 2026-10-06 08:58:26 +0000 UTC (now=2025-10-06 09:58:27.158090663 +0000 UTC))\\\\\\\"\\\\nI1006 09:58:27.158220 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1006 09:58:27.158270 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1006 09:58:27.158310 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nF1006 09:58:27.158617 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://413822c99ba2866521177cce4f1894aef17c160aebb2662da91824a04827d510\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e37e8c7400f7fd14a5c5390a5985e3d8d277e9c5ca60d5e9e9757d07d1778570\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e37e8c7400f7fd14a5c5390a5985e3d8d277e9c5ca60d5e9e9757d07d1778570\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:37Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:37 crc kubenswrapper[4824]: I1006 09:58:37.799734 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d739433-0a38-42fe-b884-b8433011967e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://464e279c5903770c34d80f31ed231b194d5628b39055c4058908039d10690059\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://efa1b62728466ef7aac0e745230348ce920eb9ec9b07fb2cd5a276150dac13fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fedc40f0217626c56d155fc7e78e0592556fef9464db2aaf09c5facc486aad8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78fac300cbe2ac52f5788306b8fc0011f6ba3a22c3546974e70998f2d394987a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:37Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:37 crc kubenswrapper[4824]: I1006 09:58:37.876060 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:37 crc kubenswrapper[4824]: I1006 09:58:37.876115 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:37 crc kubenswrapper[4824]: I1006 09:58:37.876144 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:37 crc kubenswrapper[4824]: I1006 09:58:37.876166 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:37 crc kubenswrapper[4824]: I1006 09:58:37.876181 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:37Z","lastTransitionTime":"2025-10-06T09:58:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:37 crc kubenswrapper[4824]: I1006 09:58:37.979318 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:37 crc kubenswrapper[4824]: I1006 09:58:37.979389 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:37 crc kubenswrapper[4824]: I1006 09:58:37.979410 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:37 crc kubenswrapper[4824]: I1006 09:58:37.979438 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:37 crc kubenswrapper[4824]: I1006 09:58:37.979456 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:37Z","lastTransitionTime":"2025-10-06T09:58:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:38 crc kubenswrapper[4824]: I1006 09:58:38.017599 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 09:58:38 crc kubenswrapper[4824]: E1006 09:58:38.018040 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-06 09:58:46.017962284 +0000 UTC m=+35.382385185 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 09:58:38 crc kubenswrapper[4824]: I1006 09:58:38.019041 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 09:58:38 crc kubenswrapper[4824]: I1006 09:58:38.019172 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 09:58:38 crc kubenswrapper[4824]: E1006 09:58:38.019391 4824 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 06 09:58:38 crc kubenswrapper[4824]: E1006 09:58:38.019503 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-06 09:58:46.019479471 +0000 UTC m=+35.383902372 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 06 09:58:38 crc kubenswrapper[4824]: E1006 09:58:38.019398 4824 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 06 09:58:38 crc kubenswrapper[4824]: E1006 09:58:38.019757 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-06 09:58:46.019710796 +0000 UTC m=+35.384133887 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 06 09:58:38 crc kubenswrapper[4824]: I1006 09:58:38.082830 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:38 crc kubenswrapper[4824]: I1006 09:58:38.083279 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:38 crc kubenswrapper[4824]: I1006 09:58:38.083491 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:38 crc kubenswrapper[4824]: I1006 09:58:38.083686 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:38 crc kubenswrapper[4824]: I1006 09:58:38.083851 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:38Z","lastTransitionTime":"2025-10-06T09:58:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:38 crc kubenswrapper[4824]: I1006 09:58:38.119770 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 09:58:38 crc kubenswrapper[4824]: I1006 09:58:38.120026 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 09:58:38 crc kubenswrapper[4824]: E1006 09:58:38.120121 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 06 09:58:38 crc kubenswrapper[4824]: E1006 09:58:38.120375 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 06 09:58:38 crc kubenswrapper[4824]: E1006 09:58:38.120507 4824 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 06 09:58:38 crc kubenswrapper[4824]: E1006 09:58:38.120198 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 06 09:58:38 crc kubenswrapper[4824]: E1006 09:58:38.120646 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 06 09:58:38 crc kubenswrapper[4824]: E1006 09:58:38.120667 4824 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 06 09:58:38 crc kubenswrapper[4824]: E1006 09:58:38.121338 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-06 09:58:46.120743821 +0000 UTC m=+35.485166712 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 06 09:58:38 crc kubenswrapper[4824]: E1006 09:58:38.121410 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-06 09:58:46.121390467 +0000 UTC m=+35.485813368 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 06 09:58:38 crc kubenswrapper[4824]: I1006 09:58:38.187731 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:38 crc kubenswrapper[4824]: I1006 09:58:38.187796 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:38 crc kubenswrapper[4824]: I1006 09:58:38.187815 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:38 crc kubenswrapper[4824]: I1006 09:58:38.187839 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:38 crc kubenswrapper[4824]: I1006 09:58:38.187856 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:38Z","lastTransitionTime":"2025-10-06T09:58:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:38 crc kubenswrapper[4824]: I1006 09:58:38.274110 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 09:58:38 crc kubenswrapper[4824]: I1006 09:58:38.274223 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 09:58:38 crc kubenswrapper[4824]: I1006 09:58:38.274332 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 09:58:38 crc kubenswrapper[4824]: E1006 09:58:38.274439 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 06 09:58:38 crc kubenswrapper[4824]: E1006 09:58:38.274706 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 06 09:58:38 crc kubenswrapper[4824]: E1006 09:58:38.274888 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 06 09:58:38 crc kubenswrapper[4824]: I1006 09:58:38.290922 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:38 crc kubenswrapper[4824]: I1006 09:58:38.291033 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:38 crc kubenswrapper[4824]: I1006 09:58:38.291060 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:38 crc kubenswrapper[4824]: I1006 09:58:38.291092 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:38 crc kubenswrapper[4824]: I1006 09:58:38.291113 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:38Z","lastTransitionTime":"2025-10-06T09:58:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:38 crc kubenswrapper[4824]: I1006 09:58:38.394573 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:38 crc kubenswrapper[4824]: I1006 09:58:38.394643 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:38 crc kubenswrapper[4824]: I1006 09:58:38.394662 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:38 crc kubenswrapper[4824]: I1006 09:58:38.394692 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:38 crc kubenswrapper[4824]: I1006 09:58:38.394713 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:38Z","lastTransitionTime":"2025-10-06T09:58:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:38 crc kubenswrapper[4824]: I1006 09:58:38.497657 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:38 crc kubenswrapper[4824]: I1006 09:58:38.497718 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:38 crc kubenswrapper[4824]: I1006 09:58:38.497727 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:38 crc kubenswrapper[4824]: I1006 09:58:38.497747 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:38 crc kubenswrapper[4824]: I1006 09:58:38.497762 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:38Z","lastTransitionTime":"2025-10-06T09:58:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:38 crc kubenswrapper[4824]: I1006 09:58:38.558425 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Oct 06 09:58:38 crc kubenswrapper[4824]: I1006 09:58:38.561430 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"c820d8bea16cff0d463ce68d2f2acb72d7a7c73372757ff104e14fcec1025a05"} Oct 06 09:58:38 crc kubenswrapper[4824]: I1006 09:58:38.563145 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 06 09:58:38 crc kubenswrapper[4824]: I1006 09:58:38.569464 4824 generic.go:334] "Generic (PLEG): container finished" podID="b7eaaa68-d971-4a1c-b6c5-551ded8ff27f" containerID="c8ea48b6886fd39a1c5e5b29ca1fffc2bd4552d61dc621eb2fba66fb464cf120" exitCode=0 Oct 06 09:58:38 crc kubenswrapper[4824]: I1006 09:58:38.569549 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-7c22q" event={"ID":"b7eaaa68-d971-4a1c-b6c5-551ded8ff27f","Type":"ContainerDied","Data":"c8ea48b6886fd39a1c5e5b29ca1fffc2bd4552d61dc621eb2fba66fb464cf120"} Oct 06 09:58:38 crc kubenswrapper[4824]: I1006 09:58:38.587131 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b665cda3-da4e-482e-8248-023abec382f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9235c742d8c45bf667f32979c2891fece50f4ee1bd0c7cf6a753ed2d2b9ab30e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b2a920dbc506df2426232990e31fbb60d255c00be54941826a0f094af4a92f0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d516f278f362d7c45e7149de82fa835997ac198f1165c04352910d51fb674d69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c820d8bea16cff0d463ce68d2f2acb72d7a7c73372757ff104e14fcec1025a05\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b8cdf54407cc95783919eba8e92f093bbc09eac5e908c4ba0685629948c71bd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-06T09:58:27Z\\\",\\\"message\\\":\\\"arting RequestHeaderAuthRequestController\\\\nI1006 09:58:27.156944 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1006 09:58:27.156783 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1006 09:58:27.157213 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-432254888/tls.crt::/tmp/serving-cert-432254888/tls.key\\\\\\\"\\\\nI1006 09:58:27.157407 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1006 09:58:27.157443 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1006 09:58:27.157410 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-432254888/tls.crt::/tmp/serving-cert-432254888/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759744705\\\\\\\\\\\\\\\" (2025-10-06 09:58:25 +0000 UTC to 2025-11-05 09:58:26 +0000 UTC (now=2025-10-06 09:58:27.157390076 +0000 UTC))\\\\\\\"\\\\nI1006 09:58:27.157632 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1006 09:58:27.158158 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759744707\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759744707\\\\\\\\\\\\\\\" (2025-10-06 08:58:26 +0000 UTC to 2026-10-06 08:58:26 +0000 UTC (now=2025-10-06 09:58:27.158090663 +0000 UTC))\\\\\\\"\\\\nI1006 09:58:27.158220 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1006 09:58:27.158270 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1006 09:58:27.158310 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nF1006 09:58:27.158617 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://413822c99ba2866521177cce4f1894aef17c160aebb2662da91824a04827d510\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e37e8c7400f7fd14a5c5390a5985e3d8d277e9c5ca60d5e9e9757d07d1778570\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e37e8c7400f7fd14a5c5390a5985e3d8d277e9c5ca60d5e9e9757d07d1778570\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:38Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:38 crc kubenswrapper[4824]: I1006 09:58:38.610038 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:38 crc kubenswrapper[4824]: I1006 09:58:38.610094 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:38 crc kubenswrapper[4824]: I1006 09:58:38.610114 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:38 crc kubenswrapper[4824]: I1006 09:58:38.610144 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:38 crc kubenswrapper[4824]: I1006 09:58:38.610166 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:38Z","lastTransitionTime":"2025-10-06T09:58:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:38 crc kubenswrapper[4824]: I1006 09:58:38.611501 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d739433-0a38-42fe-b884-b8433011967e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://464e279c5903770c34d80f31ed231b194d5628b39055c4058908039d10690059\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://efa1b62728466ef7aac0e745230348ce920eb9ec9b07fb2cd5a276150dac13fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fedc40f0217626c56d155fc7e78e0592556fef9464db2aaf09c5facc486aad8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78fac300cbe2ac52f5788306b8fc0011f6ba3a22c3546974e70998f2d394987a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:38Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:38 crc kubenswrapper[4824]: I1006 09:58:38.626901 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:38Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:38 crc kubenswrapper[4824]: I1006 09:58:38.649420 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-szn8q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b65abf9f-4e88-4571-960d-3ca997d9c344\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://664bf42e1324a9c9529f36f08d61e26999b3386bdd1fac77851fb7e7411e8934\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49s5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-szn8q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:38Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:38 crc kubenswrapper[4824]: I1006 09:58:38.678884 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f511ece-25eb-465f-b85c-9e5f3f886c21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6da4c07be4d507a6f20bd8b42be6e6ea4712a956f57ef164036cfded7142a87e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6da4c07be4d507a6f20bd8b42be6e6ea4712a956f57ef164036cfded7142a87e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4vwms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:38Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:38 crc kubenswrapper[4824]: I1006 09:58:38.699516 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-7c22q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7eaaa68-d971-4a1c-b6c5-551ded8ff27f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53336c7d3c509b4e022708998aedc5dee9fdaf7136ae3ee69b4181394c3a22fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53336c7d3c509b4e022708998aedc5dee9fdaf7136ae3ee69b4181394c3a22fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://652ab0a13a2bb421872edda3f7188dce8e296247052550c937f8447fea206756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://652ab0a13a2bb421872edda3f7188dce8e296247052550c937f8447fea206756\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3514dec0386bb8c016a7f555d7191ea420d5058b0ee98f97d965dd857c5376e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3514dec0386bb8c016a7f555d7191ea420d5058b0ee98f97d965dd857c5376e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b035e0bf34e3429400ac8942124a3e9834231899487d03d15609fe5624747a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b035e0bf34e3429400ac8942124a3e9834231899487d03d15609fe5624747a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6743e027a11864ad3f433d110e60451fb36d49791e03b22d437b70f89d056ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6743e027a11864ad3f433d110e60451fb36d49791e03b22d437b70f89d056ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-7c22q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:38Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:38 crc kubenswrapper[4824]: I1006 09:58:38.719368 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2035278cf272827b39a45aa41ecd5b1e41aafc1d9eb8bccd853660fb6e587159\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff3152fceadf36a55578f587623f37d748f90f2afeda5702a05504b372d399be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:38Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:38 crc kubenswrapper[4824]: I1006 09:58:38.722120 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:38 crc kubenswrapper[4824]: I1006 09:58:38.722182 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:38 crc kubenswrapper[4824]: I1006 09:58:38.722200 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:38 crc kubenswrapper[4824]: I1006 09:58:38.722230 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:38 crc kubenswrapper[4824]: I1006 09:58:38.722244 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:38Z","lastTransitionTime":"2025-10-06T09:58:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:38 crc kubenswrapper[4824]: I1006 09:58:38.739542 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xcgwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74317f75-9ebc-4327-b88c-598c2241240f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8def09a126f9bb910affad9b7fb32303b960add9fbfd5bb709d59b4aeba7d245\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v6sg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xcgwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:38Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:38 crc kubenswrapper[4824]: I1006 09:58:38.761356 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1140c8c-93f1-4fce-8c06-a69561ff8a6d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5462cb77b7d61aed7d4b4be2fbc6ba1451d267a66afd6dede2f259c15230e7c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gq49w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5512be36013baeca611ee92645a1d339da237b70d63aac43334beafe24513d44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gq49w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-khgzw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:38Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:38 crc kubenswrapper[4824]: I1006 09:58:38.783761 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f9011f6673310dffea3acd7bdc2a56f2af91a87641631490fdc63da610f047a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:38Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:38 crc kubenswrapper[4824]: I1006 09:58:38.803304 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:38Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:38 crc kubenswrapper[4824]: I1006 09:58:38.823860 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:38Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:38 crc kubenswrapper[4824]: I1006 09:58:38.825795 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:38 crc kubenswrapper[4824]: I1006 09:58:38.825850 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:38 crc kubenswrapper[4824]: I1006 09:58:38.825867 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:38 crc kubenswrapper[4824]: I1006 09:58:38.825894 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:38 crc kubenswrapper[4824]: I1006 09:58:38.825916 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:38Z","lastTransitionTime":"2025-10-06T09:58:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:38 crc kubenswrapper[4824]: I1006 09:58:38.839334 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab36ed34d6ca02b59043c46e151019620ce31733209d072c21394a79fe36be4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:38Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:38 crc kubenswrapper[4824]: I1006 09:58:38.860482 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-7c22q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7eaaa68-d971-4a1c-b6c5-551ded8ff27f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53336c7d3c509b4e022708998aedc5dee9fdaf7136ae3ee69b4181394c3a22fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53336c7d3c509b4e022708998aedc5dee9fdaf7136ae3ee69b4181394c3a22fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://652ab0a13a2bb421872edda3f7188dce8e296247052550c937f8447fea206756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://652ab0a13a2bb421872edda3f7188dce8e296247052550c937f8447fea206756\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3514dec0386bb8c016a7f555d7191ea420d5058b0ee98f97d965dd857c5376e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3514dec0386bb8c016a7f555d7191ea420d5058b0ee98f97d965dd857c5376e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b035e0bf34e3429400ac8942124a3e9834231899487d03d15609fe5624747a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b035e0bf34e3429400ac8942124a3e9834231899487d03d15609fe5624747a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6743e027a11864ad3f433d110e60451fb36d49791e03b22d437b70f89d056ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6743e027a11864ad3f433d110e60451fb36d49791e03b22d437b70f89d056ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8ea48b6886fd39a1c5e5b29ca1fffc2bd4552d61dc621eb2fba66fb464cf120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8ea48b6886fd39a1c5e5b29ca1fffc2bd4552d61dc621eb2fba66fb464cf120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-7c22q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:38Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:38 crc kubenswrapper[4824]: I1006 09:58:38.875129 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2035278cf272827b39a45aa41ecd5b1e41aafc1d9eb8bccd853660fb6e587159\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff3152fceadf36a55578f587623f37d748f90f2afeda5702a05504b372d399be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:38Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:38 crc kubenswrapper[4824]: I1006 09:58:38.887633 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xcgwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74317f75-9ebc-4327-b88c-598c2241240f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8def09a126f9bb910affad9b7fb32303b960add9fbfd5bb709d59b4aeba7d245\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v6sg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xcgwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:38Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:38 crc kubenswrapper[4824]: I1006 09:58:38.905055 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1140c8c-93f1-4fce-8c06-a69561ff8a6d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5462cb77b7d61aed7d4b4be2fbc6ba1451d267a66afd6dede2f259c15230e7c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gq49w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5512be36013baeca611ee92645a1d339da237b70d63aac43334beafe24513d44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gq49w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-khgzw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:38Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:38 crc kubenswrapper[4824]: I1006 09:58:38.920037 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f9011f6673310dffea3acd7bdc2a56f2af91a87641631490fdc63da610f047a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:38Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:38 crc kubenswrapper[4824]: I1006 09:58:38.928625 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:38 crc kubenswrapper[4824]: I1006 09:58:38.928663 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:38 crc kubenswrapper[4824]: I1006 09:58:38.928676 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:38 crc kubenswrapper[4824]: I1006 09:58:38.928694 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:38 crc kubenswrapper[4824]: I1006 09:58:38.928709 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:38Z","lastTransitionTime":"2025-10-06T09:58:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:38 crc kubenswrapper[4824]: I1006 09:58:38.937333 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:38Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:38 crc kubenswrapper[4824]: I1006 09:58:38.954627 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:38Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:38 crc kubenswrapper[4824]: I1006 09:58:38.971889 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab36ed34d6ca02b59043c46e151019620ce31733209d072c21394a79fe36be4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:38Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:38 crc kubenswrapper[4824]: I1006 09:58:38.993118 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b665cda3-da4e-482e-8248-023abec382f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9235c742d8c45bf667f32979c2891fece50f4ee1bd0c7cf6a753ed2d2b9ab30e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b2a920dbc506df2426232990e31fbb60d255c00be54941826a0f094af4a92f0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d516f278f362d7c45e7149de82fa835997ac198f1165c04352910d51fb674d69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c820d8bea16cff0d463ce68d2f2acb72d7a7c73372757ff104e14fcec1025a05\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b8cdf54407cc95783919eba8e92f093bbc09eac5e908c4ba0685629948c71bd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-06T09:58:27Z\\\",\\\"message\\\":\\\"arting RequestHeaderAuthRequestController\\\\nI1006 09:58:27.156944 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1006 09:58:27.156783 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1006 09:58:27.157213 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-432254888/tls.crt::/tmp/serving-cert-432254888/tls.key\\\\\\\"\\\\nI1006 09:58:27.157407 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1006 09:58:27.157443 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1006 09:58:27.157410 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-432254888/tls.crt::/tmp/serving-cert-432254888/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759744705\\\\\\\\\\\\\\\" (2025-10-06 09:58:25 +0000 UTC to 2025-11-05 09:58:26 +0000 UTC (now=2025-10-06 09:58:27.157390076 +0000 UTC))\\\\\\\"\\\\nI1006 09:58:27.157632 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1006 09:58:27.158158 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759744707\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759744707\\\\\\\\\\\\\\\" (2025-10-06 08:58:26 +0000 UTC to 2026-10-06 08:58:26 +0000 UTC (now=2025-10-06 09:58:27.158090663 +0000 UTC))\\\\\\\"\\\\nI1006 09:58:27.158220 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1006 09:58:27.158270 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1006 09:58:27.158310 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nF1006 09:58:27.158617 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://413822c99ba2866521177cce4f1894aef17c160aebb2662da91824a04827d510\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e37e8c7400f7fd14a5c5390a5985e3d8d277e9c5ca60d5e9e9757d07d1778570\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e37e8c7400f7fd14a5c5390a5985e3d8d277e9c5ca60d5e9e9757d07d1778570\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:38Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:39 crc kubenswrapper[4824]: I1006 09:58:39.013314 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d739433-0a38-42fe-b884-b8433011967e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://464e279c5903770c34d80f31ed231b194d5628b39055c4058908039d10690059\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://efa1b62728466ef7aac0e745230348ce920eb9ec9b07fb2cd5a276150dac13fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fedc40f0217626c56d155fc7e78e0592556fef9464db2aaf09c5facc486aad8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78fac300cbe2ac52f5788306b8fc0011f6ba3a22c3546974e70998f2d394987a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:39Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:39 crc kubenswrapper[4824]: I1006 09:58:39.030901 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:39Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:39 crc kubenswrapper[4824]: I1006 09:58:39.031261 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:39 crc kubenswrapper[4824]: I1006 09:58:39.031311 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:39 crc kubenswrapper[4824]: I1006 09:58:39.031322 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:39 crc kubenswrapper[4824]: I1006 09:58:39.031339 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:39 crc kubenswrapper[4824]: I1006 09:58:39.031352 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:39Z","lastTransitionTime":"2025-10-06T09:58:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:39 crc kubenswrapper[4824]: I1006 09:58:39.049893 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-szn8q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b65abf9f-4e88-4571-960d-3ca997d9c344\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://664bf42e1324a9c9529f36f08d61e26999b3386bdd1fac77851fb7e7411e8934\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49s5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-szn8q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:39Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:39 crc kubenswrapper[4824]: I1006 09:58:39.078467 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f511ece-25eb-465f-b85c-9e5f3f886c21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6da4c07be4d507a6f20bd8b42be6e6ea4712a956f57ef164036cfded7142a87e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6da4c07be4d507a6f20bd8b42be6e6ea4712a956f57ef164036cfded7142a87e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4vwms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:39Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:39 crc kubenswrapper[4824]: I1006 09:58:39.134308 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:39 crc kubenswrapper[4824]: I1006 09:58:39.134377 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:39 crc kubenswrapper[4824]: I1006 09:58:39.134402 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:39 crc kubenswrapper[4824]: I1006 09:58:39.134436 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:39 crc kubenswrapper[4824]: I1006 09:58:39.134458 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:39Z","lastTransitionTime":"2025-10-06T09:58:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:39 crc kubenswrapper[4824]: I1006 09:58:39.238038 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:39 crc kubenswrapper[4824]: I1006 09:58:39.238129 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:39 crc kubenswrapper[4824]: I1006 09:58:39.238147 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:39 crc kubenswrapper[4824]: I1006 09:58:39.238174 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:39 crc kubenswrapper[4824]: I1006 09:58:39.238191 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:39Z","lastTransitionTime":"2025-10-06T09:58:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:39 crc kubenswrapper[4824]: I1006 09:58:39.341682 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:39 crc kubenswrapper[4824]: I1006 09:58:39.341736 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:39 crc kubenswrapper[4824]: I1006 09:58:39.341745 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:39 crc kubenswrapper[4824]: I1006 09:58:39.341763 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:39 crc kubenswrapper[4824]: I1006 09:58:39.341776 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:39Z","lastTransitionTime":"2025-10-06T09:58:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:39 crc kubenswrapper[4824]: I1006 09:58:39.444739 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:39 crc kubenswrapper[4824]: I1006 09:58:39.444816 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:39 crc kubenswrapper[4824]: I1006 09:58:39.444836 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:39 crc kubenswrapper[4824]: I1006 09:58:39.444865 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:39 crc kubenswrapper[4824]: I1006 09:58:39.444885 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:39Z","lastTransitionTime":"2025-10-06T09:58:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:39 crc kubenswrapper[4824]: I1006 09:58:39.547813 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:39 crc kubenswrapper[4824]: I1006 09:58:39.547875 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:39 crc kubenswrapper[4824]: I1006 09:58:39.547892 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:39 crc kubenswrapper[4824]: I1006 09:58:39.547915 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:39 crc kubenswrapper[4824]: I1006 09:58:39.547932 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:39Z","lastTransitionTime":"2025-10-06T09:58:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:39 crc kubenswrapper[4824]: I1006 09:58:39.578440 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" event={"ID":"2f511ece-25eb-465f-b85c-9e5f3f886c21","Type":"ContainerStarted","Data":"d26ec372d0fabf8482f93382dbbd67379654aa4020b4e9831bd214f78c31a6d7"} Oct 06 09:58:39 crc kubenswrapper[4824]: I1006 09:58:39.578818 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" Oct 06 09:58:39 crc kubenswrapper[4824]: I1006 09:58:39.586119 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-7c22q" event={"ID":"b7eaaa68-d971-4a1c-b6c5-551ded8ff27f","Type":"ContainerStarted","Data":"3f5150c8c6cfc65ecdcd0e7bd52199dd60e2472c96874f103b49919c81b5e5f6"} Oct 06 09:58:39 crc kubenswrapper[4824]: I1006 09:58:39.607418 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-7c22q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7eaaa68-d971-4a1c-b6c5-551ded8ff27f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53336c7d3c509b4e022708998aedc5dee9fdaf7136ae3ee69b4181394c3a22fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53336c7d3c509b4e022708998aedc5dee9fdaf7136ae3ee69b4181394c3a22fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://652ab0a13a2bb421872edda3f7188dce8e296247052550c937f8447fea206756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://652ab0a13a2bb421872edda3f7188dce8e296247052550c937f8447fea206756\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3514dec0386bb8c016a7f555d7191ea420d5058b0ee98f97d965dd857c5376e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3514dec0386bb8c016a7f555d7191ea420d5058b0ee98f97d965dd857c5376e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b035e0bf34e3429400ac8942124a3e9834231899487d03d15609fe5624747a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b035e0bf34e3429400ac8942124a3e9834231899487d03d15609fe5624747a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6743e027a11864ad3f433d110e60451fb36d49791e03b22d437b70f89d056ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6743e027a11864ad3f433d110e60451fb36d49791e03b22d437b70f89d056ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8ea48b6886fd39a1c5e5b29ca1fffc2bd4552d61dc621eb2fba66fb464cf120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8ea48b6886fd39a1c5e5b29ca1fffc2bd4552d61dc621eb2fba66fb464cf120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-7c22q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:39Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:39 crc kubenswrapper[4824]: I1006 09:58:39.629891 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2035278cf272827b39a45aa41ecd5b1e41aafc1d9eb8bccd853660fb6e587159\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff3152fceadf36a55578f587623f37d748f90f2afeda5702a05504b372d399be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:39Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:39 crc kubenswrapper[4824]: I1006 09:58:39.632595 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" Oct 06 09:58:39 crc kubenswrapper[4824]: I1006 09:58:39.651929 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:39 crc kubenswrapper[4824]: I1006 09:58:39.652029 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:39 crc kubenswrapper[4824]: I1006 09:58:39.652048 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:39 crc kubenswrapper[4824]: I1006 09:58:39.652079 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:39 crc kubenswrapper[4824]: I1006 09:58:39.652100 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:39Z","lastTransitionTime":"2025-10-06T09:58:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:39 crc kubenswrapper[4824]: I1006 09:58:39.666632 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xcgwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74317f75-9ebc-4327-b88c-598c2241240f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8def09a126f9bb910affad9b7fb32303b960add9fbfd5bb709d59b4aeba7d245\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v6sg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xcgwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:39Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:39 crc kubenswrapper[4824]: I1006 09:58:39.693467 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1140c8c-93f1-4fce-8c06-a69561ff8a6d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5462cb77b7d61aed7d4b4be2fbc6ba1451d267a66afd6dede2f259c15230e7c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gq49w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5512be36013baeca611ee92645a1d339da237b70d63aac43334beafe24513d44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gq49w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-khgzw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:39Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:39 crc kubenswrapper[4824]: I1006 09:58:39.713745 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab36ed34d6ca02b59043c46e151019620ce31733209d072c21394a79fe36be4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:39Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:39 crc kubenswrapper[4824]: I1006 09:58:39.732183 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f9011f6673310dffea3acd7bdc2a56f2af91a87641631490fdc63da610f047a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:39Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:39 crc kubenswrapper[4824]: I1006 09:58:39.747512 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:39Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:39 crc kubenswrapper[4824]: I1006 09:58:39.756894 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:39 crc kubenswrapper[4824]: I1006 09:58:39.756950 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:39 crc kubenswrapper[4824]: I1006 09:58:39.756965 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:39 crc kubenswrapper[4824]: I1006 09:58:39.757008 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:39 crc kubenswrapper[4824]: I1006 09:58:39.757025 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:39Z","lastTransitionTime":"2025-10-06T09:58:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:39 crc kubenswrapper[4824]: I1006 09:58:39.768721 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:39Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:39 crc kubenswrapper[4824]: I1006 09:58:39.789648 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b665cda3-da4e-482e-8248-023abec382f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9235c742d8c45bf667f32979c2891fece50f4ee1bd0c7cf6a753ed2d2b9ab30e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b2a920dbc506df2426232990e31fbb60d255c00be54941826a0f094af4a92f0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d516f278f362d7c45e7149de82fa835997ac198f1165c04352910d51fb674d69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c820d8bea16cff0d463ce68d2f2acb72d7a7c73372757ff104e14fcec1025a05\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b8cdf54407cc95783919eba8e92f093bbc09eac5e908c4ba0685629948c71bd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-06T09:58:27Z\\\",\\\"message\\\":\\\"arting RequestHeaderAuthRequestController\\\\nI1006 09:58:27.156944 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1006 09:58:27.156783 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1006 09:58:27.157213 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-432254888/tls.crt::/tmp/serving-cert-432254888/tls.key\\\\\\\"\\\\nI1006 09:58:27.157407 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1006 09:58:27.157443 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1006 09:58:27.157410 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-432254888/tls.crt::/tmp/serving-cert-432254888/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759744705\\\\\\\\\\\\\\\" (2025-10-06 09:58:25 +0000 UTC to 2025-11-05 09:58:26 +0000 UTC (now=2025-10-06 09:58:27.157390076 +0000 UTC))\\\\\\\"\\\\nI1006 09:58:27.157632 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1006 09:58:27.158158 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759744707\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759744707\\\\\\\\\\\\\\\" (2025-10-06 08:58:26 +0000 UTC to 2026-10-06 08:58:26 +0000 UTC (now=2025-10-06 09:58:27.158090663 +0000 UTC))\\\\\\\"\\\\nI1006 09:58:27.158220 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1006 09:58:27.158270 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1006 09:58:27.158310 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nF1006 09:58:27.158617 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://413822c99ba2866521177cce4f1894aef17c160aebb2662da91824a04827d510\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e37e8c7400f7fd14a5c5390a5985e3d8d277e9c5ca60d5e9e9757d07d1778570\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e37e8c7400f7fd14a5c5390a5985e3d8d277e9c5ca60d5e9e9757d07d1778570\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:39Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:39 crc kubenswrapper[4824]: I1006 09:58:39.802946 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d739433-0a38-42fe-b884-b8433011967e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://464e279c5903770c34d80f31ed231b194d5628b39055c4058908039d10690059\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://efa1b62728466ef7aac0e745230348ce920eb9ec9b07fb2cd5a276150dac13fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fedc40f0217626c56d155fc7e78e0592556fef9464db2aaf09c5facc486aad8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78fac300cbe2ac52f5788306b8fc0011f6ba3a22c3546974e70998f2d394987a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:39Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:39 crc kubenswrapper[4824]: I1006 09:58:39.818847 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:39Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:39 crc kubenswrapper[4824]: I1006 09:58:39.833767 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-szn8q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b65abf9f-4e88-4571-960d-3ca997d9c344\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://664bf42e1324a9c9529f36f08d61e26999b3386bdd1fac77851fb7e7411e8934\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49s5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-szn8q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:39Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:39 crc kubenswrapper[4824]: I1006 09:58:39.857036 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f511ece-25eb-465f-b85c-9e5f3f886c21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10bfed3e523b7422ce1e981f98166eca78bd12dd4a7cf0634915e1f4ecc3de2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43199f86ad4a1f02bc57a926cd96b2c3c0e2ea46fb36a3f495668706be4c8041\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d6064ad3bcdafebb4c66e7d91f183cc787e9ce13660545c71a722b3ff04ce75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6213164d366243c2139d5971041d6f076cf38fd0cce4c0ca5495111963487ec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa6d59fac987f60a4340ab35d84aa059f2b7085a06362e7207a79702d17eee6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09638fbc9d163ad0b63fa57d24ea5298207ad1f55bab8671be311960be2627a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d26ec372d0fabf8482f93382dbbd67379654aa4020b4e9831bd214f78c31a6d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0644c7db28c7d26aa8d69cd5b9fbcff4d4c06b96df113137dea22a72d9fecd85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6da4c07be4d507a6f20bd8b42be6e6ea4712a956f57ef164036cfded7142a87e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6da4c07be4d507a6f20bd8b42be6e6ea4712a956f57ef164036cfded7142a87e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4vwms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:39Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:39 crc kubenswrapper[4824]: I1006 09:58:39.860190 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:39 crc kubenswrapper[4824]: I1006 09:58:39.860221 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:39 crc kubenswrapper[4824]: I1006 09:58:39.860237 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:39 crc kubenswrapper[4824]: I1006 09:58:39.860259 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:39 crc kubenswrapper[4824]: I1006 09:58:39.860272 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:39Z","lastTransitionTime":"2025-10-06T09:58:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:39 crc kubenswrapper[4824]: I1006 09:58:39.868755 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xcgwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74317f75-9ebc-4327-b88c-598c2241240f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8def09a126f9bb910affad9b7fb32303b960add9fbfd5bb709d59b4aeba7d245\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v6sg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xcgwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:39Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:39 crc kubenswrapper[4824]: I1006 09:58:39.881189 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1140c8c-93f1-4fce-8c06-a69561ff8a6d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5462cb77b7d61aed7d4b4be2fbc6ba1451d267a66afd6dede2f259c15230e7c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gq49w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5512be36013baeca611ee92645a1d339da237b70d63aac43334beafe24513d44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gq49w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-khgzw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:39Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:39 crc kubenswrapper[4824]: I1006 09:58:39.898668 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2035278cf272827b39a45aa41ecd5b1e41aafc1d9eb8bccd853660fb6e587159\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff3152fceadf36a55578f587623f37d748f90f2afeda5702a05504b372d399be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:39Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:39 crc kubenswrapper[4824]: I1006 09:58:39.920523 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f9011f6673310dffea3acd7bdc2a56f2af91a87641631490fdc63da610f047a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:39Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:39 crc kubenswrapper[4824]: I1006 09:58:39.939129 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:39Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:39 crc kubenswrapper[4824]: I1006 09:58:39.954323 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:39Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:39 crc kubenswrapper[4824]: I1006 09:58:39.962783 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:39 crc kubenswrapper[4824]: I1006 09:58:39.962856 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:39 crc kubenswrapper[4824]: I1006 09:58:39.962874 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:39 crc kubenswrapper[4824]: I1006 09:58:39.962906 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:39 crc kubenswrapper[4824]: I1006 09:58:39.962925 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:39Z","lastTransitionTime":"2025-10-06T09:58:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:39 crc kubenswrapper[4824]: I1006 09:58:39.971094 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab36ed34d6ca02b59043c46e151019620ce31733209d072c21394a79fe36be4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:39Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:39 crc kubenswrapper[4824]: I1006 09:58:39.988785 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d739433-0a38-42fe-b884-b8433011967e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://464e279c5903770c34d80f31ed231b194d5628b39055c4058908039d10690059\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://efa1b62728466ef7aac0e745230348ce920eb9ec9b07fb2cd5a276150dac13fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fedc40f0217626c56d155fc7e78e0592556fef9464db2aaf09c5facc486aad8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78fac300cbe2ac52f5788306b8fc0011f6ba3a22c3546974e70998f2d394987a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:39Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:40 crc kubenswrapper[4824]: I1006 09:58:40.004723 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:40Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:40 crc kubenswrapper[4824]: I1006 09:58:40.021527 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-szn8q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b65abf9f-4e88-4571-960d-3ca997d9c344\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://664bf42e1324a9c9529f36f08d61e26999b3386bdd1fac77851fb7e7411e8934\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49s5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-szn8q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:40Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:40 crc kubenswrapper[4824]: I1006 09:58:40.053904 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f511ece-25eb-465f-b85c-9e5f3f886c21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10bfed3e523b7422ce1e981f98166eca78bd12dd4a7cf0634915e1f4ecc3de2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43199f86ad4a1f02bc57a926cd96b2c3c0e2ea46fb36a3f495668706be4c8041\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d6064ad3bcdafebb4c66e7d91f183cc787e9ce13660545c71a722b3ff04ce75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6213164d366243c2139d5971041d6f076cf38fd0cce4c0ca5495111963487ec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa6d59fac987f60a4340ab35d84aa059f2b7085a06362e7207a79702d17eee6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09638fbc9d163ad0b63fa57d24ea5298207ad1f55bab8671be311960be2627a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d26ec372d0fabf8482f93382dbbd67379654aa4020b4e9831bd214f78c31a6d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0644c7db28c7d26aa8d69cd5b9fbcff4d4c06b96df113137dea22a72d9fecd85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6da4c07be4d507a6f20bd8b42be6e6ea4712a956f57ef164036cfded7142a87e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6da4c07be4d507a6f20bd8b42be6e6ea4712a956f57ef164036cfded7142a87e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4vwms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:40Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:40 crc kubenswrapper[4824]: I1006 09:58:40.065322 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:40 crc kubenswrapper[4824]: I1006 09:58:40.065385 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:40 crc kubenswrapper[4824]: I1006 09:58:40.065398 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:40 crc kubenswrapper[4824]: I1006 09:58:40.065420 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:40 crc kubenswrapper[4824]: I1006 09:58:40.065433 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:40Z","lastTransitionTime":"2025-10-06T09:58:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:40 crc kubenswrapper[4824]: I1006 09:58:40.072360 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b665cda3-da4e-482e-8248-023abec382f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9235c742d8c45bf667f32979c2891fece50f4ee1bd0c7cf6a753ed2d2b9ab30e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b2a920dbc506df2426232990e31fbb60d255c00be54941826a0f094af4a92f0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d516f278f362d7c45e7149de82fa835997ac198f1165c04352910d51fb674d69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c820d8bea16cff0d463ce68d2f2acb72d7a7c73372757ff104e14fcec1025a05\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b8cdf54407cc95783919eba8e92f093bbc09eac5e908c4ba0685629948c71bd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-06T09:58:27Z\\\",\\\"message\\\":\\\"arting RequestHeaderAuthRequestController\\\\nI1006 09:58:27.156944 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1006 09:58:27.156783 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1006 09:58:27.157213 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-432254888/tls.crt::/tmp/serving-cert-432254888/tls.key\\\\\\\"\\\\nI1006 09:58:27.157407 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1006 09:58:27.157443 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1006 09:58:27.157410 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-432254888/tls.crt::/tmp/serving-cert-432254888/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759744705\\\\\\\\\\\\\\\" (2025-10-06 09:58:25 +0000 UTC to 2025-11-05 09:58:26 +0000 UTC (now=2025-10-06 09:58:27.157390076 +0000 UTC))\\\\\\\"\\\\nI1006 09:58:27.157632 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1006 09:58:27.158158 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759744707\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759744707\\\\\\\\\\\\\\\" (2025-10-06 08:58:26 +0000 UTC to 2026-10-06 08:58:26 +0000 UTC (now=2025-10-06 09:58:27.158090663 +0000 UTC))\\\\\\\"\\\\nI1006 09:58:27.158220 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1006 09:58:27.158270 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1006 09:58:27.158310 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nF1006 09:58:27.158617 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://413822c99ba2866521177cce4f1894aef17c160aebb2662da91824a04827d510\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e37e8c7400f7fd14a5c5390a5985e3d8d277e9c5ca60d5e9e9757d07d1778570\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e37e8c7400f7fd14a5c5390a5985e3d8d277e9c5ca60d5e9e9757d07d1778570\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:40Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:40 crc kubenswrapper[4824]: I1006 09:58:40.095626 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-7c22q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7eaaa68-d971-4a1c-b6c5-551ded8ff27f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f5150c8c6cfc65ecdcd0e7bd52199dd60e2472c96874f103b49919c81b5e5f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53336c7d3c509b4e022708998aedc5dee9fdaf7136ae3ee69b4181394c3a22fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53336c7d3c509b4e022708998aedc5dee9fdaf7136ae3ee69b4181394c3a22fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://652ab0a13a2bb421872edda3f7188dce8e296247052550c937f8447fea206756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://652ab0a13a2bb421872edda3f7188dce8e296247052550c937f8447fea206756\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3514dec0386bb8c016a7f555d7191ea420d5058b0ee98f97d965dd857c5376e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3514dec0386bb8c016a7f555d7191ea420d5058b0ee98f97d965dd857c5376e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b035e0bf34e3429400ac8942124a3e9834231899487d03d15609fe5624747a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b035e0bf34e3429400ac8942124a3e9834231899487d03d15609fe5624747a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6743e027a11864ad3f433d110e60451fb36d49791e03b22d437b70f89d056ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6743e027a11864ad3f433d110e60451fb36d49791e03b22d437b70f89d056ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8ea48b6886fd39a1c5e5b29ca1fffc2bd4552d61dc621eb2fba66fb464cf120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8ea48b6886fd39a1c5e5b29ca1fffc2bd4552d61dc621eb2fba66fb464cf120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-7c22q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:40Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:40 crc kubenswrapper[4824]: I1006 09:58:40.168215 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:40 crc kubenswrapper[4824]: I1006 09:58:40.168339 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:40 crc kubenswrapper[4824]: I1006 09:58:40.168359 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:40 crc kubenswrapper[4824]: I1006 09:58:40.168389 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:40 crc kubenswrapper[4824]: I1006 09:58:40.168408 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:40Z","lastTransitionTime":"2025-10-06T09:58:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:40 crc kubenswrapper[4824]: I1006 09:58:40.270945 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:40 crc kubenswrapper[4824]: I1006 09:58:40.271002 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:40 crc kubenswrapper[4824]: I1006 09:58:40.271013 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:40 crc kubenswrapper[4824]: I1006 09:58:40.271029 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:40 crc kubenswrapper[4824]: I1006 09:58:40.271040 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:40Z","lastTransitionTime":"2025-10-06T09:58:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:40 crc kubenswrapper[4824]: I1006 09:58:40.273228 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 09:58:40 crc kubenswrapper[4824]: I1006 09:58:40.273236 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 09:58:40 crc kubenswrapper[4824]: I1006 09:58:40.273283 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 09:58:40 crc kubenswrapper[4824]: E1006 09:58:40.273348 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 06 09:58:40 crc kubenswrapper[4824]: E1006 09:58:40.273432 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 06 09:58:40 crc kubenswrapper[4824]: E1006 09:58:40.273578 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 06 09:58:40 crc kubenswrapper[4824]: I1006 09:58:40.374534 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:40 crc kubenswrapper[4824]: I1006 09:58:40.374577 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:40 crc kubenswrapper[4824]: I1006 09:58:40.374590 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:40 crc kubenswrapper[4824]: I1006 09:58:40.374610 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:40 crc kubenswrapper[4824]: I1006 09:58:40.374625 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:40Z","lastTransitionTime":"2025-10-06T09:58:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:40 crc kubenswrapper[4824]: I1006 09:58:40.408037 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-5rx6l"] Oct 06 09:58:40 crc kubenswrapper[4824]: I1006 09:58:40.408441 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-5rx6l" Oct 06 09:58:40 crc kubenswrapper[4824]: I1006 09:58:40.411256 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Oct 06 09:58:40 crc kubenswrapper[4824]: I1006 09:58:40.411923 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Oct 06 09:58:40 crc kubenswrapper[4824]: I1006 09:58:40.412824 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Oct 06 09:58:40 crc kubenswrapper[4824]: I1006 09:58:40.414105 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Oct 06 09:58:40 crc kubenswrapper[4824]: I1006 09:58:40.437477 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-7c22q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7eaaa68-d971-4a1c-b6c5-551ded8ff27f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f5150c8c6cfc65ecdcd0e7bd52199dd60e2472c96874f103b49919c81b5e5f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53336c7d3c509b4e022708998aedc5dee9fdaf7136ae3ee69b4181394c3a22fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53336c7d3c509b4e022708998aedc5dee9fdaf7136ae3ee69b4181394c3a22fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://652ab0a13a2bb421872edda3f7188dce8e296247052550c937f8447fea206756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://652ab0a13a2bb421872edda3f7188dce8e296247052550c937f8447fea206756\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3514dec0386bb8c016a7f555d7191ea420d5058b0ee98f97d965dd857c5376e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3514dec0386bb8c016a7f555d7191ea420d5058b0ee98f97d965dd857c5376e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b035e0bf34e3429400ac8942124a3e9834231899487d03d15609fe5624747a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b035e0bf34e3429400ac8942124a3e9834231899487d03d15609fe5624747a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6743e027a11864ad3f433d110e60451fb36d49791e03b22d437b70f89d056ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6743e027a11864ad3f433d110e60451fb36d49791e03b22d437b70f89d056ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8ea48b6886fd39a1c5e5b29ca1fffc2bd4552d61dc621eb2fba66fb464cf120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8ea48b6886fd39a1c5e5b29ca1fffc2bd4552d61dc621eb2fba66fb464cf120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-7c22q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:40Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:40 crc kubenswrapper[4824]: I1006 09:58:40.443409 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e0c320fc-94d9-4d82-81ee-ccccfd125efc-host\") pod \"node-ca-5rx6l\" (UID: \"e0c320fc-94d9-4d82-81ee-ccccfd125efc\") " pod="openshift-image-registry/node-ca-5rx6l" Oct 06 09:58:40 crc kubenswrapper[4824]: I1006 09:58:40.443462 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/e0c320fc-94d9-4d82-81ee-ccccfd125efc-serviceca\") pod \"node-ca-5rx6l\" (UID: \"e0c320fc-94d9-4d82-81ee-ccccfd125efc\") " pod="openshift-image-registry/node-ca-5rx6l" Oct 06 09:58:40 crc kubenswrapper[4824]: I1006 09:58:40.443536 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-njtbn\" (UniqueName: \"kubernetes.io/projected/e0c320fc-94d9-4d82-81ee-ccccfd125efc-kube-api-access-njtbn\") pod \"node-ca-5rx6l\" (UID: \"e0c320fc-94d9-4d82-81ee-ccccfd125efc\") " pod="openshift-image-registry/node-ca-5rx6l" Oct 06 09:58:40 crc kubenswrapper[4824]: I1006 09:58:40.456667 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2035278cf272827b39a45aa41ecd5b1e41aafc1d9eb8bccd853660fb6e587159\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff3152fceadf36a55578f587623f37d748f90f2afeda5702a05504b372d399be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:40Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:40 crc kubenswrapper[4824]: I1006 09:58:40.475226 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xcgwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74317f75-9ebc-4327-b88c-598c2241240f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8def09a126f9bb910affad9b7fb32303b960add9fbfd5bb709d59b4aeba7d245\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v6sg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xcgwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:40Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:40 crc kubenswrapper[4824]: I1006 09:58:40.477919 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:40 crc kubenswrapper[4824]: I1006 09:58:40.478041 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:40 crc kubenswrapper[4824]: I1006 09:58:40.478062 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:40 crc kubenswrapper[4824]: I1006 09:58:40.478087 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:40 crc kubenswrapper[4824]: I1006 09:58:40.478109 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:40Z","lastTransitionTime":"2025-10-06T09:58:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:40 crc kubenswrapper[4824]: I1006 09:58:40.502164 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1140c8c-93f1-4fce-8c06-a69561ff8a6d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5462cb77b7d61aed7d4b4be2fbc6ba1451d267a66afd6dede2f259c15230e7c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gq49w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5512be36013baeca611ee92645a1d339da237b70d63aac43334beafe24513d44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gq49w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-khgzw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:40Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:40 crc kubenswrapper[4824]: I1006 09:58:40.517370 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5rx6l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0c320fc-94d9-4d82-81ee-ccccfd125efc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:40Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:40Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njtbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5rx6l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:40Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:40 crc kubenswrapper[4824]: I1006 09:58:40.538407 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f9011f6673310dffea3acd7bdc2a56f2af91a87641631490fdc63da610f047a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:40Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:40 crc kubenswrapper[4824]: I1006 09:58:40.544714 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e0c320fc-94d9-4d82-81ee-ccccfd125efc-host\") pod \"node-ca-5rx6l\" (UID: \"e0c320fc-94d9-4d82-81ee-ccccfd125efc\") " pod="openshift-image-registry/node-ca-5rx6l" Oct 06 09:58:40 crc kubenswrapper[4824]: I1006 09:58:40.544771 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/e0c320fc-94d9-4d82-81ee-ccccfd125efc-serviceca\") pod \"node-ca-5rx6l\" (UID: \"e0c320fc-94d9-4d82-81ee-ccccfd125efc\") " pod="openshift-image-registry/node-ca-5rx6l" Oct 06 09:58:40 crc kubenswrapper[4824]: I1006 09:58:40.544810 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-njtbn\" (UniqueName: \"kubernetes.io/projected/e0c320fc-94d9-4d82-81ee-ccccfd125efc-kube-api-access-njtbn\") pod \"node-ca-5rx6l\" (UID: \"e0c320fc-94d9-4d82-81ee-ccccfd125efc\") " pod="openshift-image-registry/node-ca-5rx6l" Oct 06 09:58:40 crc kubenswrapper[4824]: I1006 09:58:40.544851 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e0c320fc-94d9-4d82-81ee-ccccfd125efc-host\") pod \"node-ca-5rx6l\" (UID: \"e0c320fc-94d9-4d82-81ee-ccccfd125efc\") " pod="openshift-image-registry/node-ca-5rx6l" Oct 06 09:58:40 crc kubenswrapper[4824]: I1006 09:58:40.545907 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/e0c320fc-94d9-4d82-81ee-ccccfd125efc-serviceca\") pod \"node-ca-5rx6l\" (UID: \"e0c320fc-94d9-4d82-81ee-ccccfd125efc\") " pod="openshift-image-registry/node-ca-5rx6l" Oct 06 09:58:40 crc kubenswrapper[4824]: I1006 09:58:40.558402 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:40Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:40 crc kubenswrapper[4824]: I1006 09:58:40.573142 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-njtbn\" (UniqueName: \"kubernetes.io/projected/e0c320fc-94d9-4d82-81ee-ccccfd125efc-kube-api-access-njtbn\") pod \"node-ca-5rx6l\" (UID: \"e0c320fc-94d9-4d82-81ee-ccccfd125efc\") " pod="openshift-image-registry/node-ca-5rx6l" Oct 06 09:58:40 crc kubenswrapper[4824]: I1006 09:58:40.575117 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:40Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:40 crc kubenswrapper[4824]: I1006 09:58:40.582121 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:40 crc kubenswrapper[4824]: I1006 09:58:40.582198 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:40 crc kubenswrapper[4824]: I1006 09:58:40.582219 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:40 crc kubenswrapper[4824]: I1006 09:58:40.582246 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:40 crc kubenswrapper[4824]: I1006 09:58:40.582266 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:40Z","lastTransitionTime":"2025-10-06T09:58:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:40 crc kubenswrapper[4824]: I1006 09:58:40.589409 4824 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 06 09:58:40 crc kubenswrapper[4824]: I1006 09:58:40.590134 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" Oct 06 09:58:40 crc kubenswrapper[4824]: I1006 09:58:40.597651 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab36ed34d6ca02b59043c46e151019620ce31733209d072c21394a79fe36be4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:40Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:40 crc kubenswrapper[4824]: I1006 09:58:40.621797 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b665cda3-da4e-482e-8248-023abec382f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9235c742d8c45bf667f32979c2891fece50f4ee1bd0c7cf6a753ed2d2b9ab30e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b2a920dbc506df2426232990e31fbb60d255c00be54941826a0f094af4a92f0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d516f278f362d7c45e7149de82fa835997ac198f1165c04352910d51fb674d69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c820d8bea16cff0d463ce68d2f2acb72d7a7c73372757ff104e14fcec1025a05\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b8cdf54407cc95783919eba8e92f093bbc09eac5e908c4ba0685629948c71bd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-06T09:58:27Z\\\",\\\"message\\\":\\\"arting RequestHeaderAuthRequestController\\\\nI1006 09:58:27.156944 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1006 09:58:27.156783 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1006 09:58:27.157213 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-432254888/tls.crt::/tmp/serving-cert-432254888/tls.key\\\\\\\"\\\\nI1006 09:58:27.157407 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1006 09:58:27.157443 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1006 09:58:27.157410 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-432254888/tls.crt::/tmp/serving-cert-432254888/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759744705\\\\\\\\\\\\\\\" (2025-10-06 09:58:25 +0000 UTC to 2025-11-05 09:58:26 +0000 UTC (now=2025-10-06 09:58:27.157390076 +0000 UTC))\\\\\\\"\\\\nI1006 09:58:27.157632 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1006 09:58:27.158158 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759744707\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759744707\\\\\\\\\\\\\\\" (2025-10-06 08:58:26 +0000 UTC to 2026-10-06 08:58:26 +0000 UTC (now=2025-10-06 09:58:27.158090663 +0000 UTC))\\\\\\\"\\\\nI1006 09:58:27.158220 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1006 09:58:27.158270 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1006 09:58:27.158310 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nF1006 09:58:27.158617 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://413822c99ba2866521177cce4f1894aef17c160aebb2662da91824a04827d510\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e37e8c7400f7fd14a5c5390a5985e3d8d277e9c5ca60d5e9e9757d07d1778570\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e37e8c7400f7fd14a5c5390a5985e3d8d277e9c5ca60d5e9e9757d07d1778570\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:40Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:40 crc kubenswrapper[4824]: I1006 09:58:40.625601 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" Oct 06 09:58:40 crc kubenswrapper[4824]: I1006 09:58:40.644566 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d739433-0a38-42fe-b884-b8433011967e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://464e279c5903770c34d80f31ed231b194d5628b39055c4058908039d10690059\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://efa1b62728466ef7aac0e745230348ce920eb9ec9b07fb2cd5a276150dac13fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fedc40f0217626c56d155fc7e78e0592556fef9464db2aaf09c5facc486aad8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78fac300cbe2ac52f5788306b8fc0011f6ba3a22c3546974e70998f2d394987a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:40Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:40 crc kubenswrapper[4824]: I1006 09:58:40.668509 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:40Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:40 crc kubenswrapper[4824]: I1006 09:58:40.685161 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:40 crc kubenswrapper[4824]: I1006 09:58:40.685213 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:40 crc kubenswrapper[4824]: I1006 09:58:40.685224 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:40 crc kubenswrapper[4824]: I1006 09:58:40.685249 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:40 crc kubenswrapper[4824]: I1006 09:58:40.685263 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:40Z","lastTransitionTime":"2025-10-06T09:58:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:40 crc kubenswrapper[4824]: I1006 09:58:40.690314 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-szn8q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b65abf9f-4e88-4571-960d-3ca997d9c344\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://664bf42e1324a9c9529f36f08d61e26999b3386bdd1fac77851fb7e7411e8934\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49s5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-szn8q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:40Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:40 crc kubenswrapper[4824]: I1006 09:58:40.722189 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f511ece-25eb-465f-b85c-9e5f3f886c21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10bfed3e523b7422ce1e981f98166eca78bd12dd4a7cf0634915e1f4ecc3de2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43199f86ad4a1f02bc57a926cd96b2c3c0e2ea46fb36a3f495668706be4c8041\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d6064ad3bcdafebb4c66e7d91f183cc787e9ce13660545c71a722b3ff04ce75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6213164d366243c2139d5971041d6f076cf38fd0cce4c0ca5495111963487ec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa6d59fac987f60a4340ab35d84aa059f2b7085a06362e7207a79702d17eee6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09638fbc9d163ad0b63fa57d24ea5298207ad1f55bab8671be311960be2627a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d26ec372d0fabf8482f93382dbbd67379654aa4020b4e9831bd214f78c31a6d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0644c7db28c7d26aa8d69cd5b9fbcff4d4c06b96df113137dea22a72d9fecd85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6da4c07be4d507a6f20bd8b42be6e6ea4712a956f57ef164036cfded7142a87e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6da4c07be4d507a6f20bd8b42be6e6ea4712a956f57ef164036cfded7142a87e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4vwms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:40Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:40 crc kubenswrapper[4824]: I1006 09:58:40.731351 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-5rx6l" Oct 06 09:58:40 crc kubenswrapper[4824]: W1006 09:58:40.748998 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode0c320fc_94d9_4d82_81ee_ccccfd125efc.slice/crio-ebef85c2cae855cf29b8a4c5c3300efea138c8de0dc8a2abda1cab89aaf250f7 WatchSource:0}: Error finding container ebef85c2cae855cf29b8a4c5c3300efea138c8de0dc8a2abda1cab89aaf250f7: Status 404 returned error can't find the container with id ebef85c2cae855cf29b8a4c5c3300efea138c8de0dc8a2abda1cab89aaf250f7 Oct 06 09:58:40 crc kubenswrapper[4824]: I1006 09:58:40.749504 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b665cda3-da4e-482e-8248-023abec382f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9235c742d8c45bf667f32979c2891fece50f4ee1bd0c7cf6a753ed2d2b9ab30e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b2a920dbc506df2426232990e31fbb60d255c00be54941826a0f094af4a92f0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d516f278f362d7c45e7149de82fa835997ac198f1165c04352910d51fb674d69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c820d8bea16cff0d463ce68d2f2acb72d7a7c73372757ff104e14fcec1025a05\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b8cdf54407cc95783919eba8e92f093bbc09eac5e908c4ba0685629948c71bd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-06T09:58:27Z\\\",\\\"message\\\":\\\"arting RequestHeaderAuthRequestController\\\\nI1006 09:58:27.156944 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1006 09:58:27.156783 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1006 09:58:27.157213 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-432254888/tls.crt::/tmp/serving-cert-432254888/tls.key\\\\\\\"\\\\nI1006 09:58:27.157407 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1006 09:58:27.157443 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1006 09:58:27.157410 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-432254888/tls.crt::/tmp/serving-cert-432254888/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759744705\\\\\\\\\\\\\\\" (2025-10-06 09:58:25 +0000 UTC to 2025-11-05 09:58:26 +0000 UTC (now=2025-10-06 09:58:27.157390076 +0000 UTC))\\\\\\\"\\\\nI1006 09:58:27.157632 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1006 09:58:27.158158 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759744707\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759744707\\\\\\\\\\\\\\\" (2025-10-06 08:58:26 +0000 UTC to 2026-10-06 08:58:26 +0000 UTC (now=2025-10-06 09:58:27.158090663 +0000 UTC))\\\\\\\"\\\\nI1006 09:58:27.158220 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1006 09:58:27.158270 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1006 09:58:27.158310 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nF1006 09:58:27.158617 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://413822c99ba2866521177cce4f1894aef17c160aebb2662da91824a04827d510\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e37e8c7400f7fd14a5c5390a5985e3d8d277e9c5ca60d5e9e9757d07d1778570\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e37e8c7400f7fd14a5c5390a5985e3d8d277e9c5ca60d5e9e9757d07d1778570\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:40Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:40 crc kubenswrapper[4824]: I1006 09:58:40.771768 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d739433-0a38-42fe-b884-b8433011967e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://464e279c5903770c34d80f31ed231b194d5628b39055c4058908039d10690059\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://efa1b62728466ef7aac0e745230348ce920eb9ec9b07fb2cd5a276150dac13fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fedc40f0217626c56d155fc7e78e0592556fef9464db2aaf09c5facc486aad8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78fac300cbe2ac52f5788306b8fc0011f6ba3a22c3546974e70998f2d394987a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:40Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:40 crc kubenswrapper[4824]: I1006 09:58:40.788824 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:40 crc kubenswrapper[4824]: I1006 09:58:40.788863 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:40 crc kubenswrapper[4824]: I1006 09:58:40.788872 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:40 crc kubenswrapper[4824]: I1006 09:58:40.788886 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:40 crc kubenswrapper[4824]: I1006 09:58:40.788895 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:40Z","lastTransitionTime":"2025-10-06T09:58:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:40 crc kubenswrapper[4824]: I1006 09:58:40.793123 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:40Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:40 crc kubenswrapper[4824]: I1006 09:58:40.816605 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-szn8q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b65abf9f-4e88-4571-960d-3ca997d9c344\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://664bf42e1324a9c9529f36f08d61e26999b3386bdd1fac77851fb7e7411e8934\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49s5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-szn8q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:40Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:40 crc kubenswrapper[4824]: I1006 09:58:40.847860 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f511ece-25eb-465f-b85c-9e5f3f886c21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10bfed3e523b7422ce1e981f98166eca78bd12dd4a7cf0634915e1f4ecc3de2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43199f86ad4a1f02bc57a926cd96b2c3c0e2ea46fb36a3f495668706be4c8041\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d6064ad3bcdafebb4c66e7d91f183cc787e9ce13660545c71a722b3ff04ce75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6213164d366243c2139d5971041d6f076cf38fd0cce4c0ca5495111963487ec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa6d59fac987f60a4340ab35d84aa059f2b7085a06362e7207a79702d17eee6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09638fbc9d163ad0b63fa57d24ea5298207ad1f55bab8671be311960be2627a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d26ec372d0fabf8482f93382dbbd67379654aa4020b4e9831bd214f78c31a6d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0644c7db28c7d26aa8d69cd5b9fbcff4d4c06b96df113137dea22a72d9fecd85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6da4c07be4d507a6f20bd8b42be6e6ea4712a956f57ef164036cfded7142a87e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6da4c07be4d507a6f20bd8b42be6e6ea4712a956f57ef164036cfded7142a87e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4vwms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:40Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:40 crc kubenswrapper[4824]: I1006 09:58:40.871468 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-7c22q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7eaaa68-d971-4a1c-b6c5-551ded8ff27f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f5150c8c6cfc65ecdcd0e7bd52199dd60e2472c96874f103b49919c81b5e5f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53336c7d3c509b4e022708998aedc5dee9fdaf7136ae3ee69b4181394c3a22fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53336c7d3c509b4e022708998aedc5dee9fdaf7136ae3ee69b4181394c3a22fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://652ab0a13a2bb421872edda3f7188dce8e296247052550c937f8447fea206756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://652ab0a13a2bb421872edda3f7188dce8e296247052550c937f8447fea206756\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3514dec0386bb8c016a7f555d7191ea420d5058b0ee98f97d965dd857c5376e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3514dec0386bb8c016a7f555d7191ea420d5058b0ee98f97d965dd857c5376e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b035e0bf34e3429400ac8942124a3e9834231899487d03d15609fe5624747a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b035e0bf34e3429400ac8942124a3e9834231899487d03d15609fe5624747a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6743e027a11864ad3f433d110e60451fb36d49791e03b22d437b70f89d056ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6743e027a11864ad3f433d110e60451fb36d49791e03b22d437b70f89d056ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8ea48b6886fd39a1c5e5b29ca1fffc2bd4552d61dc621eb2fba66fb464cf120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8ea48b6886fd39a1c5e5b29ca1fffc2bd4552d61dc621eb2fba66fb464cf120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-7c22q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:40Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:40 crc kubenswrapper[4824]: I1006 09:58:40.889398 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2035278cf272827b39a45aa41ecd5b1e41aafc1d9eb8bccd853660fb6e587159\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff3152fceadf36a55578f587623f37d748f90f2afeda5702a05504b372d399be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:40Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:40 crc kubenswrapper[4824]: I1006 09:58:40.891486 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:40 crc kubenswrapper[4824]: I1006 09:58:40.891544 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:40 crc kubenswrapper[4824]: I1006 09:58:40.891565 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:40 crc kubenswrapper[4824]: I1006 09:58:40.891593 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:40 crc kubenswrapper[4824]: I1006 09:58:40.891616 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:40Z","lastTransitionTime":"2025-10-06T09:58:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:40 crc kubenswrapper[4824]: I1006 09:58:40.910853 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xcgwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74317f75-9ebc-4327-b88c-598c2241240f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8def09a126f9bb910affad9b7fb32303b960add9fbfd5bb709d59b4aeba7d245\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v6sg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xcgwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:40Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:40 crc kubenswrapper[4824]: I1006 09:58:40.926618 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1140c8c-93f1-4fce-8c06-a69561ff8a6d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5462cb77b7d61aed7d4b4be2fbc6ba1451d267a66afd6dede2f259c15230e7c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gq49w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5512be36013baeca611ee92645a1d339da237b70d63aac43334beafe24513d44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gq49w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-khgzw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:40Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:40 crc kubenswrapper[4824]: I1006 09:58:40.941955 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5rx6l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0c320fc-94d9-4d82-81ee-ccccfd125efc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:40Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:40Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njtbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5rx6l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:40Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:40 crc kubenswrapper[4824]: I1006 09:58:40.957916 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f9011f6673310dffea3acd7bdc2a56f2af91a87641631490fdc63da610f047a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:40Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:40 crc kubenswrapper[4824]: I1006 09:58:40.971325 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:40Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:40 crc kubenswrapper[4824]: I1006 09:58:40.985440 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:40Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:40 crc kubenswrapper[4824]: I1006 09:58:40.996282 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:40 crc kubenswrapper[4824]: I1006 09:58:40.996333 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:40 crc kubenswrapper[4824]: I1006 09:58:40.996344 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:40 crc kubenswrapper[4824]: I1006 09:58:40.996365 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:40 crc kubenswrapper[4824]: I1006 09:58:40.996379 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:40Z","lastTransitionTime":"2025-10-06T09:58:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:40 crc kubenswrapper[4824]: I1006 09:58:40.998911 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab36ed34d6ca02b59043c46e151019620ce31733209d072c21394a79fe36be4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:40Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:41 crc kubenswrapper[4824]: I1006 09:58:41.100169 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:41 crc kubenswrapper[4824]: I1006 09:58:41.100246 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:41 crc kubenswrapper[4824]: I1006 09:58:41.100263 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:41 crc kubenswrapper[4824]: I1006 09:58:41.100290 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:41 crc kubenswrapper[4824]: I1006 09:58:41.100308 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:41Z","lastTransitionTime":"2025-10-06T09:58:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:41 crc kubenswrapper[4824]: I1006 09:58:41.206466 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:41 crc kubenswrapper[4824]: I1006 09:58:41.206524 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:41 crc kubenswrapper[4824]: I1006 09:58:41.206541 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:41 crc kubenswrapper[4824]: I1006 09:58:41.206568 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:41 crc kubenswrapper[4824]: I1006 09:58:41.206585 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:41Z","lastTransitionTime":"2025-10-06T09:58:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:41 crc kubenswrapper[4824]: I1006 09:58:41.302246 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab36ed34d6ca02b59043c46e151019620ce31733209d072c21394a79fe36be4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:41Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:41 crc kubenswrapper[4824]: I1006 09:58:41.309845 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:41 crc kubenswrapper[4824]: I1006 09:58:41.309890 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:41 crc kubenswrapper[4824]: I1006 09:58:41.309908 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:41 crc kubenswrapper[4824]: I1006 09:58:41.309932 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:41 crc kubenswrapper[4824]: I1006 09:58:41.309951 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:41Z","lastTransitionTime":"2025-10-06T09:58:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:41 crc kubenswrapper[4824]: I1006 09:58:41.319920 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5rx6l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0c320fc-94d9-4d82-81ee-ccccfd125efc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:40Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:40Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njtbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5rx6l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:41Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:41 crc kubenswrapper[4824]: I1006 09:58:41.340331 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f9011f6673310dffea3acd7bdc2a56f2af91a87641631490fdc63da610f047a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:41Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:41 crc kubenswrapper[4824]: I1006 09:58:41.375205 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:41Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:41 crc kubenswrapper[4824]: I1006 09:58:41.392799 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:41Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:41 crc kubenswrapper[4824]: I1006 09:58:41.407934 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b665cda3-da4e-482e-8248-023abec382f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9235c742d8c45bf667f32979c2891fece50f4ee1bd0c7cf6a753ed2d2b9ab30e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b2a920dbc506df2426232990e31fbb60d255c00be54941826a0f094af4a92f0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d516f278f362d7c45e7149de82fa835997ac198f1165c04352910d51fb674d69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c820d8bea16cff0d463ce68d2f2acb72d7a7c73372757ff104e14fcec1025a05\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b8cdf54407cc95783919eba8e92f093bbc09eac5e908c4ba0685629948c71bd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-06T09:58:27Z\\\",\\\"message\\\":\\\"arting RequestHeaderAuthRequestController\\\\nI1006 09:58:27.156944 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1006 09:58:27.156783 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1006 09:58:27.157213 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-432254888/tls.crt::/tmp/serving-cert-432254888/tls.key\\\\\\\"\\\\nI1006 09:58:27.157407 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1006 09:58:27.157443 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1006 09:58:27.157410 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-432254888/tls.crt::/tmp/serving-cert-432254888/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759744705\\\\\\\\\\\\\\\" (2025-10-06 09:58:25 +0000 UTC to 2025-11-05 09:58:26 +0000 UTC (now=2025-10-06 09:58:27.157390076 +0000 UTC))\\\\\\\"\\\\nI1006 09:58:27.157632 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1006 09:58:27.158158 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759744707\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759744707\\\\\\\\\\\\\\\" (2025-10-06 08:58:26 +0000 UTC to 2026-10-06 08:58:26 +0000 UTC (now=2025-10-06 09:58:27.158090663 +0000 UTC))\\\\\\\"\\\\nI1006 09:58:27.158220 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1006 09:58:27.158270 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1006 09:58:27.158310 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nF1006 09:58:27.158617 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://413822c99ba2866521177cce4f1894aef17c160aebb2662da91824a04827d510\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e37e8c7400f7fd14a5c5390a5985e3d8d277e9c5ca60d5e9e9757d07d1778570\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e37e8c7400f7fd14a5c5390a5985e3d8d277e9c5ca60d5e9e9757d07d1778570\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:41Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:41 crc kubenswrapper[4824]: I1006 09:58:41.414940 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:41 crc kubenswrapper[4824]: I1006 09:58:41.415098 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:41 crc kubenswrapper[4824]: I1006 09:58:41.415309 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:41 crc kubenswrapper[4824]: I1006 09:58:41.415405 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:41 crc kubenswrapper[4824]: I1006 09:58:41.415516 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:41Z","lastTransitionTime":"2025-10-06T09:58:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:41 crc kubenswrapper[4824]: I1006 09:58:41.422657 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d739433-0a38-42fe-b884-b8433011967e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://464e279c5903770c34d80f31ed231b194d5628b39055c4058908039d10690059\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://efa1b62728466ef7aac0e745230348ce920eb9ec9b07fb2cd5a276150dac13fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fedc40f0217626c56d155fc7e78e0592556fef9464db2aaf09c5facc486aad8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78fac300cbe2ac52f5788306b8fc0011f6ba3a22c3546974e70998f2d394987a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:41Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:41 crc kubenswrapper[4824]: I1006 09:58:41.437258 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:41Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:41 crc kubenswrapper[4824]: I1006 09:58:41.453504 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-szn8q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b65abf9f-4e88-4571-960d-3ca997d9c344\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://664bf42e1324a9c9529f36f08d61e26999b3386bdd1fac77851fb7e7411e8934\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49s5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-szn8q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:41Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:41 crc kubenswrapper[4824]: I1006 09:58:41.472774 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f511ece-25eb-465f-b85c-9e5f3f886c21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10bfed3e523b7422ce1e981f98166eca78bd12dd4a7cf0634915e1f4ecc3de2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43199f86ad4a1f02bc57a926cd96b2c3c0e2ea46fb36a3f495668706be4c8041\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d6064ad3bcdafebb4c66e7d91f183cc787e9ce13660545c71a722b3ff04ce75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6213164d366243c2139d5971041d6f076cf38fd0cce4c0ca5495111963487ec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa6d59fac987f60a4340ab35d84aa059f2b7085a06362e7207a79702d17eee6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09638fbc9d163ad0b63fa57d24ea5298207ad1f55bab8671be311960be2627a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d26ec372d0fabf8482f93382dbbd67379654aa4020b4e9831bd214f78c31a6d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0644c7db28c7d26aa8d69cd5b9fbcff4d4c06b96df113137dea22a72d9fecd85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6da4c07be4d507a6f20bd8b42be6e6ea4712a956f57ef164036cfded7142a87e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6da4c07be4d507a6f20bd8b42be6e6ea4712a956f57ef164036cfded7142a87e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4vwms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:41Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:41 crc kubenswrapper[4824]: I1006 09:58:41.489203 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-7c22q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7eaaa68-d971-4a1c-b6c5-551ded8ff27f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f5150c8c6cfc65ecdcd0e7bd52199dd60e2472c96874f103b49919c81b5e5f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53336c7d3c509b4e022708998aedc5dee9fdaf7136ae3ee69b4181394c3a22fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53336c7d3c509b4e022708998aedc5dee9fdaf7136ae3ee69b4181394c3a22fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://652ab0a13a2bb421872edda3f7188dce8e296247052550c937f8447fea206756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://652ab0a13a2bb421872edda3f7188dce8e296247052550c937f8447fea206756\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3514dec0386bb8c016a7f555d7191ea420d5058b0ee98f97d965dd857c5376e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3514dec0386bb8c016a7f555d7191ea420d5058b0ee98f97d965dd857c5376e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b035e0bf34e3429400ac8942124a3e9834231899487d03d15609fe5624747a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b035e0bf34e3429400ac8942124a3e9834231899487d03d15609fe5624747a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6743e027a11864ad3f433d110e60451fb36d49791e03b22d437b70f89d056ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6743e027a11864ad3f433d110e60451fb36d49791e03b22d437b70f89d056ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8ea48b6886fd39a1c5e5b29ca1fffc2bd4552d61dc621eb2fba66fb464cf120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8ea48b6886fd39a1c5e5b29ca1fffc2bd4552d61dc621eb2fba66fb464cf120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-7c22q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:41Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:41 crc kubenswrapper[4824]: I1006 09:58:41.507761 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2035278cf272827b39a45aa41ecd5b1e41aafc1d9eb8bccd853660fb6e587159\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff3152fceadf36a55578f587623f37d748f90f2afeda5702a05504b372d399be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:41Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:41 crc kubenswrapper[4824]: I1006 09:58:41.519086 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:41 crc kubenswrapper[4824]: I1006 09:58:41.519134 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:41 crc kubenswrapper[4824]: I1006 09:58:41.519148 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:41 crc kubenswrapper[4824]: I1006 09:58:41.519171 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:41 crc kubenswrapper[4824]: I1006 09:58:41.519187 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:41Z","lastTransitionTime":"2025-10-06T09:58:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:41 crc kubenswrapper[4824]: I1006 09:58:41.523497 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xcgwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74317f75-9ebc-4327-b88c-598c2241240f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8def09a126f9bb910affad9b7fb32303b960add9fbfd5bb709d59b4aeba7d245\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v6sg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xcgwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:41Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:41 crc kubenswrapper[4824]: I1006 09:58:41.539950 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1140c8c-93f1-4fce-8c06-a69561ff8a6d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5462cb77b7d61aed7d4b4be2fbc6ba1451d267a66afd6dede2f259c15230e7c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gq49w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5512be36013baeca611ee92645a1d339da237b70d63aac43334beafe24513d44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gq49w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-khgzw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:41Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:41 crc kubenswrapper[4824]: I1006 09:58:41.601351 4824 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 06 09:58:41 crc kubenswrapper[4824]: I1006 09:58:41.601478 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-5rx6l" event={"ID":"e0c320fc-94d9-4d82-81ee-ccccfd125efc","Type":"ContainerStarted","Data":"7f8be7bad4e9644494dfb42380c7f5481e1039a1d27ad379c623515948d87392"} Oct 06 09:58:41 crc kubenswrapper[4824]: I1006 09:58:41.602124 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-5rx6l" event={"ID":"e0c320fc-94d9-4d82-81ee-ccccfd125efc","Type":"ContainerStarted","Data":"ebef85c2cae855cf29b8a4c5c3300efea138c8de0dc8a2abda1cab89aaf250f7"} Oct 06 09:58:41 crc kubenswrapper[4824]: I1006 09:58:41.621971 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:41 crc kubenswrapper[4824]: I1006 09:58:41.622075 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:41 crc kubenswrapper[4824]: I1006 09:58:41.622095 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:41 crc kubenswrapper[4824]: I1006 09:58:41.622122 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:41 crc kubenswrapper[4824]: I1006 09:58:41.622141 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:41Z","lastTransitionTime":"2025-10-06T09:58:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:41 crc kubenswrapper[4824]: I1006 09:58:41.623565 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2035278cf272827b39a45aa41ecd5b1e41aafc1d9eb8bccd853660fb6e587159\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff3152fceadf36a55578f587623f37d748f90f2afeda5702a05504b372d399be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:41Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:41 crc kubenswrapper[4824]: I1006 09:58:41.635834 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xcgwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74317f75-9ebc-4327-b88c-598c2241240f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8def09a126f9bb910affad9b7fb32303b960add9fbfd5bb709d59b4aeba7d245\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v6sg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xcgwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:41Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:41 crc kubenswrapper[4824]: I1006 09:58:41.651109 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1140c8c-93f1-4fce-8c06-a69561ff8a6d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5462cb77b7d61aed7d4b4be2fbc6ba1451d267a66afd6dede2f259c15230e7c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gq49w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5512be36013baeca611ee92645a1d339da237b70d63aac43334beafe24513d44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gq49w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-khgzw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:41Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:41 crc kubenswrapper[4824]: I1006 09:58:41.668500 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f9011f6673310dffea3acd7bdc2a56f2af91a87641631490fdc63da610f047a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:41Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:41 crc kubenswrapper[4824]: I1006 09:58:41.693370 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:41Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:41 crc kubenswrapper[4824]: I1006 09:58:41.711910 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:41Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:41 crc kubenswrapper[4824]: I1006 09:58:41.724901 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:41 crc kubenswrapper[4824]: I1006 09:58:41.724936 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:41 crc kubenswrapper[4824]: I1006 09:58:41.724947 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:41 crc kubenswrapper[4824]: I1006 09:58:41.724966 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:41 crc kubenswrapper[4824]: I1006 09:58:41.724995 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:41Z","lastTransitionTime":"2025-10-06T09:58:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:41 crc kubenswrapper[4824]: I1006 09:58:41.729858 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab36ed34d6ca02b59043c46e151019620ce31733209d072c21394a79fe36be4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:41Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:41 crc kubenswrapper[4824]: I1006 09:58:41.744872 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5rx6l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0c320fc-94d9-4d82-81ee-ccccfd125efc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f8be7bad4e9644494dfb42380c7f5481e1039a1d27ad379c623515948d87392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njtbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5rx6l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:41Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:41 crc kubenswrapper[4824]: I1006 09:58:41.765553 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b665cda3-da4e-482e-8248-023abec382f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9235c742d8c45bf667f32979c2891fece50f4ee1bd0c7cf6a753ed2d2b9ab30e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b2a920dbc506df2426232990e31fbb60d255c00be54941826a0f094af4a92f0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d516f278f362d7c45e7149de82fa835997ac198f1165c04352910d51fb674d69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c820d8bea16cff0d463ce68d2f2acb72d7a7c73372757ff104e14fcec1025a05\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b8cdf54407cc95783919eba8e92f093bbc09eac5e908c4ba0685629948c71bd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-06T09:58:27Z\\\",\\\"message\\\":\\\"arting RequestHeaderAuthRequestController\\\\nI1006 09:58:27.156944 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1006 09:58:27.156783 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1006 09:58:27.157213 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-432254888/tls.crt::/tmp/serving-cert-432254888/tls.key\\\\\\\"\\\\nI1006 09:58:27.157407 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1006 09:58:27.157443 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1006 09:58:27.157410 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-432254888/tls.crt::/tmp/serving-cert-432254888/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759744705\\\\\\\\\\\\\\\" (2025-10-06 09:58:25 +0000 UTC to 2025-11-05 09:58:26 +0000 UTC (now=2025-10-06 09:58:27.157390076 +0000 UTC))\\\\\\\"\\\\nI1006 09:58:27.157632 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1006 09:58:27.158158 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759744707\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759744707\\\\\\\\\\\\\\\" (2025-10-06 08:58:26 +0000 UTC to 2026-10-06 08:58:26 +0000 UTC (now=2025-10-06 09:58:27.158090663 +0000 UTC))\\\\\\\"\\\\nI1006 09:58:27.158220 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1006 09:58:27.158270 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1006 09:58:27.158310 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nF1006 09:58:27.158617 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://413822c99ba2866521177cce4f1894aef17c160aebb2662da91824a04827d510\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e37e8c7400f7fd14a5c5390a5985e3d8d277e9c5ca60d5e9e9757d07d1778570\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e37e8c7400f7fd14a5c5390a5985e3d8d277e9c5ca60d5e9e9757d07d1778570\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:41Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:41 crc kubenswrapper[4824]: I1006 09:58:41.783222 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d739433-0a38-42fe-b884-b8433011967e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://464e279c5903770c34d80f31ed231b194d5628b39055c4058908039d10690059\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://efa1b62728466ef7aac0e745230348ce920eb9ec9b07fb2cd5a276150dac13fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fedc40f0217626c56d155fc7e78e0592556fef9464db2aaf09c5facc486aad8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78fac300cbe2ac52f5788306b8fc0011f6ba3a22c3546974e70998f2d394987a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:41Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:41 crc kubenswrapper[4824]: I1006 09:58:41.804145 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:41Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:41 crc kubenswrapper[4824]: I1006 09:58:41.824375 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-szn8q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b65abf9f-4e88-4571-960d-3ca997d9c344\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://664bf42e1324a9c9529f36f08d61e26999b3386bdd1fac77851fb7e7411e8934\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49s5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-szn8q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:41Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:41 crc kubenswrapper[4824]: I1006 09:58:41.827641 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:41 crc kubenswrapper[4824]: I1006 09:58:41.827728 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:41 crc kubenswrapper[4824]: I1006 09:58:41.827741 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:41 crc kubenswrapper[4824]: I1006 09:58:41.827766 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:41 crc kubenswrapper[4824]: I1006 09:58:41.827779 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:41Z","lastTransitionTime":"2025-10-06T09:58:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:41 crc kubenswrapper[4824]: I1006 09:58:41.861614 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f511ece-25eb-465f-b85c-9e5f3f886c21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10bfed3e523b7422ce1e981f98166eca78bd12dd4a7cf0634915e1f4ecc3de2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43199f86ad4a1f02bc57a926cd96b2c3c0e2ea46fb36a3f495668706be4c8041\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d6064ad3bcdafebb4c66e7d91f183cc787e9ce13660545c71a722b3ff04ce75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6213164d366243c2139d5971041d6f076cf38fd0cce4c0ca5495111963487ec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa6d59fac987f60a4340ab35d84aa059f2b7085a06362e7207a79702d17eee6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09638fbc9d163ad0b63fa57d24ea5298207ad1f55bab8671be311960be2627a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d26ec372d0fabf8482f93382dbbd67379654aa4020b4e9831bd214f78c31a6d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0644c7db28c7d26aa8d69cd5b9fbcff4d4c06b96df113137dea22a72d9fecd85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6da4c07be4d507a6f20bd8b42be6e6ea4712a956f57ef164036cfded7142a87e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6da4c07be4d507a6f20bd8b42be6e6ea4712a956f57ef164036cfded7142a87e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4vwms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:41Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:41 crc kubenswrapper[4824]: I1006 09:58:41.887737 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-7c22q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7eaaa68-d971-4a1c-b6c5-551ded8ff27f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f5150c8c6cfc65ecdcd0e7bd52199dd60e2472c96874f103b49919c81b5e5f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53336c7d3c509b4e022708998aedc5dee9fdaf7136ae3ee69b4181394c3a22fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53336c7d3c509b4e022708998aedc5dee9fdaf7136ae3ee69b4181394c3a22fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://652ab0a13a2bb421872edda3f7188dce8e296247052550c937f8447fea206756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://652ab0a13a2bb421872edda3f7188dce8e296247052550c937f8447fea206756\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3514dec0386bb8c016a7f555d7191ea420d5058b0ee98f97d965dd857c5376e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3514dec0386bb8c016a7f555d7191ea420d5058b0ee98f97d965dd857c5376e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b035e0bf34e3429400ac8942124a3e9834231899487d03d15609fe5624747a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b035e0bf34e3429400ac8942124a3e9834231899487d03d15609fe5624747a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6743e027a11864ad3f433d110e60451fb36d49791e03b22d437b70f89d056ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6743e027a11864ad3f433d110e60451fb36d49791e03b22d437b70f89d056ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8ea48b6886fd39a1c5e5b29ca1fffc2bd4552d61dc621eb2fba66fb464cf120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8ea48b6886fd39a1c5e5b29ca1fffc2bd4552d61dc621eb2fba66fb464cf120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-7c22q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:41Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:41 crc kubenswrapper[4824]: I1006 09:58:41.931236 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:41 crc kubenswrapper[4824]: I1006 09:58:41.931628 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:41 crc kubenswrapper[4824]: I1006 09:58:41.931642 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:41 crc kubenswrapper[4824]: I1006 09:58:41.931667 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:41 crc kubenswrapper[4824]: I1006 09:58:41.931682 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:41Z","lastTransitionTime":"2025-10-06T09:58:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:42 crc kubenswrapper[4824]: I1006 09:58:42.034856 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:42 crc kubenswrapper[4824]: I1006 09:58:42.034910 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:42 crc kubenswrapper[4824]: I1006 09:58:42.034922 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:42 crc kubenswrapper[4824]: I1006 09:58:42.034947 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:42 crc kubenswrapper[4824]: I1006 09:58:42.034960 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:42Z","lastTransitionTime":"2025-10-06T09:58:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:42 crc kubenswrapper[4824]: I1006 09:58:42.138425 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:42 crc kubenswrapper[4824]: I1006 09:58:42.138500 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:42 crc kubenswrapper[4824]: I1006 09:58:42.138519 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:42 crc kubenswrapper[4824]: I1006 09:58:42.138544 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:42 crc kubenswrapper[4824]: I1006 09:58:42.138562 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:42Z","lastTransitionTime":"2025-10-06T09:58:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:42 crc kubenswrapper[4824]: I1006 09:58:42.241973 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:42 crc kubenswrapper[4824]: I1006 09:58:42.242078 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:42 crc kubenswrapper[4824]: I1006 09:58:42.242095 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:42 crc kubenswrapper[4824]: I1006 09:58:42.242122 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:42 crc kubenswrapper[4824]: I1006 09:58:42.242139 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:42Z","lastTransitionTime":"2025-10-06T09:58:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:42 crc kubenswrapper[4824]: I1006 09:58:42.273241 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 09:58:42 crc kubenswrapper[4824]: I1006 09:58:42.273250 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 09:58:42 crc kubenswrapper[4824]: E1006 09:58:42.273410 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 06 09:58:42 crc kubenswrapper[4824]: I1006 09:58:42.273261 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 09:58:42 crc kubenswrapper[4824]: E1006 09:58:42.273574 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 06 09:58:42 crc kubenswrapper[4824]: E1006 09:58:42.273666 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 06 09:58:42 crc kubenswrapper[4824]: I1006 09:58:42.345459 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:42 crc kubenswrapper[4824]: I1006 09:58:42.345542 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:42 crc kubenswrapper[4824]: I1006 09:58:42.345565 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:42 crc kubenswrapper[4824]: I1006 09:58:42.345593 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:42 crc kubenswrapper[4824]: I1006 09:58:42.345619 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:42Z","lastTransitionTime":"2025-10-06T09:58:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:42 crc kubenswrapper[4824]: I1006 09:58:42.450130 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:42 crc kubenswrapper[4824]: I1006 09:58:42.450205 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:42 crc kubenswrapper[4824]: I1006 09:58:42.450222 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:42 crc kubenswrapper[4824]: I1006 09:58:42.450252 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:42 crc kubenswrapper[4824]: I1006 09:58:42.450272 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:42Z","lastTransitionTime":"2025-10-06T09:58:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:42 crc kubenswrapper[4824]: I1006 09:58:42.553717 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:42 crc kubenswrapper[4824]: I1006 09:58:42.553820 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:42 crc kubenswrapper[4824]: I1006 09:58:42.553845 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:42 crc kubenswrapper[4824]: I1006 09:58:42.553884 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:42 crc kubenswrapper[4824]: I1006 09:58:42.553911 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:42Z","lastTransitionTime":"2025-10-06T09:58:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:42 crc kubenswrapper[4824]: I1006 09:58:42.608261 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-4vwms_2f511ece-25eb-465f-b85c-9e5f3f886c21/ovnkube-controller/0.log" Oct 06 09:58:42 crc kubenswrapper[4824]: I1006 09:58:42.613044 4824 generic.go:334] "Generic (PLEG): container finished" podID="2f511ece-25eb-465f-b85c-9e5f3f886c21" containerID="d26ec372d0fabf8482f93382dbbd67379654aa4020b4e9831bd214f78c31a6d7" exitCode=1 Oct 06 09:58:42 crc kubenswrapper[4824]: I1006 09:58:42.613147 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" event={"ID":"2f511ece-25eb-465f-b85c-9e5f3f886c21","Type":"ContainerDied","Data":"d26ec372d0fabf8482f93382dbbd67379654aa4020b4e9831bd214f78c31a6d7"} Oct 06 09:58:42 crc kubenswrapper[4824]: I1006 09:58:42.614640 4824 scope.go:117] "RemoveContainer" containerID="d26ec372d0fabf8482f93382dbbd67379654aa4020b4e9831bd214f78c31a6d7" Oct 06 09:58:42 crc kubenswrapper[4824]: I1006 09:58:42.640729 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-7c22q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7eaaa68-d971-4a1c-b6c5-551ded8ff27f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f5150c8c6cfc65ecdcd0e7bd52199dd60e2472c96874f103b49919c81b5e5f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53336c7d3c509b4e022708998aedc5dee9fdaf7136ae3ee69b4181394c3a22fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53336c7d3c509b4e022708998aedc5dee9fdaf7136ae3ee69b4181394c3a22fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://652ab0a13a2bb421872edda3f7188dce8e296247052550c937f8447fea206756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://652ab0a13a2bb421872edda3f7188dce8e296247052550c937f8447fea206756\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3514dec0386bb8c016a7f555d7191ea420d5058b0ee98f97d965dd857c5376e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3514dec0386bb8c016a7f555d7191ea420d5058b0ee98f97d965dd857c5376e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b035e0bf34e3429400ac8942124a3e9834231899487d03d15609fe5624747a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b035e0bf34e3429400ac8942124a3e9834231899487d03d15609fe5624747a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6743e027a11864ad3f433d110e60451fb36d49791e03b22d437b70f89d056ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6743e027a11864ad3f433d110e60451fb36d49791e03b22d437b70f89d056ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8ea48b6886fd39a1c5e5b29ca1fffc2bd4552d61dc621eb2fba66fb464cf120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8ea48b6886fd39a1c5e5b29ca1fffc2bd4552d61dc621eb2fba66fb464cf120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-7c22q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:42Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:42 crc kubenswrapper[4824]: I1006 09:58:42.657851 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:42 crc kubenswrapper[4824]: I1006 09:58:42.658151 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:42 crc kubenswrapper[4824]: I1006 09:58:42.658320 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:42 crc kubenswrapper[4824]: I1006 09:58:42.658448 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:42 crc kubenswrapper[4824]: I1006 09:58:42.658637 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:42Z","lastTransitionTime":"2025-10-06T09:58:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:42 crc kubenswrapper[4824]: I1006 09:58:42.658965 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xcgwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74317f75-9ebc-4327-b88c-598c2241240f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8def09a126f9bb910affad9b7fb32303b960add9fbfd5bb709d59b4aeba7d245\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v6sg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xcgwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:42Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:42 crc kubenswrapper[4824]: I1006 09:58:42.681396 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1140c8c-93f1-4fce-8c06-a69561ff8a6d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5462cb77b7d61aed7d4b4be2fbc6ba1451d267a66afd6dede2f259c15230e7c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gq49w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5512be36013baeca611ee92645a1d339da237b70d63aac43334beafe24513d44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gq49w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-khgzw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:42Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:42 crc kubenswrapper[4824]: I1006 09:58:42.703680 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2035278cf272827b39a45aa41ecd5b1e41aafc1d9eb8bccd853660fb6e587159\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff3152fceadf36a55578f587623f37d748f90f2afeda5702a05504b372d399be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:42Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:42 crc kubenswrapper[4824]: I1006 09:58:42.726029 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f9011f6673310dffea3acd7bdc2a56f2af91a87641631490fdc63da610f047a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:42Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:42 crc kubenswrapper[4824]: I1006 09:58:42.741707 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:42Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:42 crc kubenswrapper[4824]: I1006 09:58:42.761924 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:42 crc kubenswrapper[4824]: I1006 09:58:42.762009 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:42 crc kubenswrapper[4824]: I1006 09:58:42.762033 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:42 crc kubenswrapper[4824]: I1006 09:58:42.762065 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:42 crc kubenswrapper[4824]: I1006 09:58:42.762089 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:42Z","lastTransitionTime":"2025-10-06T09:58:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:42 crc kubenswrapper[4824]: I1006 09:58:42.765943 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:42Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:42 crc kubenswrapper[4824]: I1006 09:58:42.787145 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab36ed34d6ca02b59043c46e151019620ce31733209d072c21394a79fe36be4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:42Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:42 crc kubenswrapper[4824]: I1006 09:58:42.808004 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5rx6l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0c320fc-94d9-4d82-81ee-ccccfd125efc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f8be7bad4e9644494dfb42380c7f5481e1039a1d27ad379c623515948d87392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njtbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5rx6l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:42Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:42 crc kubenswrapper[4824]: I1006 09:58:42.835053 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d739433-0a38-42fe-b884-b8433011967e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://464e279c5903770c34d80f31ed231b194d5628b39055c4058908039d10690059\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://efa1b62728466ef7aac0e745230348ce920eb9ec9b07fb2cd5a276150dac13fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fedc40f0217626c56d155fc7e78e0592556fef9464db2aaf09c5facc486aad8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78fac300cbe2ac52f5788306b8fc0011f6ba3a22c3546974e70998f2d394987a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:42Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:42 crc kubenswrapper[4824]: I1006 09:58:42.854035 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:42Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:42 crc kubenswrapper[4824]: I1006 09:58:42.867102 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:42 crc kubenswrapper[4824]: I1006 09:58:42.867158 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:42 crc kubenswrapper[4824]: I1006 09:58:42.867168 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:42 crc kubenswrapper[4824]: I1006 09:58:42.867193 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:42 crc kubenswrapper[4824]: I1006 09:58:42.867204 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:42Z","lastTransitionTime":"2025-10-06T09:58:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:42 crc kubenswrapper[4824]: I1006 09:58:42.869646 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-szn8q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b65abf9f-4e88-4571-960d-3ca997d9c344\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://664bf42e1324a9c9529f36f08d61e26999b3386bdd1fac77851fb7e7411e8934\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49s5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-szn8q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:42Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:42 crc kubenswrapper[4824]: I1006 09:58:42.893817 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f511ece-25eb-465f-b85c-9e5f3f886c21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10bfed3e523b7422ce1e981f98166eca78bd12dd4a7cf0634915e1f4ecc3de2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43199f86ad4a1f02bc57a926cd96b2c3c0e2ea46fb36a3f495668706be4c8041\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d6064ad3bcdafebb4c66e7d91f183cc787e9ce13660545c71a722b3ff04ce75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6213164d366243c2139d5971041d6f076cf38fd0cce4c0ca5495111963487ec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa6d59fac987f60a4340ab35d84aa059f2b7085a06362e7207a79702d17eee6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09638fbc9d163ad0b63fa57d24ea5298207ad1f55bab8671be311960be2627a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d26ec372d0fabf8482f93382dbbd67379654aa4020b4e9831bd214f78c31a6d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d26ec372d0fabf8482f93382dbbd67379654aa4020b4e9831bd214f78c31a6d7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-06T09:58:42Z\\\",\\\"message\\\":\\\" reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1006 09:58:42.062106 6129 factory.go:656] Stopping watch factory\\\\nI1006 09:58:42.062119 6129 handler.go:208] Removed *v1.Node event handler 7\\\\nI1006 09:58:42.062225 6129 handler.go:208] Removed *v1.Node event handler 2\\\\nI1006 09:58:42.062224 6129 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1006 09:58:42.062363 6129 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1006 09:58:42.062930 6129 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1006 09:58:42.063246 6129 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1006 09:58:42.064079 6129 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1006 09:58:42.064220 6129 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1006 09:58:42.064242 6129 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0644c7db28c7d26aa8d69cd5b9fbcff4d4c06b96df113137dea22a72d9fecd85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6da4c07be4d507a6f20bd8b42be6e6ea4712a956f57ef164036cfded7142a87e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6da4c07be4d507a6f20bd8b42be6e6ea4712a956f57ef164036cfded7142a87e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4vwms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:42Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:42 crc kubenswrapper[4824]: I1006 09:58:42.907752 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b665cda3-da4e-482e-8248-023abec382f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9235c742d8c45bf667f32979c2891fece50f4ee1bd0c7cf6a753ed2d2b9ab30e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b2a920dbc506df2426232990e31fbb60d255c00be54941826a0f094af4a92f0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d516f278f362d7c45e7149de82fa835997ac198f1165c04352910d51fb674d69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c820d8bea16cff0d463ce68d2f2acb72d7a7c73372757ff104e14fcec1025a05\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b8cdf54407cc95783919eba8e92f093bbc09eac5e908c4ba0685629948c71bd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-06T09:58:27Z\\\",\\\"message\\\":\\\"arting RequestHeaderAuthRequestController\\\\nI1006 09:58:27.156944 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1006 09:58:27.156783 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1006 09:58:27.157213 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-432254888/tls.crt::/tmp/serving-cert-432254888/tls.key\\\\\\\"\\\\nI1006 09:58:27.157407 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1006 09:58:27.157443 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1006 09:58:27.157410 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-432254888/tls.crt::/tmp/serving-cert-432254888/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759744705\\\\\\\\\\\\\\\" (2025-10-06 09:58:25 +0000 UTC to 2025-11-05 09:58:26 +0000 UTC (now=2025-10-06 09:58:27.157390076 +0000 UTC))\\\\\\\"\\\\nI1006 09:58:27.157632 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1006 09:58:27.158158 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759744707\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759744707\\\\\\\\\\\\\\\" (2025-10-06 08:58:26 +0000 UTC to 2026-10-06 08:58:26 +0000 UTC (now=2025-10-06 09:58:27.158090663 +0000 UTC))\\\\\\\"\\\\nI1006 09:58:27.158220 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1006 09:58:27.158270 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1006 09:58:27.158310 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nF1006 09:58:27.158617 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://413822c99ba2866521177cce4f1894aef17c160aebb2662da91824a04827d510\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e37e8c7400f7fd14a5c5390a5985e3d8d277e9c5ca60d5e9e9757d07d1778570\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e37e8c7400f7fd14a5c5390a5985e3d8d277e9c5ca60d5e9e9757d07d1778570\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:42Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:42 crc kubenswrapper[4824]: I1006 09:58:42.971847 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:42 crc kubenswrapper[4824]: I1006 09:58:42.971904 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:42 crc kubenswrapper[4824]: I1006 09:58:42.971915 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:42 crc kubenswrapper[4824]: I1006 09:58:42.971932 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:42 crc kubenswrapper[4824]: I1006 09:58:42.971943 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:42Z","lastTransitionTime":"2025-10-06T09:58:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:43 crc kubenswrapper[4824]: I1006 09:58:43.075723 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:43 crc kubenswrapper[4824]: I1006 09:58:43.075772 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:43 crc kubenswrapper[4824]: I1006 09:58:43.075783 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:43 crc kubenswrapper[4824]: I1006 09:58:43.075804 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:43 crc kubenswrapper[4824]: I1006 09:58:43.075817 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:43Z","lastTransitionTime":"2025-10-06T09:58:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:43 crc kubenswrapper[4824]: I1006 09:58:43.178765 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:43 crc kubenswrapper[4824]: I1006 09:58:43.178813 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:43 crc kubenswrapper[4824]: I1006 09:58:43.178823 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:43 crc kubenswrapper[4824]: I1006 09:58:43.178841 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:43 crc kubenswrapper[4824]: I1006 09:58:43.178855 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:43Z","lastTransitionTime":"2025-10-06T09:58:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:43 crc kubenswrapper[4824]: I1006 09:58:43.281098 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:43 crc kubenswrapper[4824]: I1006 09:58:43.281498 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:43 crc kubenswrapper[4824]: I1006 09:58:43.281566 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:43 crc kubenswrapper[4824]: I1006 09:58:43.281649 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:43 crc kubenswrapper[4824]: I1006 09:58:43.281706 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:43Z","lastTransitionTime":"2025-10-06T09:58:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:43 crc kubenswrapper[4824]: I1006 09:58:43.384912 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:43 crc kubenswrapper[4824]: I1006 09:58:43.385037 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:43 crc kubenswrapper[4824]: I1006 09:58:43.385250 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:43 crc kubenswrapper[4824]: I1006 09:58:43.385277 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:43 crc kubenswrapper[4824]: I1006 09:58:43.385296 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:43Z","lastTransitionTime":"2025-10-06T09:58:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:43 crc kubenswrapper[4824]: I1006 09:58:43.488228 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:43 crc kubenswrapper[4824]: I1006 09:58:43.488301 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:43 crc kubenswrapper[4824]: I1006 09:58:43.488338 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:43 crc kubenswrapper[4824]: I1006 09:58:43.488367 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:43 crc kubenswrapper[4824]: I1006 09:58:43.488386 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:43Z","lastTransitionTime":"2025-10-06T09:58:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:43 crc kubenswrapper[4824]: I1006 09:58:43.592141 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:43 crc kubenswrapper[4824]: I1006 09:58:43.592230 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:43 crc kubenswrapper[4824]: I1006 09:58:43.592250 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:43 crc kubenswrapper[4824]: I1006 09:58:43.592282 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:43 crc kubenswrapper[4824]: I1006 09:58:43.592307 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:43Z","lastTransitionTime":"2025-10-06T09:58:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:43 crc kubenswrapper[4824]: I1006 09:58:43.621448 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-4vwms_2f511ece-25eb-465f-b85c-9e5f3f886c21/ovnkube-controller/0.log" Oct 06 09:58:43 crc kubenswrapper[4824]: I1006 09:58:43.626530 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" event={"ID":"2f511ece-25eb-465f-b85c-9e5f3f886c21","Type":"ContainerStarted","Data":"91a256538722530a7679c5bededd29cff892d03964bd6c71f81d33989b0efdd9"} Oct 06 09:58:43 crc kubenswrapper[4824]: I1006 09:58:43.626711 4824 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 06 09:58:43 crc kubenswrapper[4824]: I1006 09:58:43.651646 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2035278cf272827b39a45aa41ecd5b1e41aafc1d9eb8bccd853660fb6e587159\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff3152fceadf36a55578f587623f37d748f90f2afeda5702a05504b372d399be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:43Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:43 crc kubenswrapper[4824]: I1006 09:58:43.669157 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xcgwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74317f75-9ebc-4327-b88c-598c2241240f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8def09a126f9bb910affad9b7fb32303b960add9fbfd5bb709d59b4aeba7d245\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v6sg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xcgwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:43Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:43 crc kubenswrapper[4824]: I1006 09:58:43.691614 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1140c8c-93f1-4fce-8c06-a69561ff8a6d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5462cb77b7d61aed7d4b4be2fbc6ba1451d267a66afd6dede2f259c15230e7c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gq49w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5512be36013baeca611ee92645a1d339da237b70d63aac43334beafe24513d44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gq49w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-khgzw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:43Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:43 crc kubenswrapper[4824]: I1006 09:58:43.698369 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:43 crc kubenswrapper[4824]: I1006 09:58:43.698459 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:43 crc kubenswrapper[4824]: I1006 09:58:43.698488 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:43 crc kubenswrapper[4824]: I1006 09:58:43.698524 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:43 crc kubenswrapper[4824]: I1006 09:58:43.698550 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:43Z","lastTransitionTime":"2025-10-06T09:58:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:43 crc kubenswrapper[4824]: I1006 09:58:43.715969 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:43Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:43 crc kubenswrapper[4824]: I1006 09:58:43.738487 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab36ed34d6ca02b59043c46e151019620ce31733209d072c21394a79fe36be4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:43Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:43 crc kubenswrapper[4824]: I1006 09:58:43.758542 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5rx6l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0c320fc-94d9-4d82-81ee-ccccfd125efc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f8be7bad4e9644494dfb42380c7f5481e1039a1d27ad379c623515948d87392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njtbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5rx6l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:43Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:43 crc kubenswrapper[4824]: I1006 09:58:43.782682 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f9011f6673310dffea3acd7bdc2a56f2af91a87641631490fdc63da610f047a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:43Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:43 crc kubenswrapper[4824]: I1006 09:58:43.801951 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:43 crc kubenswrapper[4824]: I1006 09:58:43.802049 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:43 crc kubenswrapper[4824]: I1006 09:58:43.802073 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:43 crc kubenswrapper[4824]: I1006 09:58:43.802103 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:43 crc kubenswrapper[4824]: I1006 09:58:43.802126 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:43Z","lastTransitionTime":"2025-10-06T09:58:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:43 crc kubenswrapper[4824]: I1006 09:58:43.804092 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:43Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:43 crc kubenswrapper[4824]: I1006 09:58:43.830852 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f511ece-25eb-465f-b85c-9e5f3f886c21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10bfed3e523b7422ce1e981f98166eca78bd12dd4a7cf0634915e1f4ecc3de2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43199f86ad4a1f02bc57a926cd96b2c3c0e2ea46fb36a3f495668706be4c8041\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d6064ad3bcdafebb4c66e7d91f183cc787e9ce13660545c71a722b3ff04ce75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6213164d366243c2139d5971041d6f076cf38fd0cce4c0ca5495111963487ec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa6d59fac987f60a4340ab35d84aa059f2b7085a06362e7207a79702d17eee6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09638fbc9d163ad0b63fa57d24ea5298207ad1f55bab8671be311960be2627a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91a256538722530a7679c5bededd29cff892d03964bd6c71f81d33989b0efdd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d26ec372d0fabf8482f93382dbbd67379654aa4020b4e9831bd214f78c31a6d7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-06T09:58:42Z\\\",\\\"message\\\":\\\" reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1006 09:58:42.062106 6129 factory.go:656] Stopping watch factory\\\\nI1006 09:58:42.062119 6129 handler.go:208] Removed *v1.Node event handler 7\\\\nI1006 09:58:42.062225 6129 handler.go:208] Removed *v1.Node event handler 2\\\\nI1006 09:58:42.062224 6129 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1006 09:58:42.062363 6129 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1006 09:58:42.062930 6129 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1006 09:58:42.063246 6129 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1006 09:58:42.064079 6129 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1006 09:58:42.064220 6129 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1006 09:58:42.064242 6129 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:39Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0644c7db28c7d26aa8d69cd5b9fbcff4d4c06b96df113137dea22a72d9fecd85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6da4c07be4d507a6f20bd8b42be6e6ea4712a956f57ef164036cfded7142a87e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6da4c07be4d507a6f20bd8b42be6e6ea4712a956f57ef164036cfded7142a87e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4vwms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:43Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:43 crc kubenswrapper[4824]: I1006 09:58:43.851174 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b665cda3-da4e-482e-8248-023abec382f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9235c742d8c45bf667f32979c2891fece50f4ee1bd0c7cf6a753ed2d2b9ab30e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b2a920dbc506df2426232990e31fbb60d255c00be54941826a0f094af4a92f0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d516f278f362d7c45e7149de82fa835997ac198f1165c04352910d51fb674d69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c820d8bea16cff0d463ce68d2f2acb72d7a7c73372757ff104e14fcec1025a05\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b8cdf54407cc95783919eba8e92f093bbc09eac5e908c4ba0685629948c71bd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-06T09:58:27Z\\\",\\\"message\\\":\\\"arting RequestHeaderAuthRequestController\\\\nI1006 09:58:27.156944 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1006 09:58:27.156783 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1006 09:58:27.157213 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-432254888/tls.crt::/tmp/serving-cert-432254888/tls.key\\\\\\\"\\\\nI1006 09:58:27.157407 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1006 09:58:27.157443 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1006 09:58:27.157410 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-432254888/tls.crt::/tmp/serving-cert-432254888/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759744705\\\\\\\\\\\\\\\" (2025-10-06 09:58:25 +0000 UTC to 2025-11-05 09:58:26 +0000 UTC (now=2025-10-06 09:58:27.157390076 +0000 UTC))\\\\\\\"\\\\nI1006 09:58:27.157632 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1006 09:58:27.158158 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759744707\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759744707\\\\\\\\\\\\\\\" (2025-10-06 08:58:26 +0000 UTC to 2026-10-06 08:58:26 +0000 UTC (now=2025-10-06 09:58:27.158090663 +0000 UTC))\\\\\\\"\\\\nI1006 09:58:27.158220 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1006 09:58:27.158270 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1006 09:58:27.158310 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nF1006 09:58:27.158617 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://413822c99ba2866521177cce4f1894aef17c160aebb2662da91824a04827d510\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e37e8c7400f7fd14a5c5390a5985e3d8d277e9c5ca60d5e9e9757d07d1778570\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e37e8c7400f7fd14a5c5390a5985e3d8d277e9c5ca60d5e9e9757d07d1778570\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:43Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:43 crc kubenswrapper[4824]: I1006 09:58:43.874404 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d739433-0a38-42fe-b884-b8433011967e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://464e279c5903770c34d80f31ed231b194d5628b39055c4058908039d10690059\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://efa1b62728466ef7aac0e745230348ce920eb9ec9b07fb2cd5a276150dac13fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fedc40f0217626c56d155fc7e78e0592556fef9464db2aaf09c5facc486aad8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78fac300cbe2ac52f5788306b8fc0011f6ba3a22c3546974e70998f2d394987a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:43Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:43 crc kubenswrapper[4824]: I1006 09:58:43.893860 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:43Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:43 crc kubenswrapper[4824]: I1006 09:58:43.907092 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:43 crc kubenswrapper[4824]: I1006 09:58:43.907178 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:43 crc kubenswrapper[4824]: I1006 09:58:43.907204 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:43 crc kubenswrapper[4824]: I1006 09:58:43.907240 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:43 crc kubenswrapper[4824]: I1006 09:58:43.907264 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:43Z","lastTransitionTime":"2025-10-06T09:58:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:43 crc kubenswrapper[4824]: I1006 09:58:43.925123 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-szn8q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b65abf9f-4e88-4571-960d-3ca997d9c344\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://664bf42e1324a9c9529f36f08d61e26999b3386bdd1fac77851fb7e7411e8934\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49s5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-szn8q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:43Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:43 crc kubenswrapper[4824]: I1006 09:58:43.949427 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-7c22q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7eaaa68-d971-4a1c-b6c5-551ded8ff27f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f5150c8c6cfc65ecdcd0e7bd52199dd60e2472c96874f103b49919c81b5e5f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53336c7d3c509b4e022708998aedc5dee9fdaf7136ae3ee69b4181394c3a22fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53336c7d3c509b4e022708998aedc5dee9fdaf7136ae3ee69b4181394c3a22fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://652ab0a13a2bb421872edda3f7188dce8e296247052550c937f8447fea206756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://652ab0a13a2bb421872edda3f7188dce8e296247052550c937f8447fea206756\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3514dec0386bb8c016a7f555d7191ea420d5058b0ee98f97d965dd857c5376e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3514dec0386bb8c016a7f555d7191ea420d5058b0ee98f97d965dd857c5376e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b035e0bf34e3429400ac8942124a3e9834231899487d03d15609fe5624747a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b035e0bf34e3429400ac8942124a3e9834231899487d03d15609fe5624747a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6743e027a11864ad3f433d110e60451fb36d49791e03b22d437b70f89d056ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6743e027a11864ad3f433d110e60451fb36d49791e03b22d437b70f89d056ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8ea48b6886fd39a1c5e5b29ca1fffc2bd4552d61dc621eb2fba66fb464cf120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8ea48b6886fd39a1c5e5b29ca1fffc2bd4552d61dc621eb2fba66fb464cf120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-7c22q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:43Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:44 crc kubenswrapper[4824]: I1006 09:58:44.010884 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:44 crc kubenswrapper[4824]: I1006 09:58:44.010962 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:44 crc kubenswrapper[4824]: I1006 09:58:44.011022 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:44 crc kubenswrapper[4824]: I1006 09:58:44.011055 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:44 crc kubenswrapper[4824]: I1006 09:58:44.011114 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:44Z","lastTransitionTime":"2025-10-06T09:58:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:44 crc kubenswrapper[4824]: I1006 09:58:44.115052 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:44 crc kubenswrapper[4824]: I1006 09:58:44.115135 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:44 crc kubenswrapper[4824]: I1006 09:58:44.115163 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:44 crc kubenswrapper[4824]: I1006 09:58:44.115196 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:44 crc kubenswrapper[4824]: I1006 09:58:44.115221 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:44Z","lastTransitionTime":"2025-10-06T09:58:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:44 crc kubenswrapper[4824]: I1006 09:58:44.218631 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:44 crc kubenswrapper[4824]: I1006 09:58:44.218713 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:44 crc kubenswrapper[4824]: I1006 09:58:44.218732 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:44 crc kubenswrapper[4824]: I1006 09:58:44.218762 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:44 crc kubenswrapper[4824]: I1006 09:58:44.218785 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:44Z","lastTransitionTime":"2025-10-06T09:58:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:44 crc kubenswrapper[4824]: I1006 09:58:44.273304 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 09:58:44 crc kubenswrapper[4824]: I1006 09:58:44.273352 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 09:58:44 crc kubenswrapper[4824]: I1006 09:58:44.273406 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 09:58:44 crc kubenswrapper[4824]: E1006 09:58:44.273516 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 06 09:58:44 crc kubenswrapper[4824]: E1006 09:58:44.273919 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 06 09:58:44 crc kubenswrapper[4824]: E1006 09:58:44.274103 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 06 09:58:44 crc kubenswrapper[4824]: I1006 09:58:44.322062 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:44 crc kubenswrapper[4824]: I1006 09:58:44.322133 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:44 crc kubenswrapper[4824]: I1006 09:58:44.322154 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:44 crc kubenswrapper[4824]: I1006 09:58:44.322186 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:44 crc kubenswrapper[4824]: I1006 09:58:44.322210 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:44Z","lastTransitionTime":"2025-10-06T09:58:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:44 crc kubenswrapper[4824]: I1006 09:58:44.425941 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:44 crc kubenswrapper[4824]: I1006 09:58:44.426015 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:44 crc kubenswrapper[4824]: I1006 09:58:44.426030 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:44 crc kubenswrapper[4824]: I1006 09:58:44.426049 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:44 crc kubenswrapper[4824]: I1006 09:58:44.426062 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:44Z","lastTransitionTime":"2025-10-06T09:58:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:44 crc kubenswrapper[4824]: I1006 09:58:44.529800 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:44 crc kubenswrapper[4824]: I1006 09:58:44.529874 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:44 crc kubenswrapper[4824]: I1006 09:58:44.529891 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:44 crc kubenswrapper[4824]: I1006 09:58:44.529918 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:44 crc kubenswrapper[4824]: I1006 09:58:44.529937 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:44Z","lastTransitionTime":"2025-10-06T09:58:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:44 crc kubenswrapper[4824]: I1006 09:58:44.633496 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:44 crc kubenswrapper[4824]: I1006 09:58:44.633548 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:44 crc kubenswrapper[4824]: I1006 09:58:44.633565 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:44 crc kubenswrapper[4824]: I1006 09:58:44.633592 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:44 crc kubenswrapper[4824]: I1006 09:58:44.633615 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:44Z","lastTransitionTime":"2025-10-06T09:58:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:44 crc kubenswrapper[4824]: I1006 09:58:44.635254 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-4vwms_2f511ece-25eb-465f-b85c-9e5f3f886c21/ovnkube-controller/1.log" Oct 06 09:58:44 crc kubenswrapper[4824]: I1006 09:58:44.636421 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-4vwms_2f511ece-25eb-465f-b85c-9e5f3f886c21/ovnkube-controller/0.log" Oct 06 09:58:44 crc kubenswrapper[4824]: I1006 09:58:44.642754 4824 generic.go:334] "Generic (PLEG): container finished" podID="2f511ece-25eb-465f-b85c-9e5f3f886c21" containerID="91a256538722530a7679c5bededd29cff892d03964bd6c71f81d33989b0efdd9" exitCode=1 Oct 06 09:58:44 crc kubenswrapper[4824]: I1006 09:58:44.642837 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" event={"ID":"2f511ece-25eb-465f-b85c-9e5f3f886c21","Type":"ContainerDied","Data":"91a256538722530a7679c5bededd29cff892d03964bd6c71f81d33989b0efdd9"} Oct 06 09:58:44 crc kubenswrapper[4824]: I1006 09:58:44.642943 4824 scope.go:117] "RemoveContainer" containerID="d26ec372d0fabf8482f93382dbbd67379654aa4020b4e9831bd214f78c31a6d7" Oct 06 09:58:44 crc kubenswrapper[4824]: I1006 09:58:44.644209 4824 scope.go:117] "RemoveContainer" containerID="91a256538722530a7679c5bededd29cff892d03964bd6c71f81d33989b0efdd9" Oct 06 09:58:44 crc kubenswrapper[4824]: E1006 09:58:44.644579 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-4vwms_openshift-ovn-kubernetes(2f511ece-25eb-465f-b85c-9e5f3f886c21)\"" pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" podUID="2f511ece-25eb-465f-b85c-9e5f3f886c21" Oct 06 09:58:44 crc kubenswrapper[4824]: I1006 09:58:44.666343 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lszvg"] Oct 06 09:58:44 crc kubenswrapper[4824]: I1006 09:58:44.667185 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lszvg" Oct 06 09:58:44 crc kubenswrapper[4824]: I1006 09:58:44.669958 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Oct 06 09:58:44 crc kubenswrapper[4824]: I1006 09:58:44.671281 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Oct 06 09:58:44 crc kubenswrapper[4824]: I1006 09:58:44.672153 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-7c22q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7eaaa68-d971-4a1c-b6c5-551ded8ff27f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f5150c8c6cfc65ecdcd0e7bd52199dd60e2472c96874f103b49919c81b5e5f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53336c7d3c509b4e022708998aedc5dee9fdaf7136ae3ee69b4181394c3a22fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53336c7d3c509b4e022708998aedc5dee9fdaf7136ae3ee69b4181394c3a22fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://652ab0a13a2bb421872edda3f7188dce8e296247052550c937f8447fea206756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://652ab0a13a2bb421872edda3f7188dce8e296247052550c937f8447fea206756\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3514dec0386bb8c016a7f555d7191ea420d5058b0ee98f97d965dd857c5376e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3514dec0386bb8c016a7f555d7191ea420d5058b0ee98f97d965dd857c5376e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b035e0bf34e3429400ac8942124a3e9834231899487d03d15609fe5624747a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b035e0bf34e3429400ac8942124a3e9834231899487d03d15609fe5624747a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6743e027a11864ad3f433d110e60451fb36d49791e03b22d437b70f89d056ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6743e027a11864ad3f433d110e60451fb36d49791e03b22d437b70f89d056ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8ea48b6886fd39a1c5e5b29ca1fffc2bd4552d61dc621eb2fba66fb464cf120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8ea48b6886fd39a1c5e5b29ca1fffc2bd4552d61dc621eb2fba66fb464cf120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-7c22q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:44Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:44 crc kubenswrapper[4824]: I1006 09:58:44.692854 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/0cf70ab4-4cfe-41c5-8db7-035451bafcfa-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-lszvg\" (UID: \"0cf70ab4-4cfe-41c5-8db7-035451bafcfa\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lszvg" Oct 06 09:58:44 crc kubenswrapper[4824]: I1006 09:58:44.692955 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/0cf70ab4-4cfe-41c5-8db7-035451bafcfa-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-lszvg\" (UID: \"0cf70ab4-4cfe-41c5-8db7-035451bafcfa\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lszvg" Oct 06 09:58:44 crc kubenswrapper[4824]: I1006 09:58:44.693095 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/0cf70ab4-4cfe-41c5-8db7-035451bafcfa-env-overrides\") pod \"ovnkube-control-plane-749d76644c-lszvg\" (UID: \"0cf70ab4-4cfe-41c5-8db7-035451bafcfa\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lszvg" Oct 06 09:58:44 crc kubenswrapper[4824]: I1006 09:58:44.693262 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9fkbf\" (UniqueName: \"kubernetes.io/projected/0cf70ab4-4cfe-41c5-8db7-035451bafcfa-kube-api-access-9fkbf\") pod \"ovnkube-control-plane-749d76644c-lszvg\" (UID: \"0cf70ab4-4cfe-41c5-8db7-035451bafcfa\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lszvg" Oct 06 09:58:44 crc kubenswrapper[4824]: I1006 09:58:44.698330 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2035278cf272827b39a45aa41ecd5b1e41aafc1d9eb8bccd853660fb6e587159\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff3152fceadf36a55578f587623f37d748f90f2afeda5702a05504b372d399be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:44Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:44 crc kubenswrapper[4824]: I1006 09:58:44.714941 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xcgwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74317f75-9ebc-4327-b88c-598c2241240f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8def09a126f9bb910affad9b7fb32303b960add9fbfd5bb709d59b4aeba7d245\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v6sg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xcgwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:44Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:44 crc kubenswrapper[4824]: I1006 09:58:44.733925 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1140c8c-93f1-4fce-8c06-a69561ff8a6d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5462cb77b7d61aed7d4b4be2fbc6ba1451d267a66afd6dede2f259c15230e7c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gq49w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5512be36013baeca611ee92645a1d339da237b70d63aac43334beafe24513d44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gq49w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-khgzw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:44Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:44 crc kubenswrapper[4824]: I1006 09:58:44.737116 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:44 crc kubenswrapper[4824]: I1006 09:58:44.737166 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:44 crc kubenswrapper[4824]: I1006 09:58:44.737187 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:44 crc kubenswrapper[4824]: I1006 09:58:44.737218 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:44 crc kubenswrapper[4824]: I1006 09:58:44.737240 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:44Z","lastTransitionTime":"2025-10-06T09:58:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:44 crc kubenswrapper[4824]: I1006 09:58:44.756555 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:44Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:44 crc kubenswrapper[4824]: I1006 09:58:44.776627 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab36ed34d6ca02b59043c46e151019620ce31733209d072c21394a79fe36be4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:44Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:44 crc kubenswrapper[4824]: I1006 09:58:44.793395 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5rx6l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0c320fc-94d9-4d82-81ee-ccccfd125efc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f8be7bad4e9644494dfb42380c7f5481e1039a1d27ad379c623515948d87392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njtbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5rx6l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:44Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:44 crc kubenswrapper[4824]: I1006 09:58:44.794941 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/0cf70ab4-4cfe-41c5-8db7-035451bafcfa-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-lszvg\" (UID: \"0cf70ab4-4cfe-41c5-8db7-035451bafcfa\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lszvg" Oct 06 09:58:44 crc kubenswrapper[4824]: I1006 09:58:44.795058 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/0cf70ab4-4cfe-41c5-8db7-035451bafcfa-env-overrides\") pod \"ovnkube-control-plane-749d76644c-lszvg\" (UID: \"0cf70ab4-4cfe-41c5-8db7-035451bafcfa\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lszvg" Oct 06 09:58:44 crc kubenswrapper[4824]: I1006 09:58:44.795098 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9fkbf\" (UniqueName: \"kubernetes.io/projected/0cf70ab4-4cfe-41c5-8db7-035451bafcfa-kube-api-access-9fkbf\") pod \"ovnkube-control-plane-749d76644c-lszvg\" (UID: \"0cf70ab4-4cfe-41c5-8db7-035451bafcfa\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lszvg" Oct 06 09:58:44 crc kubenswrapper[4824]: I1006 09:58:44.795164 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/0cf70ab4-4cfe-41c5-8db7-035451bafcfa-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-lszvg\" (UID: \"0cf70ab4-4cfe-41c5-8db7-035451bafcfa\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lszvg" Oct 06 09:58:44 crc kubenswrapper[4824]: I1006 09:58:44.796285 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/0cf70ab4-4cfe-41c5-8db7-035451bafcfa-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-lszvg\" (UID: \"0cf70ab4-4cfe-41c5-8db7-035451bafcfa\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lszvg" Oct 06 09:58:44 crc kubenswrapper[4824]: I1006 09:58:44.797386 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/0cf70ab4-4cfe-41c5-8db7-035451bafcfa-env-overrides\") pod \"ovnkube-control-plane-749d76644c-lszvg\" (UID: \"0cf70ab4-4cfe-41c5-8db7-035451bafcfa\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lszvg" Oct 06 09:58:44 crc kubenswrapper[4824]: I1006 09:58:44.803191 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/0cf70ab4-4cfe-41c5-8db7-035451bafcfa-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-lszvg\" (UID: \"0cf70ab4-4cfe-41c5-8db7-035451bafcfa\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lszvg" Oct 06 09:58:44 crc kubenswrapper[4824]: I1006 09:58:44.816588 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f9011f6673310dffea3acd7bdc2a56f2af91a87641631490fdc63da610f047a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:44Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:44 crc kubenswrapper[4824]: I1006 09:58:44.821731 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9fkbf\" (UniqueName: \"kubernetes.io/projected/0cf70ab4-4cfe-41c5-8db7-035451bafcfa-kube-api-access-9fkbf\") pod \"ovnkube-control-plane-749d76644c-lszvg\" (UID: \"0cf70ab4-4cfe-41c5-8db7-035451bafcfa\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lszvg" Oct 06 09:58:44 crc kubenswrapper[4824]: I1006 09:58:44.838387 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:44Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:44 crc kubenswrapper[4824]: I1006 09:58:44.841937 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:44 crc kubenswrapper[4824]: I1006 09:58:44.842075 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:44 crc kubenswrapper[4824]: I1006 09:58:44.842140 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:44 crc kubenswrapper[4824]: I1006 09:58:44.842175 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:44 crc kubenswrapper[4824]: I1006 09:58:44.842251 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:44Z","lastTransitionTime":"2025-10-06T09:58:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:44 crc kubenswrapper[4824]: I1006 09:58:44.868786 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f511ece-25eb-465f-b85c-9e5f3f886c21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10bfed3e523b7422ce1e981f98166eca78bd12dd4a7cf0634915e1f4ecc3de2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43199f86ad4a1f02bc57a926cd96b2c3c0e2ea46fb36a3f495668706be4c8041\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d6064ad3bcdafebb4c66e7d91f183cc787e9ce13660545c71a722b3ff04ce75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6213164d366243c2139d5971041d6f076cf38fd0cce4c0ca5495111963487ec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa6d59fac987f60a4340ab35d84aa059f2b7085a06362e7207a79702d17eee6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09638fbc9d163ad0b63fa57d24ea5298207ad1f55bab8671be311960be2627a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91a256538722530a7679c5bededd29cff892d03964bd6c71f81d33989b0efdd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d26ec372d0fabf8482f93382dbbd67379654aa4020b4e9831bd214f78c31a6d7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-06T09:58:42Z\\\",\\\"message\\\":\\\" reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1006 09:58:42.062106 6129 factory.go:656] Stopping watch factory\\\\nI1006 09:58:42.062119 6129 handler.go:208] Removed *v1.Node event handler 7\\\\nI1006 09:58:42.062225 6129 handler.go:208] Removed *v1.Node event handler 2\\\\nI1006 09:58:42.062224 6129 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1006 09:58:42.062363 6129 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1006 09:58:42.062930 6129 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1006 09:58:42.063246 6129 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1006 09:58:42.064079 6129 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1006 09:58:42.064220 6129 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1006 09:58:42.064242 6129 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:39Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91a256538722530a7679c5bededd29cff892d03964bd6c71f81d33989b0efdd9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-06T09:58:43Z\\\",\\\"message\\\":\\\" to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:43Z is after 2025-08-24T17:21:41Z]\\\\nI1006 09:58:43.725317 6283 obj_retry.go:365] Adding new object: *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI1006 09:58:43.725319 6283 services_controller.go:356] Processing sync for service openshift-marketplace/marketplace-operator-metrics for network=default\\\\nI1006 09:58:43.725320 6283 lb_config.go:1031] Cluster endpoints for openshift-machine-config-operator/machine-config-controller for network=default are: map[]\\\\nI1006 09:58:43.725336 6283 ovn.go:134] Ensuring zone local for Pod openshift-kube-controller-manager/kube-controller-manager-crc in node crc\\\\nI1006 09:58:43.725345 6283 obj_retry.go:386] Retry successful for *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc after 0 failed attempt(s)\\\\nI1006 09:58:43.725352 6283 default_network_controller.go:776] Recording success event on pod openshift-ku\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0644c7db28c7d26aa8d69cd5b9fbcff4d4c06b96df113137dea22a72d9fecd85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6da4c07be4d507a6f20bd8b42be6e6ea4712a956f57ef164036cfded7142a87e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6da4c07be4d507a6f20bd8b42be6e6ea4712a956f57ef164036cfded7142a87e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4vwms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:44Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:44 crc kubenswrapper[4824]: I1006 09:58:44.894179 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b665cda3-da4e-482e-8248-023abec382f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9235c742d8c45bf667f32979c2891fece50f4ee1bd0c7cf6a753ed2d2b9ab30e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b2a920dbc506df2426232990e31fbb60d255c00be54941826a0f094af4a92f0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d516f278f362d7c45e7149de82fa835997ac198f1165c04352910d51fb674d69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c820d8bea16cff0d463ce68d2f2acb72d7a7c73372757ff104e14fcec1025a05\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b8cdf54407cc95783919eba8e92f093bbc09eac5e908c4ba0685629948c71bd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-06T09:58:27Z\\\",\\\"message\\\":\\\"arting RequestHeaderAuthRequestController\\\\nI1006 09:58:27.156944 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1006 09:58:27.156783 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1006 09:58:27.157213 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-432254888/tls.crt::/tmp/serving-cert-432254888/tls.key\\\\\\\"\\\\nI1006 09:58:27.157407 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1006 09:58:27.157443 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1006 09:58:27.157410 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-432254888/tls.crt::/tmp/serving-cert-432254888/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759744705\\\\\\\\\\\\\\\" (2025-10-06 09:58:25 +0000 UTC to 2025-11-05 09:58:26 +0000 UTC (now=2025-10-06 09:58:27.157390076 +0000 UTC))\\\\\\\"\\\\nI1006 09:58:27.157632 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1006 09:58:27.158158 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759744707\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759744707\\\\\\\\\\\\\\\" (2025-10-06 08:58:26 +0000 UTC to 2026-10-06 08:58:26 +0000 UTC (now=2025-10-06 09:58:27.158090663 +0000 UTC))\\\\\\\"\\\\nI1006 09:58:27.158220 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1006 09:58:27.158270 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1006 09:58:27.158310 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nF1006 09:58:27.158617 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://413822c99ba2866521177cce4f1894aef17c160aebb2662da91824a04827d510\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e37e8c7400f7fd14a5c5390a5985e3d8d277e9c5ca60d5e9e9757d07d1778570\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e37e8c7400f7fd14a5c5390a5985e3d8d277e9c5ca60d5e9e9757d07d1778570\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:44Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:44 crc kubenswrapper[4824]: I1006 09:58:44.916557 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d739433-0a38-42fe-b884-b8433011967e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://464e279c5903770c34d80f31ed231b194d5628b39055c4058908039d10690059\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://efa1b62728466ef7aac0e745230348ce920eb9ec9b07fb2cd5a276150dac13fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fedc40f0217626c56d155fc7e78e0592556fef9464db2aaf09c5facc486aad8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78fac300cbe2ac52f5788306b8fc0011f6ba3a22c3546974e70998f2d394987a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:44Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:44 crc kubenswrapper[4824]: I1006 09:58:44.940894 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:44Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:44 crc kubenswrapper[4824]: I1006 09:58:44.945241 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:44 crc kubenswrapper[4824]: I1006 09:58:44.945328 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:44 crc kubenswrapper[4824]: I1006 09:58:44.945356 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:44 crc kubenswrapper[4824]: I1006 09:58:44.945397 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:44 crc kubenswrapper[4824]: I1006 09:58:44.945426 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:44Z","lastTransitionTime":"2025-10-06T09:58:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:44 crc kubenswrapper[4824]: I1006 09:58:44.963827 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-szn8q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b65abf9f-4e88-4571-960d-3ca997d9c344\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://664bf42e1324a9c9529f36f08d61e26999b3386bdd1fac77851fb7e7411e8934\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49s5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-szn8q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:44Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:44 crc kubenswrapper[4824]: I1006 09:58:44.986027 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2035278cf272827b39a45aa41ecd5b1e41aafc1d9eb8bccd853660fb6e587159\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff3152fceadf36a55578f587623f37d748f90f2afeda5702a05504b372d399be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:44Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:44 crc kubenswrapper[4824]: I1006 09:58:44.989035 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lszvg" Oct 06 09:58:45 crc kubenswrapper[4824]: I1006 09:58:45.004439 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xcgwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74317f75-9ebc-4327-b88c-598c2241240f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8def09a126f9bb910affad9b7fb32303b960add9fbfd5bb709d59b4aeba7d245\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v6sg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xcgwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:45Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:45 crc kubenswrapper[4824]: W1006 09:58:45.015030 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0cf70ab4_4cfe_41c5_8db7_035451bafcfa.slice/crio-89772e12b436860e435b7efd177c43dd7f073148892797d2666777765c956f15 WatchSource:0}: Error finding container 89772e12b436860e435b7efd177c43dd7f073148892797d2666777765c956f15: Status 404 returned error can't find the container with id 89772e12b436860e435b7efd177c43dd7f073148892797d2666777765c956f15 Oct 06 09:58:45 crc kubenswrapper[4824]: I1006 09:58:45.024753 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1140c8c-93f1-4fce-8c06-a69561ff8a6d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5462cb77b7d61aed7d4b4be2fbc6ba1451d267a66afd6dede2f259c15230e7c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gq49w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5512be36013baeca611ee92645a1d339da237b70d63aac43334beafe24513d44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gq49w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-khgzw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:45Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:45 crc kubenswrapper[4824]: I1006 09:58:45.042411 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5rx6l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0c320fc-94d9-4d82-81ee-ccccfd125efc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f8be7bad4e9644494dfb42380c7f5481e1039a1d27ad379c623515948d87392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njtbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5rx6l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:45Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:45 crc kubenswrapper[4824]: I1006 09:58:45.049208 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:45 crc kubenswrapper[4824]: I1006 09:58:45.049370 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:45 crc kubenswrapper[4824]: I1006 09:58:45.049394 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:45 crc kubenswrapper[4824]: I1006 09:58:45.049544 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:45 crc kubenswrapper[4824]: I1006 09:58:45.049621 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:45Z","lastTransitionTime":"2025-10-06T09:58:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:45 crc kubenswrapper[4824]: I1006 09:58:45.064422 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f9011f6673310dffea3acd7bdc2a56f2af91a87641631490fdc63da610f047a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:45Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:45 crc kubenswrapper[4824]: I1006 09:58:45.085569 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:45Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:45 crc kubenswrapper[4824]: I1006 09:58:45.107352 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:45Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:45 crc kubenswrapper[4824]: I1006 09:58:45.125350 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab36ed34d6ca02b59043c46e151019620ce31733209d072c21394a79fe36be4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:45Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:45 crc kubenswrapper[4824]: I1006 09:58:45.148274 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b665cda3-da4e-482e-8248-023abec382f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9235c742d8c45bf667f32979c2891fece50f4ee1bd0c7cf6a753ed2d2b9ab30e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b2a920dbc506df2426232990e31fbb60d255c00be54941826a0f094af4a92f0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d516f278f362d7c45e7149de82fa835997ac198f1165c04352910d51fb674d69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c820d8bea16cff0d463ce68d2f2acb72d7a7c73372757ff104e14fcec1025a05\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b8cdf54407cc95783919eba8e92f093bbc09eac5e908c4ba0685629948c71bd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-06T09:58:27Z\\\",\\\"message\\\":\\\"arting RequestHeaderAuthRequestController\\\\nI1006 09:58:27.156944 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1006 09:58:27.156783 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1006 09:58:27.157213 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-432254888/tls.crt::/tmp/serving-cert-432254888/tls.key\\\\\\\"\\\\nI1006 09:58:27.157407 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1006 09:58:27.157443 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1006 09:58:27.157410 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-432254888/tls.crt::/tmp/serving-cert-432254888/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759744705\\\\\\\\\\\\\\\" (2025-10-06 09:58:25 +0000 UTC to 2025-11-05 09:58:26 +0000 UTC (now=2025-10-06 09:58:27.157390076 +0000 UTC))\\\\\\\"\\\\nI1006 09:58:27.157632 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1006 09:58:27.158158 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759744707\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759744707\\\\\\\\\\\\\\\" (2025-10-06 08:58:26 +0000 UTC to 2026-10-06 08:58:26 +0000 UTC (now=2025-10-06 09:58:27.158090663 +0000 UTC))\\\\\\\"\\\\nI1006 09:58:27.158220 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1006 09:58:27.158270 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1006 09:58:27.158310 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nF1006 09:58:27.158617 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://413822c99ba2866521177cce4f1894aef17c160aebb2662da91824a04827d510\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e37e8c7400f7fd14a5c5390a5985e3d8d277e9c5ca60d5e9e9757d07d1778570\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e37e8c7400f7fd14a5c5390a5985e3d8d277e9c5ca60d5e9e9757d07d1778570\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:45Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:45 crc kubenswrapper[4824]: I1006 09:58:45.155387 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:45 crc kubenswrapper[4824]: I1006 09:58:45.155463 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:45 crc kubenswrapper[4824]: I1006 09:58:45.155484 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:45 crc kubenswrapper[4824]: I1006 09:58:45.155512 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:45 crc kubenswrapper[4824]: I1006 09:58:45.155531 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:45Z","lastTransitionTime":"2025-10-06T09:58:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:45 crc kubenswrapper[4824]: I1006 09:58:45.169952 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d739433-0a38-42fe-b884-b8433011967e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://464e279c5903770c34d80f31ed231b194d5628b39055c4058908039d10690059\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://efa1b62728466ef7aac0e745230348ce920eb9ec9b07fb2cd5a276150dac13fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fedc40f0217626c56d155fc7e78e0592556fef9464db2aaf09c5facc486aad8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78fac300cbe2ac52f5788306b8fc0011f6ba3a22c3546974e70998f2d394987a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:45Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:45 crc kubenswrapper[4824]: I1006 09:58:45.193765 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:45Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:45 crc kubenswrapper[4824]: I1006 09:58:45.212096 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-szn8q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b65abf9f-4e88-4571-960d-3ca997d9c344\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://664bf42e1324a9c9529f36f08d61e26999b3386bdd1fac77851fb7e7411e8934\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49s5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-szn8q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:45Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:45 crc kubenswrapper[4824]: I1006 09:58:45.246478 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f511ece-25eb-465f-b85c-9e5f3f886c21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10bfed3e523b7422ce1e981f98166eca78bd12dd4a7cf0634915e1f4ecc3de2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43199f86ad4a1f02bc57a926cd96b2c3c0e2ea46fb36a3f495668706be4c8041\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d6064ad3bcdafebb4c66e7d91f183cc787e9ce13660545c71a722b3ff04ce75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6213164d366243c2139d5971041d6f076cf38fd0cce4c0ca5495111963487ec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa6d59fac987f60a4340ab35d84aa059f2b7085a06362e7207a79702d17eee6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09638fbc9d163ad0b63fa57d24ea5298207ad1f55bab8671be311960be2627a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91a256538722530a7679c5bededd29cff892d03964bd6c71f81d33989b0efdd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d26ec372d0fabf8482f93382dbbd67379654aa4020b4e9831bd214f78c31a6d7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-06T09:58:42Z\\\",\\\"message\\\":\\\" reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1006 09:58:42.062106 6129 factory.go:656] Stopping watch factory\\\\nI1006 09:58:42.062119 6129 handler.go:208] Removed *v1.Node event handler 7\\\\nI1006 09:58:42.062225 6129 handler.go:208] Removed *v1.Node event handler 2\\\\nI1006 09:58:42.062224 6129 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1006 09:58:42.062363 6129 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1006 09:58:42.062930 6129 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1006 09:58:42.063246 6129 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1006 09:58:42.064079 6129 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1006 09:58:42.064220 6129 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1006 09:58:42.064242 6129 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:39Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91a256538722530a7679c5bededd29cff892d03964bd6c71f81d33989b0efdd9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-06T09:58:43Z\\\",\\\"message\\\":\\\" to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:43Z is after 2025-08-24T17:21:41Z]\\\\nI1006 09:58:43.725317 6283 obj_retry.go:365] Adding new object: *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI1006 09:58:43.725319 6283 services_controller.go:356] Processing sync for service openshift-marketplace/marketplace-operator-metrics for network=default\\\\nI1006 09:58:43.725320 6283 lb_config.go:1031] Cluster endpoints for openshift-machine-config-operator/machine-config-controller for network=default are: map[]\\\\nI1006 09:58:43.725336 6283 ovn.go:134] Ensuring zone local for Pod openshift-kube-controller-manager/kube-controller-manager-crc in node crc\\\\nI1006 09:58:43.725345 6283 obj_retry.go:386] Retry successful for *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc after 0 failed attempt(s)\\\\nI1006 09:58:43.725352 6283 default_network_controller.go:776] Recording success event on pod openshift-ku\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0644c7db28c7d26aa8d69cd5b9fbcff4d4c06b96df113137dea22a72d9fecd85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6da4c07be4d507a6f20bd8b42be6e6ea4712a956f57ef164036cfded7142a87e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6da4c07be4d507a6f20bd8b42be6e6ea4712a956f57ef164036cfded7142a87e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4vwms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:45Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:45 crc kubenswrapper[4824]: I1006 09:58:45.259123 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:45 crc kubenswrapper[4824]: I1006 09:58:45.259184 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:45 crc kubenswrapper[4824]: I1006 09:58:45.259197 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:45 crc kubenswrapper[4824]: I1006 09:58:45.259222 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:45 crc kubenswrapper[4824]: I1006 09:58:45.259237 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:45Z","lastTransitionTime":"2025-10-06T09:58:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:45 crc kubenswrapper[4824]: I1006 09:58:45.266335 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lszvg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0cf70ab4-4cfe-41c5-8db7-035451bafcfa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fkbf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fkbf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lszvg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:45Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:45 crc kubenswrapper[4824]: I1006 09:58:45.289875 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-7c22q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7eaaa68-d971-4a1c-b6c5-551ded8ff27f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f5150c8c6cfc65ecdcd0e7bd52199dd60e2472c96874f103b49919c81b5e5f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53336c7d3c509b4e022708998aedc5dee9fdaf7136ae3ee69b4181394c3a22fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53336c7d3c509b4e022708998aedc5dee9fdaf7136ae3ee69b4181394c3a22fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://652ab0a13a2bb421872edda3f7188dce8e296247052550c937f8447fea206756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://652ab0a13a2bb421872edda3f7188dce8e296247052550c937f8447fea206756\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3514dec0386bb8c016a7f555d7191ea420d5058b0ee98f97d965dd857c5376e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3514dec0386bb8c016a7f555d7191ea420d5058b0ee98f97d965dd857c5376e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b035e0bf34e3429400ac8942124a3e9834231899487d03d15609fe5624747a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b035e0bf34e3429400ac8942124a3e9834231899487d03d15609fe5624747a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6743e027a11864ad3f433d110e60451fb36d49791e03b22d437b70f89d056ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6743e027a11864ad3f433d110e60451fb36d49791e03b22d437b70f89d056ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8ea48b6886fd39a1c5e5b29ca1fffc2bd4552d61dc621eb2fba66fb464cf120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8ea48b6886fd39a1c5e5b29ca1fffc2bd4552d61dc621eb2fba66fb464cf120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-7c22q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:45Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:45 crc kubenswrapper[4824]: I1006 09:58:45.362615 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:45 crc kubenswrapper[4824]: I1006 09:58:45.362675 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:45 crc kubenswrapper[4824]: I1006 09:58:45.362696 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:45 crc kubenswrapper[4824]: I1006 09:58:45.362723 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:45 crc kubenswrapper[4824]: I1006 09:58:45.362741 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:45Z","lastTransitionTime":"2025-10-06T09:58:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:45 crc kubenswrapper[4824]: I1006 09:58:45.466941 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:45 crc kubenswrapper[4824]: I1006 09:58:45.467034 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:45 crc kubenswrapper[4824]: I1006 09:58:45.467054 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:45 crc kubenswrapper[4824]: I1006 09:58:45.467088 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:45 crc kubenswrapper[4824]: I1006 09:58:45.467109 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:45Z","lastTransitionTime":"2025-10-06T09:58:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:45 crc kubenswrapper[4824]: I1006 09:58:45.571033 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:45 crc kubenswrapper[4824]: I1006 09:58:45.571101 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:45 crc kubenswrapper[4824]: I1006 09:58:45.571124 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:45 crc kubenswrapper[4824]: I1006 09:58:45.571152 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:45 crc kubenswrapper[4824]: I1006 09:58:45.571175 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:45Z","lastTransitionTime":"2025-10-06T09:58:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:45 crc kubenswrapper[4824]: I1006 09:58:45.656410 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-4vwms_2f511ece-25eb-465f-b85c-9e5f3f886c21/ovnkube-controller/1.log" Oct 06 09:58:45 crc kubenswrapper[4824]: I1006 09:58:45.662344 4824 scope.go:117] "RemoveContainer" containerID="91a256538722530a7679c5bededd29cff892d03964bd6c71f81d33989b0efdd9" Oct 06 09:58:45 crc kubenswrapper[4824]: E1006 09:58:45.663114 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-4vwms_openshift-ovn-kubernetes(2f511ece-25eb-465f-b85c-9e5f3f886c21)\"" pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" podUID="2f511ece-25eb-465f-b85c-9e5f3f886c21" Oct 06 09:58:45 crc kubenswrapper[4824]: I1006 09:58:45.666475 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lszvg" event={"ID":"0cf70ab4-4cfe-41c5-8db7-035451bafcfa","Type":"ContainerStarted","Data":"519653c6b241cd89ef3ece003667d420a7a55cdf513f4577a69e14ade5fb4c76"} Oct 06 09:58:45 crc kubenswrapper[4824]: I1006 09:58:45.666546 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lszvg" event={"ID":"0cf70ab4-4cfe-41c5-8db7-035451bafcfa","Type":"ContainerStarted","Data":"a7b80b5b375b3b32132501ee0d661e06e199654c696ac76ad56e7ba96918cc77"} Oct 06 09:58:45 crc kubenswrapper[4824]: I1006 09:58:45.666576 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lszvg" event={"ID":"0cf70ab4-4cfe-41c5-8db7-035451bafcfa","Type":"ContainerStarted","Data":"89772e12b436860e435b7efd177c43dd7f073148892797d2666777765c956f15"} Oct 06 09:58:45 crc kubenswrapper[4824]: I1006 09:58:45.674250 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:45 crc kubenswrapper[4824]: I1006 09:58:45.674308 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:45 crc kubenswrapper[4824]: I1006 09:58:45.674327 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:45 crc kubenswrapper[4824]: I1006 09:58:45.674356 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:45 crc kubenswrapper[4824]: I1006 09:58:45.674373 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:45Z","lastTransitionTime":"2025-10-06T09:58:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:45 crc kubenswrapper[4824]: I1006 09:58:45.689027 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2035278cf272827b39a45aa41ecd5b1e41aafc1d9eb8bccd853660fb6e587159\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff3152fceadf36a55578f587623f37d748f90f2afeda5702a05504b372d399be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:45Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:45 crc kubenswrapper[4824]: I1006 09:58:45.704822 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xcgwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74317f75-9ebc-4327-b88c-598c2241240f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8def09a126f9bb910affad9b7fb32303b960add9fbfd5bb709d59b4aeba7d245\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v6sg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xcgwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:45Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:45 crc kubenswrapper[4824]: I1006 09:58:45.726183 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1140c8c-93f1-4fce-8c06-a69561ff8a6d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5462cb77b7d61aed7d4b4be2fbc6ba1451d267a66afd6dede2f259c15230e7c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gq49w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5512be36013baeca611ee92645a1d339da237b70d63aac43334beafe24513d44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gq49w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-khgzw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:45Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:45 crc kubenswrapper[4824]: I1006 09:58:45.746844 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:45Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:45 crc kubenswrapper[4824]: I1006 09:58:45.762336 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab36ed34d6ca02b59043c46e151019620ce31733209d072c21394a79fe36be4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:45Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:45 crc kubenswrapper[4824]: I1006 09:58:45.778183 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:45 crc kubenswrapper[4824]: I1006 09:58:45.778252 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:45 crc kubenswrapper[4824]: I1006 09:58:45.778270 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:45 crc kubenswrapper[4824]: I1006 09:58:45.778296 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:45 crc kubenswrapper[4824]: I1006 09:58:45.778327 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:45Z","lastTransitionTime":"2025-10-06T09:58:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:45 crc kubenswrapper[4824]: I1006 09:58:45.781628 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5rx6l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0c320fc-94d9-4d82-81ee-ccccfd125efc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f8be7bad4e9644494dfb42380c7f5481e1039a1d27ad379c623515948d87392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njtbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5rx6l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:45Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:45 crc kubenswrapper[4824]: I1006 09:58:45.801809 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f9011f6673310dffea3acd7bdc2a56f2af91a87641631490fdc63da610f047a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:45Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:45 crc kubenswrapper[4824]: I1006 09:58:45.820677 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:45Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:45 crc kubenswrapper[4824]: I1006 09:58:45.848128 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f511ece-25eb-465f-b85c-9e5f3f886c21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10bfed3e523b7422ce1e981f98166eca78bd12dd4a7cf0634915e1f4ecc3de2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43199f86ad4a1f02bc57a926cd96b2c3c0e2ea46fb36a3f495668706be4c8041\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d6064ad3bcdafebb4c66e7d91f183cc787e9ce13660545c71a722b3ff04ce75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6213164d366243c2139d5971041d6f076cf38fd0cce4c0ca5495111963487ec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa6d59fac987f60a4340ab35d84aa059f2b7085a06362e7207a79702d17eee6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09638fbc9d163ad0b63fa57d24ea5298207ad1f55bab8671be311960be2627a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91a256538722530a7679c5bededd29cff892d03964bd6c71f81d33989b0efdd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91a256538722530a7679c5bededd29cff892d03964bd6c71f81d33989b0efdd9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-06T09:58:43Z\\\",\\\"message\\\":\\\" to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:43Z is after 2025-08-24T17:21:41Z]\\\\nI1006 09:58:43.725317 6283 obj_retry.go:365] Adding new object: *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI1006 09:58:43.725319 6283 services_controller.go:356] Processing sync for service openshift-marketplace/marketplace-operator-metrics for network=default\\\\nI1006 09:58:43.725320 6283 lb_config.go:1031] Cluster endpoints for openshift-machine-config-operator/machine-config-controller for network=default are: map[]\\\\nI1006 09:58:43.725336 6283 ovn.go:134] Ensuring zone local for Pod openshift-kube-controller-manager/kube-controller-manager-crc in node crc\\\\nI1006 09:58:43.725345 6283 obj_retry.go:386] Retry successful for *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc after 0 failed attempt(s)\\\\nI1006 09:58:43.725352 6283 default_network_controller.go:776] Recording success event on pod openshift-ku\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:42Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-4vwms_openshift-ovn-kubernetes(2f511ece-25eb-465f-b85c-9e5f3f886c21)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0644c7db28c7d26aa8d69cd5b9fbcff4d4c06b96df113137dea22a72d9fecd85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6da4c07be4d507a6f20bd8b42be6e6ea4712a956f57ef164036cfded7142a87e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6da4c07be4d507a6f20bd8b42be6e6ea4712a956f57ef164036cfded7142a87e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4vwms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:45Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:45 crc kubenswrapper[4824]: I1006 09:58:45.867182 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lszvg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0cf70ab4-4cfe-41c5-8db7-035451bafcfa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fkbf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fkbf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lszvg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:45Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:45 crc kubenswrapper[4824]: I1006 09:58:45.881740 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:45 crc kubenswrapper[4824]: I1006 09:58:45.881798 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:45 crc kubenswrapper[4824]: I1006 09:58:45.881818 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:45 crc kubenswrapper[4824]: I1006 09:58:45.881850 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:45 crc kubenswrapper[4824]: I1006 09:58:45.881869 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:45Z","lastTransitionTime":"2025-10-06T09:58:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:45 crc kubenswrapper[4824]: I1006 09:58:45.888732 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b665cda3-da4e-482e-8248-023abec382f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9235c742d8c45bf667f32979c2891fece50f4ee1bd0c7cf6a753ed2d2b9ab30e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b2a920dbc506df2426232990e31fbb60d255c00be54941826a0f094af4a92f0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d516f278f362d7c45e7149de82fa835997ac198f1165c04352910d51fb674d69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c820d8bea16cff0d463ce68d2f2acb72d7a7c73372757ff104e14fcec1025a05\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b8cdf54407cc95783919eba8e92f093bbc09eac5e908c4ba0685629948c71bd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-06T09:58:27Z\\\",\\\"message\\\":\\\"arting RequestHeaderAuthRequestController\\\\nI1006 09:58:27.156944 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1006 09:58:27.156783 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1006 09:58:27.157213 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-432254888/tls.crt::/tmp/serving-cert-432254888/tls.key\\\\\\\"\\\\nI1006 09:58:27.157407 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1006 09:58:27.157443 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1006 09:58:27.157410 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-432254888/tls.crt::/tmp/serving-cert-432254888/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759744705\\\\\\\\\\\\\\\" (2025-10-06 09:58:25 +0000 UTC to 2025-11-05 09:58:26 +0000 UTC (now=2025-10-06 09:58:27.157390076 +0000 UTC))\\\\\\\"\\\\nI1006 09:58:27.157632 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1006 09:58:27.158158 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759744707\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759744707\\\\\\\\\\\\\\\" (2025-10-06 08:58:26 +0000 UTC to 2026-10-06 08:58:26 +0000 UTC (now=2025-10-06 09:58:27.158090663 +0000 UTC))\\\\\\\"\\\\nI1006 09:58:27.158220 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1006 09:58:27.158270 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1006 09:58:27.158310 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nF1006 09:58:27.158617 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://413822c99ba2866521177cce4f1894aef17c160aebb2662da91824a04827d510\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e37e8c7400f7fd14a5c5390a5985e3d8d277e9c5ca60d5e9e9757d07d1778570\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e37e8c7400f7fd14a5c5390a5985e3d8d277e9c5ca60d5e9e9757d07d1778570\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:45Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:45 crc kubenswrapper[4824]: I1006 09:58:45.907613 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d739433-0a38-42fe-b884-b8433011967e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://464e279c5903770c34d80f31ed231b194d5628b39055c4058908039d10690059\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://efa1b62728466ef7aac0e745230348ce920eb9ec9b07fb2cd5a276150dac13fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fedc40f0217626c56d155fc7e78e0592556fef9464db2aaf09c5facc486aad8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78fac300cbe2ac52f5788306b8fc0011f6ba3a22c3546974e70998f2d394987a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:45Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:45 crc kubenswrapper[4824]: I1006 09:58:45.933271 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:45Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:45 crc kubenswrapper[4824]: I1006 09:58:45.956231 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-szn8q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b65abf9f-4e88-4571-960d-3ca997d9c344\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://664bf42e1324a9c9529f36f08d61e26999b3386bdd1fac77851fb7e7411e8934\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49s5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-szn8q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:45Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:45 crc kubenswrapper[4824]: I1006 09:58:45.985903 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:45 crc kubenswrapper[4824]: I1006 09:58:45.985997 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:45 crc kubenswrapper[4824]: I1006 09:58:45.986014 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:45 crc kubenswrapper[4824]: I1006 09:58:45.986042 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:45 crc kubenswrapper[4824]: I1006 09:58:45.986061 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:45Z","lastTransitionTime":"2025-10-06T09:58:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:45 crc kubenswrapper[4824]: I1006 09:58:45.986433 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-7c22q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7eaaa68-d971-4a1c-b6c5-551ded8ff27f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f5150c8c6cfc65ecdcd0e7bd52199dd60e2472c96874f103b49919c81b5e5f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53336c7d3c509b4e022708998aedc5dee9fdaf7136ae3ee69b4181394c3a22fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53336c7d3c509b4e022708998aedc5dee9fdaf7136ae3ee69b4181394c3a22fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://652ab0a13a2bb421872edda3f7188dce8e296247052550c937f8447fea206756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://652ab0a13a2bb421872edda3f7188dce8e296247052550c937f8447fea206756\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3514dec0386bb8c016a7f555d7191ea420d5058b0ee98f97d965dd857c5376e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3514dec0386bb8c016a7f555d7191ea420d5058b0ee98f97d965dd857c5376e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b035e0bf34e3429400ac8942124a3e9834231899487d03d15609fe5624747a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b035e0bf34e3429400ac8942124a3e9834231899487d03d15609fe5624747a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6743e027a11864ad3f433d110e60451fb36d49791e03b22d437b70f89d056ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6743e027a11864ad3f433d110e60451fb36d49791e03b22d437b70f89d056ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8ea48b6886fd39a1c5e5b29ca1fffc2bd4552d61dc621eb2fba66fb464cf120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8ea48b6886fd39a1c5e5b29ca1fffc2bd4552d61dc621eb2fba66fb464cf120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-7c22q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:45Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:46 crc kubenswrapper[4824]: I1006 09:58:46.019480 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2035278cf272827b39a45aa41ecd5b1e41aafc1d9eb8bccd853660fb6e587159\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff3152fceadf36a55578f587623f37d748f90f2afeda5702a05504b372d399be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:46Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:46 crc kubenswrapper[4824]: I1006 09:58:46.040867 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xcgwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74317f75-9ebc-4327-b88c-598c2241240f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8def09a126f9bb910affad9b7fb32303b960add9fbfd5bb709d59b4aeba7d245\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v6sg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xcgwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:46Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:46 crc kubenswrapper[4824]: I1006 09:58:46.074507 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1140c8c-93f1-4fce-8c06-a69561ff8a6d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5462cb77b7d61aed7d4b4be2fbc6ba1451d267a66afd6dede2f259c15230e7c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gq49w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5512be36013baeca611ee92645a1d339da237b70d63aac43334beafe24513d44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gq49w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-khgzw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:46Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:46 crc kubenswrapper[4824]: I1006 09:58:46.088532 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:46 crc kubenswrapper[4824]: I1006 09:58:46.088581 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:46 crc kubenswrapper[4824]: I1006 09:58:46.088589 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:46 crc kubenswrapper[4824]: I1006 09:58:46.088608 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:46 crc kubenswrapper[4824]: I1006 09:58:46.088619 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:46Z","lastTransitionTime":"2025-10-06T09:58:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:46 crc kubenswrapper[4824]: I1006 09:58:46.093558 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab36ed34d6ca02b59043c46e151019620ce31733209d072c21394a79fe36be4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:46Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:46 crc kubenswrapper[4824]: I1006 09:58:46.105285 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5rx6l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0c320fc-94d9-4d82-81ee-ccccfd125efc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f8be7bad4e9644494dfb42380c7f5481e1039a1d27ad379c623515948d87392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njtbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5rx6l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:46Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:46 crc kubenswrapper[4824]: I1006 09:58:46.113123 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 09:58:46 crc kubenswrapper[4824]: E1006 09:58:46.113303 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-06 09:59:02.1132721 +0000 UTC m=+51.477695111 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 09:58:46 crc kubenswrapper[4824]: I1006 09:58:46.113408 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 09:58:46 crc kubenswrapper[4824]: I1006 09:58:46.113555 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 09:58:46 crc kubenswrapper[4824]: E1006 09:58:46.113620 4824 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 06 09:58:46 crc kubenswrapper[4824]: E1006 09:58:46.113672 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-06 09:59:02.113662469 +0000 UTC m=+51.478085530 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 06 09:58:46 crc kubenswrapper[4824]: E1006 09:58:46.113811 4824 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 06 09:58:46 crc kubenswrapper[4824]: E1006 09:58:46.113914 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-06 09:59:02.113880914 +0000 UTC m=+51.478303925 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 06 09:58:46 crc kubenswrapper[4824]: I1006 09:58:46.128188 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f9011f6673310dffea3acd7bdc2a56f2af91a87641631490fdc63da610f047a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:46Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:46 crc kubenswrapper[4824]: I1006 09:58:46.145351 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:46Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:46 crc kubenswrapper[4824]: I1006 09:58:46.165027 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:46Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:46 crc kubenswrapper[4824]: I1006 09:58:46.178767 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lszvg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0cf70ab4-4cfe-41c5-8db7-035451bafcfa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7b80b5b375b3b32132501ee0d661e06e199654c696ac76ad56e7ba96918cc77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fkbf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://519653c6b241cd89ef3ece003667d420a7a55cdf513f4577a69e14ade5fb4c76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fkbf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lszvg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:46Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:46 crc kubenswrapper[4824]: I1006 09:58:46.191564 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:46 crc kubenswrapper[4824]: I1006 09:58:46.191604 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:46 crc kubenswrapper[4824]: I1006 09:58:46.191617 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:46 crc kubenswrapper[4824]: I1006 09:58:46.191636 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:46 crc kubenswrapper[4824]: I1006 09:58:46.191647 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:46Z","lastTransitionTime":"2025-10-06T09:58:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:46 crc kubenswrapper[4824]: I1006 09:58:46.197368 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b665cda3-da4e-482e-8248-023abec382f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9235c742d8c45bf667f32979c2891fece50f4ee1bd0c7cf6a753ed2d2b9ab30e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b2a920dbc506df2426232990e31fbb60d255c00be54941826a0f094af4a92f0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d516f278f362d7c45e7149de82fa835997ac198f1165c04352910d51fb674d69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c820d8bea16cff0d463ce68d2f2acb72d7a7c73372757ff104e14fcec1025a05\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b8cdf54407cc95783919eba8e92f093bbc09eac5e908c4ba0685629948c71bd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-06T09:58:27Z\\\",\\\"message\\\":\\\"arting RequestHeaderAuthRequestController\\\\nI1006 09:58:27.156944 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1006 09:58:27.156783 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1006 09:58:27.157213 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-432254888/tls.crt::/tmp/serving-cert-432254888/tls.key\\\\\\\"\\\\nI1006 09:58:27.157407 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1006 09:58:27.157443 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1006 09:58:27.157410 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-432254888/tls.crt::/tmp/serving-cert-432254888/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759744705\\\\\\\\\\\\\\\" (2025-10-06 09:58:25 +0000 UTC to 2025-11-05 09:58:26 +0000 UTC (now=2025-10-06 09:58:27.157390076 +0000 UTC))\\\\\\\"\\\\nI1006 09:58:27.157632 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1006 09:58:27.158158 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759744707\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759744707\\\\\\\\\\\\\\\" (2025-10-06 08:58:26 +0000 UTC to 2026-10-06 08:58:26 +0000 UTC (now=2025-10-06 09:58:27.158090663 +0000 UTC))\\\\\\\"\\\\nI1006 09:58:27.158220 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1006 09:58:27.158270 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1006 09:58:27.158310 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nF1006 09:58:27.158617 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://413822c99ba2866521177cce4f1894aef17c160aebb2662da91824a04827d510\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e37e8c7400f7fd14a5c5390a5985e3d8d277e9c5ca60d5e9e9757d07d1778570\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e37e8c7400f7fd14a5c5390a5985e3d8d277e9c5ca60d5e9e9757d07d1778570\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:46Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:46 crc kubenswrapper[4824]: I1006 09:58:46.209075 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-gnw94"] Oct 06 09:58:46 crc kubenswrapper[4824]: I1006 09:58:46.210238 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gnw94" Oct 06 09:58:46 crc kubenswrapper[4824]: E1006 09:58:46.210364 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gnw94" podUID="7cf1d355-64cb-48a6-acfd-7d258b7afa10" Oct 06 09:58:46 crc kubenswrapper[4824]: I1006 09:58:46.215213 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 09:58:46 crc kubenswrapper[4824]: I1006 09:58:46.215367 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 09:58:46 crc kubenswrapper[4824]: E1006 09:58:46.215386 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 06 09:58:46 crc kubenswrapper[4824]: I1006 09:58:46.215380 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d739433-0a38-42fe-b884-b8433011967e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://464e279c5903770c34d80f31ed231b194d5628b39055c4058908039d10690059\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://efa1b62728466ef7aac0e745230348ce920eb9ec9b07fb2cd5a276150dac13fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fedc40f0217626c56d155fc7e78e0592556fef9464db2aaf09c5facc486aad8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78fac300cbe2ac52f5788306b8fc0011f6ba3a22c3546974e70998f2d394987a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:46Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:46 crc kubenswrapper[4824]: E1006 09:58:46.215480 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 06 09:58:46 crc kubenswrapper[4824]: E1006 09:58:46.215660 4824 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 06 09:58:46 crc kubenswrapper[4824]: E1006 09:58:46.215675 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 06 09:58:46 crc kubenswrapper[4824]: E1006 09:58:46.215731 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 06 09:58:46 crc kubenswrapper[4824]: E1006 09:58:46.215753 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-06 09:59:02.215725229 +0000 UTC m=+51.580148140 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 06 09:58:46 crc kubenswrapper[4824]: E1006 09:58:46.215761 4824 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 06 09:58:46 crc kubenswrapper[4824]: E1006 09:58:46.215876 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-06 09:59:02.215840512 +0000 UTC m=+51.580263523 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 06 09:58:46 crc kubenswrapper[4824]: I1006 09:58:46.232629 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:46Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:46 crc kubenswrapper[4824]: I1006 09:58:46.248954 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-szn8q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b65abf9f-4e88-4571-960d-3ca997d9c344\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://664bf42e1324a9c9529f36f08d61e26999b3386bdd1fac77851fb7e7411e8934\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49s5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-szn8q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:46Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:46 crc kubenswrapper[4824]: I1006 09:58:46.273557 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 09:58:46 crc kubenswrapper[4824]: I1006 09:58:46.273565 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 09:58:46 crc kubenswrapper[4824]: E1006 09:58:46.273745 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 06 09:58:46 crc kubenswrapper[4824]: I1006 09:58:46.273861 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 09:58:46 crc kubenswrapper[4824]: E1006 09:58:46.274092 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 06 09:58:46 crc kubenswrapper[4824]: E1006 09:58:46.274341 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 06 09:58:46 crc kubenswrapper[4824]: I1006 09:58:46.279697 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f511ece-25eb-465f-b85c-9e5f3f886c21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10bfed3e523b7422ce1e981f98166eca78bd12dd4a7cf0634915e1f4ecc3de2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43199f86ad4a1f02bc57a926cd96b2c3c0e2ea46fb36a3f495668706be4c8041\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d6064ad3bcdafebb4c66e7d91f183cc787e9ce13660545c71a722b3ff04ce75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6213164d366243c2139d5971041d6f076cf38fd0cce4c0ca5495111963487ec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa6d59fac987f60a4340ab35d84aa059f2b7085a06362e7207a79702d17eee6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09638fbc9d163ad0b63fa57d24ea5298207ad1f55bab8671be311960be2627a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91a256538722530a7679c5bededd29cff892d03964bd6c71f81d33989b0efdd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91a256538722530a7679c5bededd29cff892d03964bd6c71f81d33989b0efdd9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-06T09:58:43Z\\\",\\\"message\\\":\\\" to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:43Z is after 2025-08-24T17:21:41Z]\\\\nI1006 09:58:43.725317 6283 obj_retry.go:365] Adding new object: *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI1006 09:58:43.725319 6283 services_controller.go:356] Processing sync for service openshift-marketplace/marketplace-operator-metrics for network=default\\\\nI1006 09:58:43.725320 6283 lb_config.go:1031] Cluster endpoints for openshift-machine-config-operator/machine-config-controller for network=default are: map[]\\\\nI1006 09:58:43.725336 6283 ovn.go:134] Ensuring zone local for Pod openshift-kube-controller-manager/kube-controller-manager-crc in node crc\\\\nI1006 09:58:43.725345 6283 obj_retry.go:386] Retry successful for *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc after 0 failed attempt(s)\\\\nI1006 09:58:43.725352 6283 default_network_controller.go:776] Recording success event on pod openshift-ku\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:42Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-4vwms_openshift-ovn-kubernetes(2f511ece-25eb-465f-b85c-9e5f3f886c21)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0644c7db28c7d26aa8d69cd5b9fbcff4d4c06b96df113137dea22a72d9fecd85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6da4c07be4d507a6f20bd8b42be6e6ea4712a956f57ef164036cfded7142a87e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6da4c07be4d507a6f20bd8b42be6e6ea4712a956f57ef164036cfded7142a87e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4vwms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:46Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:46 crc kubenswrapper[4824]: I1006 09:58:46.295223 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:46 crc kubenswrapper[4824]: I1006 09:58:46.295291 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:46 crc kubenswrapper[4824]: I1006 09:58:46.295312 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:46 crc kubenswrapper[4824]: I1006 09:58:46.295342 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:46 crc kubenswrapper[4824]: I1006 09:58:46.295366 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:46Z","lastTransitionTime":"2025-10-06T09:58:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:46 crc kubenswrapper[4824]: I1006 09:58:46.304698 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-7c22q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7eaaa68-d971-4a1c-b6c5-551ded8ff27f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f5150c8c6cfc65ecdcd0e7bd52199dd60e2472c96874f103b49919c81b5e5f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53336c7d3c509b4e022708998aedc5dee9fdaf7136ae3ee69b4181394c3a22fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53336c7d3c509b4e022708998aedc5dee9fdaf7136ae3ee69b4181394c3a22fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://652ab0a13a2bb421872edda3f7188dce8e296247052550c937f8447fea206756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://652ab0a13a2bb421872edda3f7188dce8e296247052550c937f8447fea206756\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3514dec0386bb8c016a7f555d7191ea420d5058b0ee98f97d965dd857c5376e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3514dec0386bb8c016a7f555d7191ea420d5058b0ee98f97d965dd857c5376e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b035e0bf34e3429400ac8942124a3e9834231899487d03d15609fe5624747a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b035e0bf34e3429400ac8942124a3e9834231899487d03d15609fe5624747a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6743e027a11864ad3f433d110e60451fb36d49791e03b22d437b70f89d056ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6743e027a11864ad3f433d110e60451fb36d49791e03b22d437b70f89d056ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8ea48b6886fd39a1c5e5b29ca1fffc2bd4552d61dc621eb2fba66fb464cf120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8ea48b6886fd39a1c5e5b29ca1fffc2bd4552d61dc621eb2fba66fb464cf120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-7c22q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:46Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:46 crc kubenswrapper[4824]: I1006 09:58:46.317005 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7cf1d355-64cb-48a6-acfd-7d258b7afa10-metrics-certs\") pod \"network-metrics-daemon-gnw94\" (UID: \"7cf1d355-64cb-48a6-acfd-7d258b7afa10\") " pod="openshift-multus/network-metrics-daemon-gnw94" Oct 06 09:58:46 crc kubenswrapper[4824]: I1006 09:58:46.317158 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-psbq9\" (UniqueName: \"kubernetes.io/projected/7cf1d355-64cb-48a6-acfd-7d258b7afa10-kube-api-access-psbq9\") pod \"network-metrics-daemon-gnw94\" (UID: \"7cf1d355-64cb-48a6-acfd-7d258b7afa10\") " pod="openshift-multus/network-metrics-daemon-gnw94" Oct 06 09:58:46 crc kubenswrapper[4824]: I1006 09:58:46.323840 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2035278cf272827b39a45aa41ecd5b1e41aafc1d9eb8bccd853660fb6e587159\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff3152fceadf36a55578f587623f37d748f90f2afeda5702a05504b372d399be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:46Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:46 crc kubenswrapper[4824]: I1006 09:58:46.341894 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xcgwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74317f75-9ebc-4327-b88c-598c2241240f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8def09a126f9bb910affad9b7fb32303b960add9fbfd5bb709d59b4aeba7d245\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v6sg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xcgwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:46Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:46 crc kubenswrapper[4824]: I1006 09:58:46.370932 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1140c8c-93f1-4fce-8c06-a69561ff8a6d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5462cb77b7d61aed7d4b4be2fbc6ba1451d267a66afd6dede2f259c15230e7c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gq49w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5512be36013baeca611ee92645a1d339da237b70d63aac43334beafe24513d44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gq49w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-khgzw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:46Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:46 crc kubenswrapper[4824]: I1006 09:58:46.393267 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:46Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:46 crc kubenswrapper[4824]: I1006 09:58:46.398487 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:46 crc kubenswrapper[4824]: I1006 09:58:46.398579 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:46 crc kubenswrapper[4824]: I1006 09:58:46.398604 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:46 crc kubenswrapper[4824]: I1006 09:58:46.398640 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:46 crc kubenswrapper[4824]: I1006 09:58:46.398661 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:46Z","lastTransitionTime":"2025-10-06T09:58:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:46 crc kubenswrapper[4824]: I1006 09:58:46.413892 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:46Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:46 crc kubenswrapper[4824]: I1006 09:58:46.418361 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-psbq9\" (UniqueName: \"kubernetes.io/projected/7cf1d355-64cb-48a6-acfd-7d258b7afa10-kube-api-access-psbq9\") pod \"network-metrics-daemon-gnw94\" (UID: \"7cf1d355-64cb-48a6-acfd-7d258b7afa10\") " pod="openshift-multus/network-metrics-daemon-gnw94" Oct 06 09:58:46 crc kubenswrapper[4824]: I1006 09:58:46.418426 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7cf1d355-64cb-48a6-acfd-7d258b7afa10-metrics-certs\") pod \"network-metrics-daemon-gnw94\" (UID: \"7cf1d355-64cb-48a6-acfd-7d258b7afa10\") " pod="openshift-multus/network-metrics-daemon-gnw94" Oct 06 09:58:46 crc kubenswrapper[4824]: E1006 09:58:46.418593 4824 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 06 09:58:46 crc kubenswrapper[4824]: E1006 09:58:46.418665 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7cf1d355-64cb-48a6-acfd-7d258b7afa10-metrics-certs podName:7cf1d355-64cb-48a6-acfd-7d258b7afa10 nodeName:}" failed. No retries permitted until 2025-10-06 09:58:46.91863915 +0000 UTC m=+36.283062051 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/7cf1d355-64cb-48a6-acfd-7d258b7afa10-metrics-certs") pod "network-metrics-daemon-gnw94" (UID: "7cf1d355-64cb-48a6-acfd-7d258b7afa10") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 06 09:58:46 crc kubenswrapper[4824]: I1006 09:58:46.432604 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab36ed34d6ca02b59043c46e151019620ce31733209d072c21394a79fe36be4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:46Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:46 crc kubenswrapper[4824]: I1006 09:58:46.437556 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-psbq9\" (UniqueName: \"kubernetes.io/projected/7cf1d355-64cb-48a6-acfd-7d258b7afa10-kube-api-access-psbq9\") pod \"network-metrics-daemon-gnw94\" (UID: \"7cf1d355-64cb-48a6-acfd-7d258b7afa10\") " pod="openshift-multus/network-metrics-daemon-gnw94" Oct 06 09:58:46 crc kubenswrapper[4824]: I1006 09:58:46.447107 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5rx6l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0c320fc-94d9-4d82-81ee-ccccfd125efc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f8be7bad4e9644494dfb42380c7f5481e1039a1d27ad379c623515948d87392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njtbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5rx6l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:46Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:46 crc kubenswrapper[4824]: I1006 09:58:46.469887 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f9011f6673310dffea3acd7bdc2a56f2af91a87641631490fdc63da610f047a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:46Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:46 crc kubenswrapper[4824]: I1006 09:58:46.491059 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-szn8q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b65abf9f-4e88-4571-960d-3ca997d9c344\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://664bf42e1324a9c9529f36f08d61e26999b3386bdd1fac77851fb7e7411e8934\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49s5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-szn8q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:46Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:46 crc kubenswrapper[4824]: I1006 09:58:46.503412 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:46 crc kubenswrapper[4824]: I1006 09:58:46.503481 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:46 crc kubenswrapper[4824]: I1006 09:58:46.503502 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:46 crc kubenswrapper[4824]: I1006 09:58:46.503533 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:46 crc kubenswrapper[4824]: I1006 09:58:46.503555 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:46Z","lastTransitionTime":"2025-10-06T09:58:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:46 crc kubenswrapper[4824]: I1006 09:58:46.516175 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f511ece-25eb-465f-b85c-9e5f3f886c21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10bfed3e523b7422ce1e981f98166eca78bd12dd4a7cf0634915e1f4ecc3de2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43199f86ad4a1f02bc57a926cd96b2c3c0e2ea46fb36a3f495668706be4c8041\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d6064ad3bcdafebb4c66e7d91f183cc787e9ce13660545c71a722b3ff04ce75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6213164d366243c2139d5971041d6f076cf38fd0cce4c0ca5495111963487ec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa6d59fac987f60a4340ab35d84aa059f2b7085a06362e7207a79702d17eee6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09638fbc9d163ad0b63fa57d24ea5298207ad1f55bab8671be311960be2627a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91a256538722530a7679c5bededd29cff892d03964bd6c71f81d33989b0efdd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91a256538722530a7679c5bededd29cff892d03964bd6c71f81d33989b0efdd9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-06T09:58:43Z\\\",\\\"message\\\":\\\" to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:43Z is after 2025-08-24T17:21:41Z]\\\\nI1006 09:58:43.725317 6283 obj_retry.go:365] Adding new object: *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI1006 09:58:43.725319 6283 services_controller.go:356] Processing sync for service openshift-marketplace/marketplace-operator-metrics for network=default\\\\nI1006 09:58:43.725320 6283 lb_config.go:1031] Cluster endpoints for openshift-machine-config-operator/machine-config-controller for network=default are: map[]\\\\nI1006 09:58:43.725336 6283 ovn.go:134] Ensuring zone local for Pod openshift-kube-controller-manager/kube-controller-manager-crc in node crc\\\\nI1006 09:58:43.725345 6283 obj_retry.go:386] Retry successful for *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc after 0 failed attempt(s)\\\\nI1006 09:58:43.725352 6283 default_network_controller.go:776] Recording success event on pod openshift-ku\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:42Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-4vwms_openshift-ovn-kubernetes(2f511ece-25eb-465f-b85c-9e5f3f886c21)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0644c7db28c7d26aa8d69cd5b9fbcff4d4c06b96df113137dea22a72d9fecd85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6da4c07be4d507a6f20bd8b42be6e6ea4712a956f57ef164036cfded7142a87e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6da4c07be4d507a6f20bd8b42be6e6ea4712a956f57ef164036cfded7142a87e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4vwms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:46Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:46 crc kubenswrapper[4824]: I1006 09:58:46.535863 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lszvg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0cf70ab4-4cfe-41c5-8db7-035451bafcfa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7b80b5b375b3b32132501ee0d661e06e199654c696ac76ad56e7ba96918cc77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fkbf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://519653c6b241cd89ef3ece003667d420a7a55cdf513f4577a69e14ade5fb4c76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fkbf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lszvg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:46Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:46 crc kubenswrapper[4824]: I1006 09:58:46.551091 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gnw94" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7cf1d355-64cb-48a6-acfd-7d258b7afa10\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-psbq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-psbq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:46Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gnw94\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:46Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:46 crc kubenswrapper[4824]: I1006 09:58:46.571126 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b665cda3-da4e-482e-8248-023abec382f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9235c742d8c45bf667f32979c2891fece50f4ee1bd0c7cf6a753ed2d2b9ab30e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b2a920dbc506df2426232990e31fbb60d255c00be54941826a0f094af4a92f0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d516f278f362d7c45e7149de82fa835997ac198f1165c04352910d51fb674d69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c820d8bea16cff0d463ce68d2f2acb72d7a7c73372757ff104e14fcec1025a05\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b8cdf54407cc95783919eba8e92f093bbc09eac5e908c4ba0685629948c71bd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-06T09:58:27Z\\\",\\\"message\\\":\\\"arting RequestHeaderAuthRequestController\\\\nI1006 09:58:27.156944 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1006 09:58:27.156783 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1006 09:58:27.157213 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-432254888/tls.crt::/tmp/serving-cert-432254888/tls.key\\\\\\\"\\\\nI1006 09:58:27.157407 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1006 09:58:27.157443 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1006 09:58:27.157410 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-432254888/tls.crt::/tmp/serving-cert-432254888/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759744705\\\\\\\\\\\\\\\" (2025-10-06 09:58:25 +0000 UTC to 2025-11-05 09:58:26 +0000 UTC (now=2025-10-06 09:58:27.157390076 +0000 UTC))\\\\\\\"\\\\nI1006 09:58:27.157632 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1006 09:58:27.158158 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759744707\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759744707\\\\\\\\\\\\\\\" (2025-10-06 08:58:26 +0000 UTC to 2026-10-06 08:58:26 +0000 UTC (now=2025-10-06 09:58:27.158090663 +0000 UTC))\\\\\\\"\\\\nI1006 09:58:27.158220 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1006 09:58:27.158270 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1006 09:58:27.158310 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nF1006 09:58:27.158617 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://413822c99ba2866521177cce4f1894aef17c160aebb2662da91824a04827d510\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e37e8c7400f7fd14a5c5390a5985e3d8d277e9c5ca60d5e9e9757d07d1778570\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e37e8c7400f7fd14a5c5390a5985e3d8d277e9c5ca60d5e9e9757d07d1778570\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:46Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:46 crc kubenswrapper[4824]: I1006 09:58:46.589801 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d739433-0a38-42fe-b884-b8433011967e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://464e279c5903770c34d80f31ed231b194d5628b39055c4058908039d10690059\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://efa1b62728466ef7aac0e745230348ce920eb9ec9b07fb2cd5a276150dac13fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fedc40f0217626c56d155fc7e78e0592556fef9464db2aaf09c5facc486aad8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78fac300cbe2ac52f5788306b8fc0011f6ba3a22c3546974e70998f2d394987a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:46Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:46 crc kubenswrapper[4824]: I1006 09:58:46.604025 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:46Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:46 crc kubenswrapper[4824]: I1006 09:58:46.606761 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:46 crc kubenswrapper[4824]: I1006 09:58:46.606825 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:46 crc kubenswrapper[4824]: I1006 09:58:46.606845 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:46 crc kubenswrapper[4824]: I1006 09:58:46.606872 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:46 crc kubenswrapper[4824]: I1006 09:58:46.606894 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:46Z","lastTransitionTime":"2025-10-06T09:58:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:46 crc kubenswrapper[4824]: I1006 09:58:46.629072 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-7c22q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7eaaa68-d971-4a1c-b6c5-551ded8ff27f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f5150c8c6cfc65ecdcd0e7bd52199dd60e2472c96874f103b49919c81b5e5f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53336c7d3c509b4e022708998aedc5dee9fdaf7136ae3ee69b4181394c3a22fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53336c7d3c509b4e022708998aedc5dee9fdaf7136ae3ee69b4181394c3a22fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://652ab0a13a2bb421872edda3f7188dce8e296247052550c937f8447fea206756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://652ab0a13a2bb421872edda3f7188dce8e296247052550c937f8447fea206756\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3514dec0386bb8c016a7f555d7191ea420d5058b0ee98f97d965dd857c5376e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3514dec0386bb8c016a7f555d7191ea420d5058b0ee98f97d965dd857c5376e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b035e0bf34e3429400ac8942124a3e9834231899487d03d15609fe5624747a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b035e0bf34e3429400ac8942124a3e9834231899487d03d15609fe5624747a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6743e027a11864ad3f433d110e60451fb36d49791e03b22d437b70f89d056ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6743e027a11864ad3f433d110e60451fb36d49791e03b22d437b70f89d056ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8ea48b6886fd39a1c5e5b29ca1fffc2bd4552d61dc621eb2fba66fb464cf120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8ea48b6886fd39a1c5e5b29ca1fffc2bd4552d61dc621eb2fba66fb464cf120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-7c22q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:46Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:46 crc kubenswrapper[4824]: I1006 09:58:46.710691 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:46 crc kubenswrapper[4824]: I1006 09:58:46.710744 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:46 crc kubenswrapper[4824]: I1006 09:58:46.710757 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:46 crc kubenswrapper[4824]: I1006 09:58:46.710780 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:46 crc kubenswrapper[4824]: I1006 09:58:46.710797 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:46Z","lastTransitionTime":"2025-10-06T09:58:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:46 crc kubenswrapper[4824]: I1006 09:58:46.814450 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:46 crc kubenswrapper[4824]: I1006 09:58:46.814530 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:46 crc kubenswrapper[4824]: I1006 09:58:46.814554 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:46 crc kubenswrapper[4824]: I1006 09:58:46.814592 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:46 crc kubenswrapper[4824]: I1006 09:58:46.814619 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:46Z","lastTransitionTime":"2025-10-06T09:58:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:46 crc kubenswrapper[4824]: I1006 09:58:46.898532 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:46 crc kubenswrapper[4824]: I1006 09:58:46.898678 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:46 crc kubenswrapper[4824]: I1006 09:58:46.898709 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:46 crc kubenswrapper[4824]: I1006 09:58:46.898747 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:46 crc kubenswrapper[4824]: I1006 09:58:46.898771 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:46Z","lastTransitionTime":"2025-10-06T09:58:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:46 crc kubenswrapper[4824]: E1006 09:58:46.916822 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T09:58:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T09:58:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T09:58:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T09:58:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"740c2e7e-c94c-411f-b080-eb42c9aecdca\\\",\\\"systemUUID\\\":\\\"77448683-227c-4c6b-a334-eb212350758a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:46Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:46 crc kubenswrapper[4824]: I1006 09:58:46.920701 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" Oct 06 09:58:46 crc kubenswrapper[4824]: I1006 09:58:46.922200 4824 scope.go:117] "RemoveContainer" containerID="91a256538722530a7679c5bededd29cff892d03964bd6c71f81d33989b0efdd9" Oct 06 09:58:46 crc kubenswrapper[4824]: I1006 09:58:46.922412 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:46 crc kubenswrapper[4824]: I1006 09:58:46.922462 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:46 crc kubenswrapper[4824]: I1006 09:58:46.922475 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:46 crc kubenswrapper[4824]: I1006 09:58:46.922493 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:46 crc kubenswrapper[4824]: E1006 09:58:46.922487 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-4vwms_openshift-ovn-kubernetes(2f511ece-25eb-465f-b85c-9e5f3f886c21)\"" pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" podUID="2f511ece-25eb-465f-b85c-9e5f3f886c21" Oct 06 09:58:46 crc kubenswrapper[4824]: I1006 09:58:46.922508 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:46Z","lastTransitionTime":"2025-10-06T09:58:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:46 crc kubenswrapper[4824]: I1006 09:58:46.924672 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7cf1d355-64cb-48a6-acfd-7d258b7afa10-metrics-certs\") pod \"network-metrics-daemon-gnw94\" (UID: \"7cf1d355-64cb-48a6-acfd-7d258b7afa10\") " pod="openshift-multus/network-metrics-daemon-gnw94" Oct 06 09:58:46 crc kubenswrapper[4824]: E1006 09:58:46.924868 4824 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 06 09:58:46 crc kubenswrapper[4824]: E1006 09:58:46.924947 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7cf1d355-64cb-48a6-acfd-7d258b7afa10-metrics-certs podName:7cf1d355-64cb-48a6-acfd-7d258b7afa10 nodeName:}" failed. No retries permitted until 2025-10-06 09:58:47.924926311 +0000 UTC m=+37.289349262 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/7cf1d355-64cb-48a6-acfd-7d258b7afa10-metrics-certs") pod "network-metrics-daemon-gnw94" (UID: "7cf1d355-64cb-48a6-acfd-7d258b7afa10") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 06 09:58:46 crc kubenswrapper[4824]: E1006 09:58:46.938952 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T09:58:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T09:58:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T09:58:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T09:58:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"740c2e7e-c94c-411f-b080-eb42c9aecdca\\\",\\\"systemUUID\\\":\\\"77448683-227c-4c6b-a334-eb212350758a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:46Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:46 crc kubenswrapper[4824]: I1006 09:58:46.944475 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:46 crc kubenswrapper[4824]: I1006 09:58:46.944546 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:46 crc kubenswrapper[4824]: I1006 09:58:46.944565 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:46 crc kubenswrapper[4824]: I1006 09:58:46.944592 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:46 crc kubenswrapper[4824]: I1006 09:58:46.944614 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:46Z","lastTransitionTime":"2025-10-06T09:58:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:46 crc kubenswrapper[4824]: E1006 09:58:46.963787 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T09:58:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T09:58:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T09:58:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T09:58:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"740c2e7e-c94c-411f-b080-eb42c9aecdca\\\",\\\"systemUUID\\\":\\\"77448683-227c-4c6b-a334-eb212350758a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:46Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:46 crc kubenswrapper[4824]: I1006 09:58:46.969141 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:46 crc kubenswrapper[4824]: I1006 09:58:46.969176 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:46 crc kubenswrapper[4824]: I1006 09:58:46.969187 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:46 crc kubenswrapper[4824]: I1006 09:58:46.969203 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:46 crc kubenswrapper[4824]: I1006 09:58:46.969214 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:46Z","lastTransitionTime":"2025-10-06T09:58:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:46 crc kubenswrapper[4824]: E1006 09:58:46.989045 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T09:58:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T09:58:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T09:58:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T09:58:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"740c2e7e-c94c-411f-b080-eb42c9aecdca\\\",\\\"systemUUID\\\":\\\"77448683-227c-4c6b-a334-eb212350758a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:46Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:46 crc kubenswrapper[4824]: I1006 09:58:46.995443 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:46 crc kubenswrapper[4824]: I1006 09:58:46.995495 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:46 crc kubenswrapper[4824]: I1006 09:58:46.995508 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:46 crc kubenswrapper[4824]: I1006 09:58:46.995531 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:46 crc kubenswrapper[4824]: I1006 09:58:46.995546 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:46Z","lastTransitionTime":"2025-10-06T09:58:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:47 crc kubenswrapper[4824]: E1006 09:58:47.013169 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T09:58:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T09:58:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T09:58:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T09:58:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"740c2e7e-c94c-411f-b080-eb42c9aecdca\\\",\\\"systemUUID\\\":\\\"77448683-227c-4c6b-a334-eb212350758a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:47Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:47 crc kubenswrapper[4824]: E1006 09:58:47.013494 4824 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 06 09:58:47 crc kubenswrapper[4824]: I1006 09:58:47.015579 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:47 crc kubenswrapper[4824]: I1006 09:58:47.015628 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:47 crc kubenswrapper[4824]: I1006 09:58:47.015646 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:47 crc kubenswrapper[4824]: I1006 09:58:47.015671 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:47 crc kubenswrapper[4824]: I1006 09:58:47.015692 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:47Z","lastTransitionTime":"2025-10-06T09:58:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:47 crc kubenswrapper[4824]: I1006 09:58:47.119521 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:47 crc kubenswrapper[4824]: I1006 09:58:47.119592 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:47 crc kubenswrapper[4824]: I1006 09:58:47.119604 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:47 crc kubenswrapper[4824]: I1006 09:58:47.119625 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:47 crc kubenswrapper[4824]: I1006 09:58:47.119637 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:47Z","lastTransitionTime":"2025-10-06T09:58:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:47 crc kubenswrapper[4824]: I1006 09:58:47.223312 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:47 crc kubenswrapper[4824]: I1006 09:58:47.223389 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:47 crc kubenswrapper[4824]: I1006 09:58:47.223409 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:47 crc kubenswrapper[4824]: I1006 09:58:47.223441 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:47 crc kubenswrapper[4824]: I1006 09:58:47.223464 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:47Z","lastTransitionTime":"2025-10-06T09:58:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:47 crc kubenswrapper[4824]: I1006 09:58:47.326364 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:47 crc kubenswrapper[4824]: I1006 09:58:47.326456 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:47 crc kubenswrapper[4824]: I1006 09:58:47.326477 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:47 crc kubenswrapper[4824]: I1006 09:58:47.326910 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:47 crc kubenswrapper[4824]: I1006 09:58:47.327162 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:47Z","lastTransitionTime":"2025-10-06T09:58:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:47 crc kubenswrapper[4824]: I1006 09:58:47.432943 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:47 crc kubenswrapper[4824]: I1006 09:58:47.433043 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:47 crc kubenswrapper[4824]: I1006 09:58:47.433063 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:47 crc kubenswrapper[4824]: I1006 09:58:47.433093 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:47 crc kubenswrapper[4824]: I1006 09:58:47.433115 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:47Z","lastTransitionTime":"2025-10-06T09:58:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:47 crc kubenswrapper[4824]: I1006 09:58:47.536605 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:47 crc kubenswrapper[4824]: I1006 09:58:47.536693 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:47 crc kubenswrapper[4824]: I1006 09:58:47.536712 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:47 crc kubenswrapper[4824]: I1006 09:58:47.536741 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:47 crc kubenswrapper[4824]: I1006 09:58:47.536761 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:47Z","lastTransitionTime":"2025-10-06T09:58:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:47 crc kubenswrapper[4824]: I1006 09:58:47.640278 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:47 crc kubenswrapper[4824]: I1006 09:58:47.640338 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:47 crc kubenswrapper[4824]: I1006 09:58:47.640356 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:47 crc kubenswrapper[4824]: I1006 09:58:47.640385 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:47 crc kubenswrapper[4824]: I1006 09:58:47.640404 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:47Z","lastTransitionTime":"2025-10-06T09:58:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:47 crc kubenswrapper[4824]: I1006 09:58:47.743413 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:47 crc kubenswrapper[4824]: I1006 09:58:47.743473 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:47 crc kubenswrapper[4824]: I1006 09:58:47.743493 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:47 crc kubenswrapper[4824]: I1006 09:58:47.743526 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:47 crc kubenswrapper[4824]: I1006 09:58:47.743548 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:47Z","lastTransitionTime":"2025-10-06T09:58:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:47 crc kubenswrapper[4824]: I1006 09:58:47.847245 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:47 crc kubenswrapper[4824]: I1006 09:58:47.847339 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:47 crc kubenswrapper[4824]: I1006 09:58:47.847361 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:47 crc kubenswrapper[4824]: I1006 09:58:47.847389 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:47 crc kubenswrapper[4824]: I1006 09:58:47.847408 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:47Z","lastTransitionTime":"2025-10-06T09:58:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:47 crc kubenswrapper[4824]: I1006 09:58:47.936391 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7cf1d355-64cb-48a6-acfd-7d258b7afa10-metrics-certs\") pod \"network-metrics-daemon-gnw94\" (UID: \"7cf1d355-64cb-48a6-acfd-7d258b7afa10\") " pod="openshift-multus/network-metrics-daemon-gnw94" Oct 06 09:58:47 crc kubenswrapper[4824]: E1006 09:58:47.936686 4824 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 06 09:58:47 crc kubenswrapper[4824]: E1006 09:58:47.936801 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7cf1d355-64cb-48a6-acfd-7d258b7afa10-metrics-certs podName:7cf1d355-64cb-48a6-acfd-7d258b7afa10 nodeName:}" failed. No retries permitted until 2025-10-06 09:58:49.936764206 +0000 UTC m=+39.301187287 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/7cf1d355-64cb-48a6-acfd-7d258b7afa10-metrics-certs") pod "network-metrics-daemon-gnw94" (UID: "7cf1d355-64cb-48a6-acfd-7d258b7afa10") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 06 09:58:47 crc kubenswrapper[4824]: I1006 09:58:47.950635 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:47 crc kubenswrapper[4824]: I1006 09:58:47.950705 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:47 crc kubenswrapper[4824]: I1006 09:58:47.950734 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:47 crc kubenswrapper[4824]: I1006 09:58:47.950775 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:47 crc kubenswrapper[4824]: I1006 09:58:47.950806 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:47Z","lastTransitionTime":"2025-10-06T09:58:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:48 crc kubenswrapper[4824]: I1006 09:58:48.054871 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:48 crc kubenswrapper[4824]: I1006 09:58:48.054970 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:48 crc kubenswrapper[4824]: I1006 09:58:48.055019 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:48 crc kubenswrapper[4824]: I1006 09:58:48.055046 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:48 crc kubenswrapper[4824]: I1006 09:58:48.055066 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:48Z","lastTransitionTime":"2025-10-06T09:58:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:48 crc kubenswrapper[4824]: I1006 09:58:48.158459 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:48 crc kubenswrapper[4824]: I1006 09:58:48.158528 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:48 crc kubenswrapper[4824]: I1006 09:58:48.158552 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:48 crc kubenswrapper[4824]: I1006 09:58:48.158583 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:48 crc kubenswrapper[4824]: I1006 09:58:48.158604 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:48Z","lastTransitionTime":"2025-10-06T09:58:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:48 crc kubenswrapper[4824]: I1006 09:58:48.262634 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:48 crc kubenswrapper[4824]: I1006 09:58:48.262690 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:48 crc kubenswrapper[4824]: I1006 09:58:48.262710 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:48 crc kubenswrapper[4824]: I1006 09:58:48.262730 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:48 crc kubenswrapper[4824]: I1006 09:58:48.262742 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:48Z","lastTransitionTime":"2025-10-06T09:58:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:48 crc kubenswrapper[4824]: I1006 09:58:48.273219 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 09:58:48 crc kubenswrapper[4824]: I1006 09:58:48.273222 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gnw94" Oct 06 09:58:48 crc kubenswrapper[4824]: I1006 09:58:48.273302 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 09:58:48 crc kubenswrapper[4824]: I1006 09:58:48.273369 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 09:58:48 crc kubenswrapper[4824]: E1006 09:58:48.273482 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 06 09:58:48 crc kubenswrapper[4824]: E1006 09:58:48.273636 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 06 09:58:48 crc kubenswrapper[4824]: E1006 09:58:48.273804 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gnw94" podUID="7cf1d355-64cb-48a6-acfd-7d258b7afa10" Oct 06 09:58:48 crc kubenswrapper[4824]: E1006 09:58:48.273928 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 06 09:58:48 crc kubenswrapper[4824]: I1006 09:58:48.366094 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:48 crc kubenswrapper[4824]: I1006 09:58:48.366159 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:48 crc kubenswrapper[4824]: I1006 09:58:48.366183 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:48 crc kubenswrapper[4824]: I1006 09:58:48.366251 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:48 crc kubenswrapper[4824]: I1006 09:58:48.366271 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:48Z","lastTransitionTime":"2025-10-06T09:58:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:48 crc kubenswrapper[4824]: I1006 09:58:48.469689 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:48 crc kubenswrapper[4824]: I1006 09:58:48.469754 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:48 crc kubenswrapper[4824]: I1006 09:58:48.469776 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:48 crc kubenswrapper[4824]: I1006 09:58:48.469808 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:48 crc kubenswrapper[4824]: I1006 09:58:48.469834 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:48Z","lastTransitionTime":"2025-10-06T09:58:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:48 crc kubenswrapper[4824]: I1006 09:58:48.573144 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:48 crc kubenswrapper[4824]: I1006 09:58:48.573198 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:48 crc kubenswrapper[4824]: I1006 09:58:48.573218 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:48 crc kubenswrapper[4824]: I1006 09:58:48.573246 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:48 crc kubenswrapper[4824]: I1006 09:58:48.573266 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:48Z","lastTransitionTime":"2025-10-06T09:58:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:48 crc kubenswrapper[4824]: I1006 09:58:48.676322 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:48 crc kubenswrapper[4824]: I1006 09:58:48.676400 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:48 crc kubenswrapper[4824]: I1006 09:58:48.676414 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:48 crc kubenswrapper[4824]: I1006 09:58:48.676437 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:48 crc kubenswrapper[4824]: I1006 09:58:48.676452 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:48Z","lastTransitionTime":"2025-10-06T09:58:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:48 crc kubenswrapper[4824]: I1006 09:58:48.779423 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:48 crc kubenswrapper[4824]: I1006 09:58:48.779522 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:48 crc kubenswrapper[4824]: I1006 09:58:48.779549 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:48 crc kubenswrapper[4824]: I1006 09:58:48.779584 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:48 crc kubenswrapper[4824]: I1006 09:58:48.779609 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:48Z","lastTransitionTime":"2025-10-06T09:58:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:48 crc kubenswrapper[4824]: I1006 09:58:48.882712 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:48 crc kubenswrapper[4824]: I1006 09:58:48.882791 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:48 crc kubenswrapper[4824]: I1006 09:58:48.882809 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:48 crc kubenswrapper[4824]: I1006 09:58:48.882840 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:48 crc kubenswrapper[4824]: I1006 09:58:48.882860 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:48Z","lastTransitionTime":"2025-10-06T09:58:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:48 crc kubenswrapper[4824]: I1006 09:58:48.986149 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:48 crc kubenswrapper[4824]: I1006 09:58:48.986212 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:48 crc kubenswrapper[4824]: I1006 09:58:48.986232 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:48 crc kubenswrapper[4824]: I1006 09:58:48.986258 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:48 crc kubenswrapper[4824]: I1006 09:58:48.986276 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:48Z","lastTransitionTime":"2025-10-06T09:58:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:49 crc kubenswrapper[4824]: I1006 09:58:49.089279 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:49 crc kubenswrapper[4824]: I1006 09:58:49.089355 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:49 crc kubenswrapper[4824]: I1006 09:58:49.089375 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:49 crc kubenswrapper[4824]: I1006 09:58:49.089413 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:49 crc kubenswrapper[4824]: I1006 09:58:49.089434 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:49Z","lastTransitionTime":"2025-10-06T09:58:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:49 crc kubenswrapper[4824]: I1006 09:58:49.192533 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:49 crc kubenswrapper[4824]: I1006 09:58:49.192608 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:49 crc kubenswrapper[4824]: I1006 09:58:49.192629 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:49 crc kubenswrapper[4824]: I1006 09:58:49.192658 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:49 crc kubenswrapper[4824]: I1006 09:58:49.192679 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:49Z","lastTransitionTime":"2025-10-06T09:58:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:49 crc kubenswrapper[4824]: I1006 09:58:49.295167 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:49 crc kubenswrapper[4824]: I1006 09:58:49.295234 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:49 crc kubenswrapper[4824]: I1006 09:58:49.295246 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:49 crc kubenswrapper[4824]: I1006 09:58:49.295337 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:49 crc kubenswrapper[4824]: I1006 09:58:49.295354 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:49Z","lastTransitionTime":"2025-10-06T09:58:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:49 crc kubenswrapper[4824]: I1006 09:58:49.399423 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:49 crc kubenswrapper[4824]: I1006 09:58:49.399476 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:49 crc kubenswrapper[4824]: I1006 09:58:49.399488 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:49 crc kubenswrapper[4824]: I1006 09:58:49.399510 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:49 crc kubenswrapper[4824]: I1006 09:58:49.399521 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:49Z","lastTransitionTime":"2025-10-06T09:58:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:49 crc kubenswrapper[4824]: I1006 09:58:49.502393 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:49 crc kubenswrapper[4824]: I1006 09:58:49.502463 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:49 crc kubenswrapper[4824]: I1006 09:58:49.502487 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:49 crc kubenswrapper[4824]: I1006 09:58:49.502511 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:49 crc kubenswrapper[4824]: I1006 09:58:49.502532 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:49Z","lastTransitionTime":"2025-10-06T09:58:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:49 crc kubenswrapper[4824]: I1006 09:58:49.605735 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:49 crc kubenswrapper[4824]: I1006 09:58:49.605846 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:49 crc kubenswrapper[4824]: I1006 09:58:49.605868 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:49 crc kubenswrapper[4824]: I1006 09:58:49.605899 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:49 crc kubenswrapper[4824]: I1006 09:58:49.605919 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:49Z","lastTransitionTime":"2025-10-06T09:58:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:49 crc kubenswrapper[4824]: I1006 09:58:49.709193 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:49 crc kubenswrapper[4824]: I1006 09:58:49.709268 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:49 crc kubenswrapper[4824]: I1006 09:58:49.709282 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:49 crc kubenswrapper[4824]: I1006 09:58:49.709501 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:49 crc kubenswrapper[4824]: I1006 09:58:49.709515 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:49Z","lastTransitionTime":"2025-10-06T09:58:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:49 crc kubenswrapper[4824]: I1006 09:58:49.813145 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:49 crc kubenswrapper[4824]: I1006 09:58:49.813231 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:49 crc kubenswrapper[4824]: I1006 09:58:49.813254 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:49 crc kubenswrapper[4824]: I1006 09:58:49.813286 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:49 crc kubenswrapper[4824]: I1006 09:58:49.813307 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:49Z","lastTransitionTime":"2025-10-06T09:58:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:49 crc kubenswrapper[4824]: I1006 09:58:49.917033 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:49 crc kubenswrapper[4824]: I1006 09:58:49.917110 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:49 crc kubenswrapper[4824]: I1006 09:58:49.917127 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:49 crc kubenswrapper[4824]: I1006 09:58:49.917154 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:49 crc kubenswrapper[4824]: I1006 09:58:49.917174 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:49Z","lastTransitionTime":"2025-10-06T09:58:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:49 crc kubenswrapper[4824]: I1006 09:58:49.963112 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7cf1d355-64cb-48a6-acfd-7d258b7afa10-metrics-certs\") pod \"network-metrics-daemon-gnw94\" (UID: \"7cf1d355-64cb-48a6-acfd-7d258b7afa10\") " pod="openshift-multus/network-metrics-daemon-gnw94" Oct 06 09:58:49 crc kubenswrapper[4824]: E1006 09:58:49.963427 4824 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 06 09:58:49 crc kubenswrapper[4824]: E1006 09:58:49.963560 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7cf1d355-64cb-48a6-acfd-7d258b7afa10-metrics-certs podName:7cf1d355-64cb-48a6-acfd-7d258b7afa10 nodeName:}" failed. No retries permitted until 2025-10-06 09:58:53.963518301 +0000 UTC m=+43.327941352 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/7cf1d355-64cb-48a6-acfd-7d258b7afa10-metrics-certs") pod "network-metrics-daemon-gnw94" (UID: "7cf1d355-64cb-48a6-acfd-7d258b7afa10") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 06 09:58:50 crc kubenswrapper[4824]: I1006 09:58:50.020945 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:50 crc kubenswrapper[4824]: I1006 09:58:50.021052 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:50 crc kubenswrapper[4824]: I1006 09:58:50.021073 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:50 crc kubenswrapper[4824]: I1006 09:58:50.021101 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:50 crc kubenswrapper[4824]: I1006 09:58:50.021120 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:50Z","lastTransitionTime":"2025-10-06T09:58:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:50 crc kubenswrapper[4824]: I1006 09:58:50.124426 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:50 crc kubenswrapper[4824]: I1006 09:58:50.124491 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:50 crc kubenswrapper[4824]: I1006 09:58:50.124504 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:50 crc kubenswrapper[4824]: I1006 09:58:50.124527 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:50 crc kubenswrapper[4824]: I1006 09:58:50.124544 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:50Z","lastTransitionTime":"2025-10-06T09:58:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:50 crc kubenswrapper[4824]: I1006 09:58:50.227760 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:50 crc kubenswrapper[4824]: I1006 09:58:50.227810 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:50 crc kubenswrapper[4824]: I1006 09:58:50.227824 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:50 crc kubenswrapper[4824]: I1006 09:58:50.227845 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:50 crc kubenswrapper[4824]: I1006 09:58:50.227858 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:50Z","lastTransitionTime":"2025-10-06T09:58:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:50 crc kubenswrapper[4824]: I1006 09:58:50.273350 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 09:58:50 crc kubenswrapper[4824]: I1006 09:58:50.273392 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 09:58:50 crc kubenswrapper[4824]: I1006 09:58:50.273454 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gnw94" Oct 06 09:58:50 crc kubenswrapper[4824]: I1006 09:58:50.273367 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 09:58:50 crc kubenswrapper[4824]: E1006 09:58:50.273508 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 06 09:58:50 crc kubenswrapper[4824]: E1006 09:58:50.273763 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gnw94" podUID="7cf1d355-64cb-48a6-acfd-7d258b7afa10" Oct 06 09:58:50 crc kubenswrapper[4824]: E1006 09:58:50.273857 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 06 09:58:50 crc kubenswrapper[4824]: E1006 09:58:50.273944 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 06 09:58:50 crc kubenswrapper[4824]: I1006 09:58:50.332088 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:50 crc kubenswrapper[4824]: I1006 09:58:50.332143 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:50 crc kubenswrapper[4824]: I1006 09:58:50.332155 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:50 crc kubenswrapper[4824]: I1006 09:58:50.332176 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:50 crc kubenswrapper[4824]: I1006 09:58:50.332197 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:50Z","lastTransitionTime":"2025-10-06T09:58:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:50 crc kubenswrapper[4824]: I1006 09:58:50.435395 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:50 crc kubenswrapper[4824]: I1006 09:58:50.435444 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:50 crc kubenswrapper[4824]: I1006 09:58:50.435458 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:50 crc kubenswrapper[4824]: I1006 09:58:50.435477 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:50 crc kubenswrapper[4824]: I1006 09:58:50.435491 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:50Z","lastTransitionTime":"2025-10-06T09:58:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:50 crc kubenswrapper[4824]: I1006 09:58:50.538171 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:50 crc kubenswrapper[4824]: I1006 09:58:50.538268 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:50 crc kubenswrapper[4824]: I1006 09:58:50.538291 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:50 crc kubenswrapper[4824]: I1006 09:58:50.538328 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:50 crc kubenswrapper[4824]: I1006 09:58:50.538357 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:50Z","lastTransitionTime":"2025-10-06T09:58:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:50 crc kubenswrapper[4824]: I1006 09:58:50.641887 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:50 crc kubenswrapper[4824]: I1006 09:58:50.641945 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:50 crc kubenswrapper[4824]: I1006 09:58:50.641958 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:50 crc kubenswrapper[4824]: I1006 09:58:50.642004 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:50 crc kubenswrapper[4824]: I1006 09:58:50.642019 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:50Z","lastTransitionTime":"2025-10-06T09:58:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:50 crc kubenswrapper[4824]: I1006 09:58:50.745373 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:50 crc kubenswrapper[4824]: I1006 09:58:50.745420 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:50 crc kubenswrapper[4824]: I1006 09:58:50.745433 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:50 crc kubenswrapper[4824]: I1006 09:58:50.745450 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:50 crc kubenswrapper[4824]: I1006 09:58:50.745461 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:50Z","lastTransitionTime":"2025-10-06T09:58:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:50 crc kubenswrapper[4824]: I1006 09:58:50.848654 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:50 crc kubenswrapper[4824]: I1006 09:58:50.848721 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:50 crc kubenswrapper[4824]: I1006 09:58:50.848742 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:50 crc kubenswrapper[4824]: I1006 09:58:50.848769 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:50 crc kubenswrapper[4824]: I1006 09:58:50.848791 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:50Z","lastTransitionTime":"2025-10-06T09:58:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:50 crc kubenswrapper[4824]: I1006 09:58:50.951721 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:50 crc kubenswrapper[4824]: I1006 09:58:50.951788 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:50 crc kubenswrapper[4824]: I1006 09:58:50.951807 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:50 crc kubenswrapper[4824]: I1006 09:58:50.951833 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:50 crc kubenswrapper[4824]: I1006 09:58:50.951852 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:50Z","lastTransitionTime":"2025-10-06T09:58:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:51 crc kubenswrapper[4824]: I1006 09:58:51.055304 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:51 crc kubenswrapper[4824]: I1006 09:58:51.055389 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:51 crc kubenswrapper[4824]: I1006 09:58:51.055417 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:51 crc kubenswrapper[4824]: I1006 09:58:51.055452 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:51 crc kubenswrapper[4824]: I1006 09:58:51.055481 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:51Z","lastTransitionTime":"2025-10-06T09:58:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:51 crc kubenswrapper[4824]: I1006 09:58:51.158779 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:51 crc kubenswrapper[4824]: I1006 09:58:51.158854 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:51 crc kubenswrapper[4824]: I1006 09:58:51.158873 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:51 crc kubenswrapper[4824]: I1006 09:58:51.158902 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:51 crc kubenswrapper[4824]: I1006 09:58:51.158923 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:51Z","lastTransitionTime":"2025-10-06T09:58:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:51 crc kubenswrapper[4824]: I1006 09:58:51.261749 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:51 crc kubenswrapper[4824]: I1006 09:58:51.261834 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:51 crc kubenswrapper[4824]: I1006 09:58:51.261860 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:51 crc kubenswrapper[4824]: I1006 09:58:51.261891 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:51 crc kubenswrapper[4824]: I1006 09:58:51.261914 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:51Z","lastTransitionTime":"2025-10-06T09:58:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:51 crc kubenswrapper[4824]: I1006 09:58:51.302958 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-7c22q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7eaaa68-d971-4a1c-b6c5-551ded8ff27f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f5150c8c6cfc65ecdcd0e7bd52199dd60e2472c96874f103b49919c81b5e5f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53336c7d3c509b4e022708998aedc5dee9fdaf7136ae3ee69b4181394c3a22fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53336c7d3c509b4e022708998aedc5dee9fdaf7136ae3ee69b4181394c3a22fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://652ab0a13a2bb421872edda3f7188dce8e296247052550c937f8447fea206756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://652ab0a13a2bb421872edda3f7188dce8e296247052550c937f8447fea206756\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3514dec0386bb8c016a7f555d7191ea420d5058b0ee98f97d965dd857c5376e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3514dec0386bb8c016a7f555d7191ea420d5058b0ee98f97d965dd857c5376e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b035e0bf34e3429400ac8942124a3e9834231899487d03d15609fe5624747a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b035e0bf34e3429400ac8942124a3e9834231899487d03d15609fe5624747a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6743e027a11864ad3f433d110e60451fb36d49791e03b22d437b70f89d056ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6743e027a11864ad3f433d110e60451fb36d49791e03b22d437b70f89d056ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8ea48b6886fd39a1c5e5b29ca1fffc2bd4552d61dc621eb2fba66fb464cf120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8ea48b6886fd39a1c5e5b29ca1fffc2bd4552d61dc621eb2fba66fb464cf120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-7c22q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:51Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:51 crc kubenswrapper[4824]: I1006 09:58:51.324344 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2035278cf272827b39a45aa41ecd5b1e41aafc1d9eb8bccd853660fb6e587159\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff3152fceadf36a55578f587623f37d748f90f2afeda5702a05504b372d399be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:51Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:51 crc kubenswrapper[4824]: I1006 09:58:51.337183 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xcgwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74317f75-9ebc-4327-b88c-598c2241240f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8def09a126f9bb910affad9b7fb32303b960add9fbfd5bb709d59b4aeba7d245\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v6sg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xcgwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:51Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:51 crc kubenswrapper[4824]: I1006 09:58:51.352068 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1140c8c-93f1-4fce-8c06-a69561ff8a6d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5462cb77b7d61aed7d4b4be2fbc6ba1451d267a66afd6dede2f259c15230e7c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gq49w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5512be36013baeca611ee92645a1d339da237b70d63aac43334beafe24513d44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gq49w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-khgzw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:51Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:51 crc kubenswrapper[4824]: I1006 09:58:51.364389 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:51 crc kubenswrapper[4824]: I1006 09:58:51.364459 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:51 crc kubenswrapper[4824]: I1006 09:58:51.364478 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:51 crc kubenswrapper[4824]: I1006 09:58:51.364509 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:51 crc kubenswrapper[4824]: I1006 09:58:51.364530 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:51Z","lastTransitionTime":"2025-10-06T09:58:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:51 crc kubenswrapper[4824]: I1006 09:58:51.376321 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f9011f6673310dffea3acd7bdc2a56f2af91a87641631490fdc63da610f047a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:51Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:51 crc kubenswrapper[4824]: I1006 09:58:51.396061 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:51Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:51 crc kubenswrapper[4824]: I1006 09:58:51.410144 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:51Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:51 crc kubenswrapper[4824]: I1006 09:58:51.425958 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab36ed34d6ca02b59043c46e151019620ce31733209d072c21394a79fe36be4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:51Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:51 crc kubenswrapper[4824]: I1006 09:58:51.437530 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5rx6l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0c320fc-94d9-4d82-81ee-ccccfd125efc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f8be7bad4e9644494dfb42380c7f5481e1039a1d27ad379c623515948d87392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njtbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5rx6l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:51Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:51 crc kubenswrapper[4824]: I1006 09:58:51.457342 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b665cda3-da4e-482e-8248-023abec382f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9235c742d8c45bf667f32979c2891fece50f4ee1bd0c7cf6a753ed2d2b9ab30e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b2a920dbc506df2426232990e31fbb60d255c00be54941826a0f094af4a92f0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d516f278f362d7c45e7149de82fa835997ac198f1165c04352910d51fb674d69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c820d8bea16cff0d463ce68d2f2acb72d7a7c73372757ff104e14fcec1025a05\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b8cdf54407cc95783919eba8e92f093bbc09eac5e908c4ba0685629948c71bd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-06T09:58:27Z\\\",\\\"message\\\":\\\"arting RequestHeaderAuthRequestController\\\\nI1006 09:58:27.156944 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1006 09:58:27.156783 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1006 09:58:27.157213 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-432254888/tls.crt::/tmp/serving-cert-432254888/tls.key\\\\\\\"\\\\nI1006 09:58:27.157407 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1006 09:58:27.157443 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1006 09:58:27.157410 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-432254888/tls.crt::/tmp/serving-cert-432254888/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759744705\\\\\\\\\\\\\\\" (2025-10-06 09:58:25 +0000 UTC to 2025-11-05 09:58:26 +0000 UTC (now=2025-10-06 09:58:27.157390076 +0000 UTC))\\\\\\\"\\\\nI1006 09:58:27.157632 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1006 09:58:27.158158 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759744707\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759744707\\\\\\\\\\\\\\\" (2025-10-06 08:58:26 +0000 UTC to 2026-10-06 08:58:26 +0000 UTC (now=2025-10-06 09:58:27.158090663 +0000 UTC))\\\\\\\"\\\\nI1006 09:58:27.158220 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1006 09:58:27.158270 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1006 09:58:27.158310 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nF1006 09:58:27.158617 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://413822c99ba2866521177cce4f1894aef17c160aebb2662da91824a04827d510\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e37e8c7400f7fd14a5c5390a5985e3d8d277e9c5ca60d5e9e9757d07d1778570\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e37e8c7400f7fd14a5c5390a5985e3d8d277e9c5ca60d5e9e9757d07d1778570\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:51Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:51 crc kubenswrapper[4824]: I1006 09:58:51.467106 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:51 crc kubenswrapper[4824]: I1006 09:58:51.467179 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:51 crc kubenswrapper[4824]: I1006 09:58:51.467198 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:51 crc kubenswrapper[4824]: I1006 09:58:51.467229 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:51 crc kubenswrapper[4824]: I1006 09:58:51.467248 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:51Z","lastTransitionTime":"2025-10-06T09:58:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:51 crc kubenswrapper[4824]: I1006 09:58:51.473254 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d739433-0a38-42fe-b884-b8433011967e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://464e279c5903770c34d80f31ed231b194d5628b39055c4058908039d10690059\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://efa1b62728466ef7aac0e745230348ce920eb9ec9b07fb2cd5a276150dac13fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fedc40f0217626c56d155fc7e78e0592556fef9464db2aaf09c5facc486aad8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78fac300cbe2ac52f5788306b8fc0011f6ba3a22c3546974e70998f2d394987a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:51Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:51 crc kubenswrapper[4824]: I1006 09:58:51.485704 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:51Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:51 crc kubenswrapper[4824]: I1006 09:58:51.500107 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-szn8q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b65abf9f-4e88-4571-960d-3ca997d9c344\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://664bf42e1324a9c9529f36f08d61e26999b3386bdd1fac77851fb7e7411e8934\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49s5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-szn8q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:51Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:51 crc kubenswrapper[4824]: I1006 09:58:51.526962 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f511ece-25eb-465f-b85c-9e5f3f886c21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10bfed3e523b7422ce1e981f98166eca78bd12dd4a7cf0634915e1f4ecc3de2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43199f86ad4a1f02bc57a926cd96b2c3c0e2ea46fb36a3f495668706be4c8041\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d6064ad3bcdafebb4c66e7d91f183cc787e9ce13660545c71a722b3ff04ce75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6213164d366243c2139d5971041d6f076cf38fd0cce4c0ca5495111963487ec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa6d59fac987f60a4340ab35d84aa059f2b7085a06362e7207a79702d17eee6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09638fbc9d163ad0b63fa57d24ea5298207ad1f55bab8671be311960be2627a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91a256538722530a7679c5bededd29cff892d03964bd6c71f81d33989b0efdd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91a256538722530a7679c5bededd29cff892d03964bd6c71f81d33989b0efdd9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-06T09:58:43Z\\\",\\\"message\\\":\\\" to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:43Z is after 2025-08-24T17:21:41Z]\\\\nI1006 09:58:43.725317 6283 obj_retry.go:365] Adding new object: *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI1006 09:58:43.725319 6283 services_controller.go:356] Processing sync for service openshift-marketplace/marketplace-operator-metrics for network=default\\\\nI1006 09:58:43.725320 6283 lb_config.go:1031] Cluster endpoints for openshift-machine-config-operator/machine-config-controller for network=default are: map[]\\\\nI1006 09:58:43.725336 6283 ovn.go:134] Ensuring zone local for Pod openshift-kube-controller-manager/kube-controller-manager-crc in node crc\\\\nI1006 09:58:43.725345 6283 obj_retry.go:386] Retry successful for *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc after 0 failed attempt(s)\\\\nI1006 09:58:43.725352 6283 default_network_controller.go:776] Recording success event on pod openshift-ku\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:42Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-4vwms_openshift-ovn-kubernetes(2f511ece-25eb-465f-b85c-9e5f3f886c21)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0644c7db28c7d26aa8d69cd5b9fbcff4d4c06b96df113137dea22a72d9fecd85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6da4c07be4d507a6f20bd8b42be6e6ea4712a956f57ef164036cfded7142a87e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6da4c07be4d507a6f20bd8b42be6e6ea4712a956f57ef164036cfded7142a87e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4vwms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:51Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:51 crc kubenswrapper[4824]: I1006 09:58:51.543738 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lszvg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0cf70ab4-4cfe-41c5-8db7-035451bafcfa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7b80b5b375b3b32132501ee0d661e06e199654c696ac76ad56e7ba96918cc77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fkbf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://519653c6b241cd89ef3ece003667d420a7a55cdf513f4577a69e14ade5fb4c76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fkbf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lszvg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:51Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:51 crc kubenswrapper[4824]: I1006 09:58:51.559054 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gnw94" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7cf1d355-64cb-48a6-acfd-7d258b7afa10\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-psbq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-psbq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:46Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gnw94\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:51Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:51 crc kubenswrapper[4824]: I1006 09:58:51.570281 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:51 crc kubenswrapper[4824]: I1006 09:58:51.570339 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:51 crc kubenswrapper[4824]: I1006 09:58:51.570359 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:51 crc kubenswrapper[4824]: I1006 09:58:51.570409 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:51 crc kubenswrapper[4824]: I1006 09:58:51.570431 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:51Z","lastTransitionTime":"2025-10-06T09:58:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:51 crc kubenswrapper[4824]: I1006 09:58:51.673605 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:51 crc kubenswrapper[4824]: I1006 09:58:51.673677 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:51 crc kubenswrapper[4824]: I1006 09:58:51.673697 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:51 crc kubenswrapper[4824]: I1006 09:58:51.673726 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:51 crc kubenswrapper[4824]: I1006 09:58:51.673747 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:51Z","lastTransitionTime":"2025-10-06T09:58:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:51 crc kubenswrapper[4824]: I1006 09:58:51.777702 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:51 crc kubenswrapper[4824]: I1006 09:58:51.777759 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:51 crc kubenswrapper[4824]: I1006 09:58:51.777778 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:51 crc kubenswrapper[4824]: I1006 09:58:51.777804 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:51 crc kubenswrapper[4824]: I1006 09:58:51.777822 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:51Z","lastTransitionTime":"2025-10-06T09:58:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:51 crc kubenswrapper[4824]: I1006 09:58:51.880543 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:51 crc kubenswrapper[4824]: I1006 09:58:51.880625 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:51 crc kubenswrapper[4824]: I1006 09:58:51.880647 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:51 crc kubenswrapper[4824]: I1006 09:58:51.880683 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:51 crc kubenswrapper[4824]: I1006 09:58:51.880705 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:51Z","lastTransitionTime":"2025-10-06T09:58:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:51 crc kubenswrapper[4824]: I1006 09:58:51.984446 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:51 crc kubenswrapper[4824]: I1006 09:58:51.984525 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:51 crc kubenswrapper[4824]: I1006 09:58:51.984544 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:51 crc kubenswrapper[4824]: I1006 09:58:51.984574 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:51 crc kubenswrapper[4824]: I1006 09:58:51.984594 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:51Z","lastTransitionTime":"2025-10-06T09:58:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:52 crc kubenswrapper[4824]: I1006 09:58:52.087668 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:52 crc kubenswrapper[4824]: I1006 09:58:52.087753 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:52 crc kubenswrapper[4824]: I1006 09:58:52.087771 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:52 crc kubenswrapper[4824]: I1006 09:58:52.087800 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:52 crc kubenswrapper[4824]: I1006 09:58:52.087819 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:52Z","lastTransitionTime":"2025-10-06T09:58:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:52 crc kubenswrapper[4824]: I1006 09:58:52.192140 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:52 crc kubenswrapper[4824]: I1006 09:58:52.192215 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:52 crc kubenswrapper[4824]: I1006 09:58:52.192234 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:52 crc kubenswrapper[4824]: I1006 09:58:52.192267 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:52 crc kubenswrapper[4824]: I1006 09:58:52.192291 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:52Z","lastTransitionTime":"2025-10-06T09:58:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:52 crc kubenswrapper[4824]: I1006 09:58:52.273648 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 09:58:52 crc kubenswrapper[4824]: I1006 09:58:52.273794 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gnw94" Oct 06 09:58:52 crc kubenswrapper[4824]: I1006 09:58:52.273801 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 09:58:52 crc kubenswrapper[4824]: E1006 09:58:52.273895 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 06 09:58:52 crc kubenswrapper[4824]: I1006 09:58:52.273936 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 09:58:52 crc kubenswrapper[4824]: E1006 09:58:52.274096 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gnw94" podUID="7cf1d355-64cb-48a6-acfd-7d258b7afa10" Oct 06 09:58:52 crc kubenswrapper[4824]: E1006 09:58:52.274264 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 06 09:58:52 crc kubenswrapper[4824]: E1006 09:58:52.274448 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 06 09:58:52 crc kubenswrapper[4824]: I1006 09:58:52.295871 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:52 crc kubenswrapper[4824]: I1006 09:58:52.295925 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:52 crc kubenswrapper[4824]: I1006 09:58:52.295943 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:52 crc kubenswrapper[4824]: I1006 09:58:52.295970 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:52 crc kubenswrapper[4824]: I1006 09:58:52.296016 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:52Z","lastTransitionTime":"2025-10-06T09:58:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:52 crc kubenswrapper[4824]: I1006 09:58:52.400132 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:52 crc kubenswrapper[4824]: I1006 09:58:52.400211 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:52 crc kubenswrapper[4824]: I1006 09:58:52.400229 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:52 crc kubenswrapper[4824]: I1006 09:58:52.400259 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:52 crc kubenswrapper[4824]: I1006 09:58:52.400280 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:52Z","lastTransitionTime":"2025-10-06T09:58:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:52 crc kubenswrapper[4824]: I1006 09:58:52.503454 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:52 crc kubenswrapper[4824]: I1006 09:58:52.503781 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:52 crc kubenswrapper[4824]: I1006 09:58:52.503946 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:52 crc kubenswrapper[4824]: I1006 09:58:52.504184 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:52 crc kubenswrapper[4824]: I1006 09:58:52.504360 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:52Z","lastTransitionTime":"2025-10-06T09:58:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:52 crc kubenswrapper[4824]: I1006 09:58:52.558261 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 06 09:58:52 crc kubenswrapper[4824]: I1006 09:58:52.580814 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2035278cf272827b39a45aa41ecd5b1e41aafc1d9eb8bccd853660fb6e587159\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff3152fceadf36a55578f587623f37d748f90f2afeda5702a05504b372d399be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:52Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:52 crc kubenswrapper[4824]: I1006 09:58:52.598654 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xcgwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74317f75-9ebc-4327-b88c-598c2241240f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8def09a126f9bb910affad9b7fb32303b960add9fbfd5bb709d59b4aeba7d245\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v6sg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xcgwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:52Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:52 crc kubenswrapper[4824]: I1006 09:58:52.607185 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:52 crc kubenswrapper[4824]: I1006 09:58:52.607253 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:52 crc kubenswrapper[4824]: I1006 09:58:52.607270 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:52 crc kubenswrapper[4824]: I1006 09:58:52.607305 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:52 crc kubenswrapper[4824]: I1006 09:58:52.607328 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:52Z","lastTransitionTime":"2025-10-06T09:58:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:52 crc kubenswrapper[4824]: I1006 09:58:52.616686 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1140c8c-93f1-4fce-8c06-a69561ff8a6d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5462cb77b7d61aed7d4b4be2fbc6ba1451d267a66afd6dede2f259c15230e7c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gq49w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5512be36013baeca611ee92645a1d339da237b70d63aac43334beafe24513d44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gq49w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-khgzw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:52Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:52 crc kubenswrapper[4824]: I1006 09:58:52.634162 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:52Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:52 crc kubenswrapper[4824]: I1006 09:58:52.656176 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab36ed34d6ca02b59043c46e151019620ce31733209d072c21394a79fe36be4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:52Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:52 crc kubenswrapper[4824]: I1006 09:58:52.673220 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5rx6l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0c320fc-94d9-4d82-81ee-ccccfd125efc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f8be7bad4e9644494dfb42380c7f5481e1039a1d27ad379c623515948d87392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njtbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5rx6l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:52Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:52 crc kubenswrapper[4824]: I1006 09:58:52.696045 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f9011f6673310dffea3acd7bdc2a56f2af91a87641631490fdc63da610f047a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:52Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:52 crc kubenswrapper[4824]: I1006 09:58:52.711036 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:52 crc kubenswrapper[4824]: I1006 09:58:52.711097 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:52 crc kubenswrapper[4824]: I1006 09:58:52.711115 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:52 crc kubenswrapper[4824]: I1006 09:58:52.711140 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:52 crc kubenswrapper[4824]: I1006 09:58:52.711162 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:52Z","lastTransitionTime":"2025-10-06T09:58:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:52 crc kubenswrapper[4824]: I1006 09:58:52.717700 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:52Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:52 crc kubenswrapper[4824]: I1006 09:58:52.755339 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f511ece-25eb-465f-b85c-9e5f3f886c21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10bfed3e523b7422ce1e981f98166eca78bd12dd4a7cf0634915e1f4ecc3de2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43199f86ad4a1f02bc57a926cd96b2c3c0e2ea46fb36a3f495668706be4c8041\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d6064ad3bcdafebb4c66e7d91f183cc787e9ce13660545c71a722b3ff04ce75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6213164d366243c2139d5971041d6f076cf38fd0cce4c0ca5495111963487ec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa6d59fac987f60a4340ab35d84aa059f2b7085a06362e7207a79702d17eee6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09638fbc9d163ad0b63fa57d24ea5298207ad1f55bab8671be311960be2627a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91a256538722530a7679c5bededd29cff892d03964bd6c71f81d33989b0efdd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91a256538722530a7679c5bededd29cff892d03964bd6c71f81d33989b0efdd9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-06T09:58:43Z\\\",\\\"message\\\":\\\" to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:43Z is after 2025-08-24T17:21:41Z]\\\\nI1006 09:58:43.725317 6283 obj_retry.go:365] Adding new object: *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI1006 09:58:43.725319 6283 services_controller.go:356] Processing sync for service openshift-marketplace/marketplace-operator-metrics for network=default\\\\nI1006 09:58:43.725320 6283 lb_config.go:1031] Cluster endpoints for openshift-machine-config-operator/machine-config-controller for network=default are: map[]\\\\nI1006 09:58:43.725336 6283 ovn.go:134] Ensuring zone local for Pod openshift-kube-controller-manager/kube-controller-manager-crc in node crc\\\\nI1006 09:58:43.725345 6283 obj_retry.go:386] Retry successful for *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc after 0 failed attempt(s)\\\\nI1006 09:58:43.725352 6283 default_network_controller.go:776] Recording success event on pod openshift-ku\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:42Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-4vwms_openshift-ovn-kubernetes(2f511ece-25eb-465f-b85c-9e5f3f886c21)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0644c7db28c7d26aa8d69cd5b9fbcff4d4c06b96df113137dea22a72d9fecd85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6da4c07be4d507a6f20bd8b42be6e6ea4712a956f57ef164036cfded7142a87e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6da4c07be4d507a6f20bd8b42be6e6ea4712a956f57ef164036cfded7142a87e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4vwms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:52Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:52 crc kubenswrapper[4824]: I1006 09:58:52.776333 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lszvg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0cf70ab4-4cfe-41c5-8db7-035451bafcfa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7b80b5b375b3b32132501ee0d661e06e199654c696ac76ad56e7ba96918cc77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fkbf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://519653c6b241cd89ef3ece003667d420a7a55cdf513f4577a69e14ade5fb4c76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fkbf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lszvg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:52Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:52 crc kubenswrapper[4824]: I1006 09:58:52.794952 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gnw94" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7cf1d355-64cb-48a6-acfd-7d258b7afa10\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-psbq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-psbq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:46Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gnw94\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:52Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:52 crc kubenswrapper[4824]: I1006 09:58:52.815244 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:52 crc kubenswrapper[4824]: I1006 09:58:52.815322 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:52 crc kubenswrapper[4824]: I1006 09:58:52.815342 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:52 crc kubenswrapper[4824]: I1006 09:58:52.815373 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:52 crc kubenswrapper[4824]: I1006 09:58:52.815395 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:52Z","lastTransitionTime":"2025-10-06T09:58:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:52 crc kubenswrapper[4824]: I1006 09:58:52.819020 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b665cda3-da4e-482e-8248-023abec382f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9235c742d8c45bf667f32979c2891fece50f4ee1bd0c7cf6a753ed2d2b9ab30e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b2a920dbc506df2426232990e31fbb60d255c00be54941826a0f094af4a92f0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d516f278f362d7c45e7149de82fa835997ac198f1165c04352910d51fb674d69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c820d8bea16cff0d463ce68d2f2acb72d7a7c73372757ff104e14fcec1025a05\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b8cdf54407cc95783919eba8e92f093bbc09eac5e908c4ba0685629948c71bd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-06T09:58:27Z\\\",\\\"message\\\":\\\"arting RequestHeaderAuthRequestController\\\\nI1006 09:58:27.156944 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1006 09:58:27.156783 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1006 09:58:27.157213 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-432254888/tls.crt::/tmp/serving-cert-432254888/tls.key\\\\\\\"\\\\nI1006 09:58:27.157407 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1006 09:58:27.157443 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1006 09:58:27.157410 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-432254888/tls.crt::/tmp/serving-cert-432254888/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759744705\\\\\\\\\\\\\\\" (2025-10-06 09:58:25 +0000 UTC to 2025-11-05 09:58:26 +0000 UTC (now=2025-10-06 09:58:27.157390076 +0000 UTC))\\\\\\\"\\\\nI1006 09:58:27.157632 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1006 09:58:27.158158 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759744707\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759744707\\\\\\\\\\\\\\\" (2025-10-06 08:58:26 +0000 UTC to 2026-10-06 08:58:26 +0000 UTC (now=2025-10-06 09:58:27.158090663 +0000 UTC))\\\\\\\"\\\\nI1006 09:58:27.158220 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1006 09:58:27.158270 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1006 09:58:27.158310 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nF1006 09:58:27.158617 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://413822c99ba2866521177cce4f1894aef17c160aebb2662da91824a04827d510\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e37e8c7400f7fd14a5c5390a5985e3d8d277e9c5ca60d5e9e9757d07d1778570\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e37e8c7400f7fd14a5c5390a5985e3d8d277e9c5ca60d5e9e9757d07d1778570\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:52Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:52 crc kubenswrapper[4824]: I1006 09:58:52.839470 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d739433-0a38-42fe-b884-b8433011967e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://464e279c5903770c34d80f31ed231b194d5628b39055c4058908039d10690059\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://efa1b62728466ef7aac0e745230348ce920eb9ec9b07fb2cd5a276150dac13fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fedc40f0217626c56d155fc7e78e0592556fef9464db2aaf09c5facc486aad8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78fac300cbe2ac52f5788306b8fc0011f6ba3a22c3546974e70998f2d394987a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:52Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:52 crc kubenswrapper[4824]: I1006 09:58:52.859822 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:52Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:52 crc kubenswrapper[4824]: I1006 09:58:52.881553 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-szn8q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b65abf9f-4e88-4571-960d-3ca997d9c344\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://664bf42e1324a9c9529f36f08d61e26999b3386bdd1fac77851fb7e7411e8934\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49s5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-szn8q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:52Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:52 crc kubenswrapper[4824]: I1006 09:58:52.906350 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-7c22q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7eaaa68-d971-4a1c-b6c5-551ded8ff27f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f5150c8c6cfc65ecdcd0e7bd52199dd60e2472c96874f103b49919c81b5e5f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53336c7d3c509b4e022708998aedc5dee9fdaf7136ae3ee69b4181394c3a22fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53336c7d3c509b4e022708998aedc5dee9fdaf7136ae3ee69b4181394c3a22fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://652ab0a13a2bb421872edda3f7188dce8e296247052550c937f8447fea206756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://652ab0a13a2bb421872edda3f7188dce8e296247052550c937f8447fea206756\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3514dec0386bb8c016a7f555d7191ea420d5058b0ee98f97d965dd857c5376e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3514dec0386bb8c016a7f555d7191ea420d5058b0ee98f97d965dd857c5376e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b035e0bf34e3429400ac8942124a3e9834231899487d03d15609fe5624747a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b035e0bf34e3429400ac8942124a3e9834231899487d03d15609fe5624747a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6743e027a11864ad3f433d110e60451fb36d49791e03b22d437b70f89d056ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6743e027a11864ad3f433d110e60451fb36d49791e03b22d437b70f89d056ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8ea48b6886fd39a1c5e5b29ca1fffc2bd4552d61dc621eb2fba66fb464cf120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8ea48b6886fd39a1c5e5b29ca1fffc2bd4552d61dc621eb2fba66fb464cf120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-7c22q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:52Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:52 crc kubenswrapper[4824]: I1006 09:58:52.918754 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:52 crc kubenswrapper[4824]: I1006 09:58:52.918822 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:52 crc kubenswrapper[4824]: I1006 09:58:52.918840 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:52 crc kubenswrapper[4824]: I1006 09:58:52.918866 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:52 crc kubenswrapper[4824]: I1006 09:58:52.918887 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:52Z","lastTransitionTime":"2025-10-06T09:58:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:53 crc kubenswrapper[4824]: I1006 09:58:53.022613 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:53 crc kubenswrapper[4824]: I1006 09:58:53.022671 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:53 crc kubenswrapper[4824]: I1006 09:58:53.022681 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:53 crc kubenswrapper[4824]: I1006 09:58:53.022696 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:53 crc kubenswrapper[4824]: I1006 09:58:53.022707 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:53Z","lastTransitionTime":"2025-10-06T09:58:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:53 crc kubenswrapper[4824]: I1006 09:58:53.126599 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:53 crc kubenswrapper[4824]: I1006 09:58:53.126675 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:53 crc kubenswrapper[4824]: I1006 09:58:53.126692 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:53 crc kubenswrapper[4824]: I1006 09:58:53.126718 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:53 crc kubenswrapper[4824]: I1006 09:58:53.126737 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:53Z","lastTransitionTime":"2025-10-06T09:58:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:53 crc kubenswrapper[4824]: I1006 09:58:53.230642 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:53 crc kubenswrapper[4824]: I1006 09:58:53.230705 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:53 crc kubenswrapper[4824]: I1006 09:58:53.230727 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:53 crc kubenswrapper[4824]: I1006 09:58:53.230758 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:53 crc kubenswrapper[4824]: I1006 09:58:53.230781 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:53Z","lastTransitionTime":"2025-10-06T09:58:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:53 crc kubenswrapper[4824]: I1006 09:58:53.334667 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:53 crc kubenswrapper[4824]: I1006 09:58:53.334727 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:53 crc kubenswrapper[4824]: I1006 09:58:53.334743 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:53 crc kubenswrapper[4824]: I1006 09:58:53.334772 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:53 crc kubenswrapper[4824]: I1006 09:58:53.334791 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:53Z","lastTransitionTime":"2025-10-06T09:58:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:53 crc kubenswrapper[4824]: I1006 09:58:53.438863 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:53 crc kubenswrapper[4824]: I1006 09:58:53.438918 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:53 crc kubenswrapper[4824]: I1006 09:58:53.438936 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:53 crc kubenswrapper[4824]: I1006 09:58:53.438960 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:53 crc kubenswrapper[4824]: I1006 09:58:53.439012 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:53Z","lastTransitionTime":"2025-10-06T09:58:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:53 crc kubenswrapper[4824]: I1006 09:58:53.542303 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:53 crc kubenswrapper[4824]: I1006 09:58:53.542359 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:53 crc kubenswrapper[4824]: I1006 09:58:53.542376 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:53 crc kubenswrapper[4824]: I1006 09:58:53.542400 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:53 crc kubenswrapper[4824]: I1006 09:58:53.542421 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:53Z","lastTransitionTime":"2025-10-06T09:58:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:53 crc kubenswrapper[4824]: I1006 09:58:53.645838 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:53 crc kubenswrapper[4824]: I1006 09:58:53.645902 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:53 crc kubenswrapper[4824]: I1006 09:58:53.645919 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:53 crc kubenswrapper[4824]: I1006 09:58:53.645944 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:53 crc kubenswrapper[4824]: I1006 09:58:53.645965 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:53Z","lastTransitionTime":"2025-10-06T09:58:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:53 crc kubenswrapper[4824]: I1006 09:58:53.750406 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:53 crc kubenswrapper[4824]: I1006 09:58:53.750456 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:53 crc kubenswrapper[4824]: I1006 09:58:53.750466 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:53 crc kubenswrapper[4824]: I1006 09:58:53.750487 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:53 crc kubenswrapper[4824]: I1006 09:58:53.750502 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:53Z","lastTransitionTime":"2025-10-06T09:58:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:53 crc kubenswrapper[4824]: I1006 09:58:53.854297 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:53 crc kubenswrapper[4824]: I1006 09:58:53.854385 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:53 crc kubenswrapper[4824]: I1006 09:58:53.854407 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:53 crc kubenswrapper[4824]: I1006 09:58:53.854438 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:53 crc kubenswrapper[4824]: I1006 09:58:53.854462 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:53Z","lastTransitionTime":"2025-10-06T09:58:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:53 crc kubenswrapper[4824]: I1006 09:58:53.958115 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:53 crc kubenswrapper[4824]: I1006 09:58:53.958173 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:53 crc kubenswrapper[4824]: I1006 09:58:53.958192 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:53 crc kubenswrapper[4824]: I1006 09:58:53.958221 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:53 crc kubenswrapper[4824]: I1006 09:58:53.958240 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:53Z","lastTransitionTime":"2025-10-06T09:58:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:54 crc kubenswrapper[4824]: I1006 09:58:54.007642 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7cf1d355-64cb-48a6-acfd-7d258b7afa10-metrics-certs\") pod \"network-metrics-daemon-gnw94\" (UID: \"7cf1d355-64cb-48a6-acfd-7d258b7afa10\") " pod="openshift-multus/network-metrics-daemon-gnw94" Oct 06 09:58:54 crc kubenswrapper[4824]: E1006 09:58:54.007927 4824 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 06 09:58:54 crc kubenswrapper[4824]: E1006 09:58:54.008073 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7cf1d355-64cb-48a6-acfd-7d258b7afa10-metrics-certs podName:7cf1d355-64cb-48a6-acfd-7d258b7afa10 nodeName:}" failed. No retries permitted until 2025-10-06 09:59:02.008043042 +0000 UTC m=+51.372465933 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/7cf1d355-64cb-48a6-acfd-7d258b7afa10-metrics-certs") pod "network-metrics-daemon-gnw94" (UID: "7cf1d355-64cb-48a6-acfd-7d258b7afa10") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 06 09:58:54 crc kubenswrapper[4824]: I1006 09:58:54.062627 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:54 crc kubenswrapper[4824]: I1006 09:58:54.062698 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:54 crc kubenswrapper[4824]: I1006 09:58:54.062716 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:54 crc kubenswrapper[4824]: I1006 09:58:54.062746 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:54 crc kubenswrapper[4824]: I1006 09:58:54.062765 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:54Z","lastTransitionTime":"2025-10-06T09:58:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:54 crc kubenswrapper[4824]: I1006 09:58:54.166917 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:54 crc kubenswrapper[4824]: I1006 09:58:54.167020 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:54 crc kubenswrapper[4824]: I1006 09:58:54.167044 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:54 crc kubenswrapper[4824]: I1006 09:58:54.167073 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:54 crc kubenswrapper[4824]: I1006 09:58:54.167097 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:54Z","lastTransitionTime":"2025-10-06T09:58:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:54 crc kubenswrapper[4824]: I1006 09:58:54.270421 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:54 crc kubenswrapper[4824]: I1006 09:58:54.270494 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:54 crc kubenswrapper[4824]: I1006 09:58:54.270513 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:54 crc kubenswrapper[4824]: I1006 09:58:54.270545 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:54 crc kubenswrapper[4824]: I1006 09:58:54.270567 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:54Z","lastTransitionTime":"2025-10-06T09:58:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:54 crc kubenswrapper[4824]: I1006 09:58:54.273642 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 09:58:54 crc kubenswrapper[4824]: I1006 09:58:54.273663 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 09:58:54 crc kubenswrapper[4824]: I1006 09:58:54.273721 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gnw94" Oct 06 09:58:54 crc kubenswrapper[4824]: E1006 09:58:54.273819 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 06 09:58:54 crc kubenswrapper[4824]: I1006 09:58:54.273881 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 09:58:54 crc kubenswrapper[4824]: E1006 09:58:54.274068 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 06 09:58:54 crc kubenswrapper[4824]: E1006 09:58:54.274202 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 06 09:58:54 crc kubenswrapper[4824]: E1006 09:58:54.274322 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gnw94" podUID="7cf1d355-64cb-48a6-acfd-7d258b7afa10" Oct 06 09:58:54 crc kubenswrapper[4824]: I1006 09:58:54.374162 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:54 crc kubenswrapper[4824]: I1006 09:58:54.374229 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:54 crc kubenswrapper[4824]: I1006 09:58:54.374247 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:54 crc kubenswrapper[4824]: I1006 09:58:54.374277 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:54 crc kubenswrapper[4824]: I1006 09:58:54.374295 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:54Z","lastTransitionTime":"2025-10-06T09:58:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:54 crc kubenswrapper[4824]: I1006 09:58:54.478590 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:54 crc kubenswrapper[4824]: I1006 09:58:54.478692 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:54 crc kubenswrapper[4824]: I1006 09:58:54.478718 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:54 crc kubenswrapper[4824]: I1006 09:58:54.478755 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:54 crc kubenswrapper[4824]: I1006 09:58:54.478780 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:54Z","lastTransitionTime":"2025-10-06T09:58:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:54 crc kubenswrapper[4824]: I1006 09:58:54.582510 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:54 crc kubenswrapper[4824]: I1006 09:58:54.582618 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:54 crc kubenswrapper[4824]: I1006 09:58:54.582646 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:54 crc kubenswrapper[4824]: I1006 09:58:54.582686 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:54 crc kubenswrapper[4824]: I1006 09:58:54.582712 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:54Z","lastTransitionTime":"2025-10-06T09:58:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:54 crc kubenswrapper[4824]: I1006 09:58:54.686387 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:54 crc kubenswrapper[4824]: I1006 09:58:54.686842 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:54 crc kubenswrapper[4824]: I1006 09:58:54.687023 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:54 crc kubenswrapper[4824]: I1006 09:58:54.687231 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:54 crc kubenswrapper[4824]: I1006 09:58:54.687355 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:54Z","lastTransitionTime":"2025-10-06T09:58:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:54 crc kubenswrapper[4824]: I1006 09:58:54.790695 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:54 crc kubenswrapper[4824]: I1006 09:58:54.790765 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:54 crc kubenswrapper[4824]: I1006 09:58:54.790783 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:54 crc kubenswrapper[4824]: I1006 09:58:54.790810 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:54 crc kubenswrapper[4824]: I1006 09:58:54.790829 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:54Z","lastTransitionTime":"2025-10-06T09:58:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:54 crc kubenswrapper[4824]: I1006 09:58:54.894251 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:54 crc kubenswrapper[4824]: I1006 09:58:54.894331 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:54 crc kubenswrapper[4824]: I1006 09:58:54.894350 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:54 crc kubenswrapper[4824]: I1006 09:58:54.894380 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:54 crc kubenswrapper[4824]: I1006 09:58:54.894401 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:54Z","lastTransitionTime":"2025-10-06T09:58:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:54 crc kubenswrapper[4824]: I1006 09:58:54.997851 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:54 crc kubenswrapper[4824]: I1006 09:58:54.997908 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:54 crc kubenswrapper[4824]: I1006 09:58:54.997925 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:54 crc kubenswrapper[4824]: I1006 09:58:54.997949 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:54 crc kubenswrapper[4824]: I1006 09:58:54.997966 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:54Z","lastTransitionTime":"2025-10-06T09:58:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:55 crc kubenswrapper[4824]: I1006 09:58:55.101013 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:55 crc kubenswrapper[4824]: I1006 09:58:55.101060 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:55 crc kubenswrapper[4824]: I1006 09:58:55.101072 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:55 crc kubenswrapper[4824]: I1006 09:58:55.101090 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:55 crc kubenswrapper[4824]: I1006 09:58:55.101104 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:55Z","lastTransitionTime":"2025-10-06T09:58:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:55 crc kubenswrapper[4824]: I1006 09:58:55.204149 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:55 crc kubenswrapper[4824]: I1006 09:58:55.204214 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:55 crc kubenswrapper[4824]: I1006 09:58:55.204230 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:55 crc kubenswrapper[4824]: I1006 09:58:55.204258 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:55 crc kubenswrapper[4824]: I1006 09:58:55.204278 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:55Z","lastTransitionTime":"2025-10-06T09:58:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:55 crc kubenswrapper[4824]: I1006 09:58:55.307691 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:55 crc kubenswrapper[4824]: I1006 09:58:55.307767 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:55 crc kubenswrapper[4824]: I1006 09:58:55.307793 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:55 crc kubenswrapper[4824]: I1006 09:58:55.307823 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:55 crc kubenswrapper[4824]: I1006 09:58:55.307848 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:55Z","lastTransitionTime":"2025-10-06T09:58:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:55 crc kubenswrapper[4824]: I1006 09:58:55.411845 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:55 crc kubenswrapper[4824]: I1006 09:58:55.411924 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:55 crc kubenswrapper[4824]: I1006 09:58:55.411945 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:55 crc kubenswrapper[4824]: I1006 09:58:55.412015 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:55 crc kubenswrapper[4824]: I1006 09:58:55.412043 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:55Z","lastTransitionTime":"2025-10-06T09:58:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:55 crc kubenswrapper[4824]: I1006 09:58:55.515973 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:55 crc kubenswrapper[4824]: I1006 09:58:55.516069 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:55 crc kubenswrapper[4824]: I1006 09:58:55.516084 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:55 crc kubenswrapper[4824]: I1006 09:58:55.516110 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:55 crc kubenswrapper[4824]: I1006 09:58:55.516123 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:55Z","lastTransitionTime":"2025-10-06T09:58:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:55 crc kubenswrapper[4824]: I1006 09:58:55.619611 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:55 crc kubenswrapper[4824]: I1006 09:58:55.619677 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:55 crc kubenswrapper[4824]: I1006 09:58:55.619697 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:55 crc kubenswrapper[4824]: I1006 09:58:55.619730 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:55 crc kubenswrapper[4824]: I1006 09:58:55.619749 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:55Z","lastTransitionTime":"2025-10-06T09:58:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:55 crc kubenswrapper[4824]: I1006 09:58:55.723707 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:55 crc kubenswrapper[4824]: I1006 09:58:55.723770 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:55 crc kubenswrapper[4824]: I1006 09:58:55.723790 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:55 crc kubenswrapper[4824]: I1006 09:58:55.723816 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:55 crc kubenswrapper[4824]: I1006 09:58:55.723836 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:55Z","lastTransitionTime":"2025-10-06T09:58:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:55 crc kubenswrapper[4824]: I1006 09:58:55.827904 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:55 crc kubenswrapper[4824]: I1006 09:58:55.828047 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:55 crc kubenswrapper[4824]: I1006 09:58:55.828074 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:55 crc kubenswrapper[4824]: I1006 09:58:55.828112 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:55 crc kubenswrapper[4824]: I1006 09:58:55.828140 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:55Z","lastTransitionTime":"2025-10-06T09:58:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:55 crc kubenswrapper[4824]: I1006 09:58:55.931443 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:55 crc kubenswrapper[4824]: I1006 09:58:55.931497 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:55 crc kubenswrapper[4824]: I1006 09:58:55.931508 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:55 crc kubenswrapper[4824]: I1006 09:58:55.931530 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:55 crc kubenswrapper[4824]: I1006 09:58:55.931543 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:55Z","lastTransitionTime":"2025-10-06T09:58:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:56 crc kubenswrapper[4824]: I1006 09:58:56.034621 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:56 crc kubenswrapper[4824]: I1006 09:58:56.034676 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:56 crc kubenswrapper[4824]: I1006 09:58:56.034687 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:56 crc kubenswrapper[4824]: I1006 09:58:56.034707 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:56 crc kubenswrapper[4824]: I1006 09:58:56.034719 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:56Z","lastTransitionTime":"2025-10-06T09:58:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:56 crc kubenswrapper[4824]: I1006 09:58:56.138186 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:56 crc kubenswrapper[4824]: I1006 09:58:56.138237 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:56 crc kubenswrapper[4824]: I1006 09:58:56.138253 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:56 crc kubenswrapper[4824]: I1006 09:58:56.138279 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:56 crc kubenswrapper[4824]: I1006 09:58:56.138297 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:56Z","lastTransitionTime":"2025-10-06T09:58:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:56 crc kubenswrapper[4824]: I1006 09:58:56.241954 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:56 crc kubenswrapper[4824]: I1006 09:58:56.242041 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:56 crc kubenswrapper[4824]: I1006 09:58:56.242059 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:56 crc kubenswrapper[4824]: I1006 09:58:56.242088 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:56 crc kubenswrapper[4824]: I1006 09:58:56.242109 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:56Z","lastTransitionTime":"2025-10-06T09:58:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:56 crc kubenswrapper[4824]: I1006 09:58:56.273571 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 09:58:56 crc kubenswrapper[4824]: I1006 09:58:56.273600 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gnw94" Oct 06 09:58:56 crc kubenswrapper[4824]: I1006 09:58:56.273611 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 09:58:56 crc kubenswrapper[4824]: I1006 09:58:56.273794 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 09:58:56 crc kubenswrapper[4824]: E1006 09:58:56.273872 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 06 09:58:56 crc kubenswrapper[4824]: E1006 09:58:56.273969 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gnw94" podUID="7cf1d355-64cb-48a6-acfd-7d258b7afa10" Oct 06 09:58:56 crc kubenswrapper[4824]: E1006 09:58:56.274130 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 06 09:58:56 crc kubenswrapper[4824]: E1006 09:58:56.274186 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 06 09:58:56 crc kubenswrapper[4824]: I1006 09:58:56.345559 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:56 crc kubenswrapper[4824]: I1006 09:58:56.345618 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:56 crc kubenswrapper[4824]: I1006 09:58:56.345637 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:56 crc kubenswrapper[4824]: I1006 09:58:56.345660 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:56 crc kubenswrapper[4824]: I1006 09:58:56.345679 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:56Z","lastTransitionTime":"2025-10-06T09:58:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:56 crc kubenswrapper[4824]: I1006 09:58:56.449051 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:56 crc kubenswrapper[4824]: I1006 09:58:56.449142 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:56 crc kubenswrapper[4824]: I1006 09:58:56.449162 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:56 crc kubenswrapper[4824]: I1006 09:58:56.449196 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:56 crc kubenswrapper[4824]: I1006 09:58:56.449218 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:56Z","lastTransitionTime":"2025-10-06T09:58:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:56 crc kubenswrapper[4824]: I1006 09:58:56.553280 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:56 crc kubenswrapper[4824]: I1006 09:58:56.553357 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:56 crc kubenswrapper[4824]: I1006 09:58:56.553380 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:56 crc kubenswrapper[4824]: I1006 09:58:56.553412 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:56 crc kubenswrapper[4824]: I1006 09:58:56.553436 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:56Z","lastTransitionTime":"2025-10-06T09:58:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:56 crc kubenswrapper[4824]: I1006 09:58:56.657146 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:56 crc kubenswrapper[4824]: I1006 09:58:56.657205 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:56 crc kubenswrapper[4824]: I1006 09:58:56.657224 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:56 crc kubenswrapper[4824]: I1006 09:58:56.657253 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:56 crc kubenswrapper[4824]: I1006 09:58:56.657272 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:56Z","lastTransitionTime":"2025-10-06T09:58:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:56 crc kubenswrapper[4824]: I1006 09:58:56.760916 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:56 crc kubenswrapper[4824]: I1006 09:58:56.761028 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:56 crc kubenswrapper[4824]: I1006 09:58:56.761054 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:56 crc kubenswrapper[4824]: I1006 09:58:56.761084 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:56 crc kubenswrapper[4824]: I1006 09:58:56.761107 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:56Z","lastTransitionTime":"2025-10-06T09:58:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:56 crc kubenswrapper[4824]: I1006 09:58:56.865243 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:56 crc kubenswrapper[4824]: I1006 09:58:56.865317 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:56 crc kubenswrapper[4824]: I1006 09:58:56.865337 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:56 crc kubenswrapper[4824]: I1006 09:58:56.865367 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:56 crc kubenswrapper[4824]: I1006 09:58:56.865386 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:56Z","lastTransitionTime":"2025-10-06T09:58:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:56 crc kubenswrapper[4824]: I1006 09:58:56.969042 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:56 crc kubenswrapper[4824]: I1006 09:58:56.969108 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:56 crc kubenswrapper[4824]: I1006 09:58:56.969128 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:56 crc kubenswrapper[4824]: I1006 09:58:56.969156 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:56 crc kubenswrapper[4824]: I1006 09:58:56.969174 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:56Z","lastTransitionTime":"2025-10-06T09:58:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:57 crc kubenswrapper[4824]: I1006 09:58:57.032616 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:57 crc kubenswrapper[4824]: I1006 09:58:57.032671 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:57 crc kubenswrapper[4824]: I1006 09:58:57.032689 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:57 crc kubenswrapper[4824]: I1006 09:58:57.032709 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:57 crc kubenswrapper[4824]: I1006 09:58:57.032731 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:57Z","lastTransitionTime":"2025-10-06T09:58:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:57 crc kubenswrapper[4824]: E1006 09:58:57.054356 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T09:58:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T09:58:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T09:58:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T09:58:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"740c2e7e-c94c-411f-b080-eb42c9aecdca\\\",\\\"systemUUID\\\":\\\"77448683-227c-4c6b-a334-eb212350758a\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:57Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:57 crc kubenswrapper[4824]: I1006 09:58:57.060535 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:57 crc kubenswrapper[4824]: I1006 09:58:57.060641 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:57 crc kubenswrapper[4824]: I1006 09:58:57.060704 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:57 crc kubenswrapper[4824]: I1006 09:58:57.060766 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:57 crc kubenswrapper[4824]: I1006 09:58:57.060834 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:57Z","lastTransitionTime":"2025-10-06T09:58:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:57 crc kubenswrapper[4824]: E1006 09:58:57.081177 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T09:58:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T09:58:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T09:58:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T09:58:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"740c2e7e-c94c-411f-b080-eb42c9aecdca\\\",\\\"systemUUID\\\":\\\"77448683-227c-4c6b-a334-eb212350758a\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:57Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:57 crc kubenswrapper[4824]: I1006 09:58:57.086440 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:57 crc kubenswrapper[4824]: I1006 09:58:57.086546 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:57 crc kubenswrapper[4824]: I1006 09:58:57.086627 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:57 crc kubenswrapper[4824]: I1006 09:58:57.086698 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:57 crc kubenswrapper[4824]: I1006 09:58:57.086755 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:57Z","lastTransitionTime":"2025-10-06T09:58:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:57 crc kubenswrapper[4824]: E1006 09:58:57.105175 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T09:58:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T09:58:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T09:58:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T09:58:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"740c2e7e-c94c-411f-b080-eb42c9aecdca\\\",\\\"systemUUID\\\":\\\"77448683-227c-4c6b-a334-eb212350758a\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:57Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:57 crc kubenswrapper[4824]: I1006 09:58:57.111185 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:57 crc kubenswrapper[4824]: I1006 09:58:57.111243 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:57 crc kubenswrapper[4824]: I1006 09:58:57.111261 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:57 crc kubenswrapper[4824]: I1006 09:58:57.111284 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:57 crc kubenswrapper[4824]: I1006 09:58:57.111301 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:57Z","lastTransitionTime":"2025-10-06T09:58:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:57 crc kubenswrapper[4824]: E1006 09:58:57.140202 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T09:58:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T09:58:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T09:58:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T09:58:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"740c2e7e-c94c-411f-b080-eb42c9aecdca\\\",\\\"systemUUID\\\":\\\"77448683-227c-4c6b-a334-eb212350758a\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:57Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:57 crc kubenswrapper[4824]: I1006 09:58:57.145968 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:57 crc kubenswrapper[4824]: I1006 09:58:57.146049 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:57 crc kubenswrapper[4824]: I1006 09:58:57.146072 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:57 crc kubenswrapper[4824]: I1006 09:58:57.146097 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:57 crc kubenswrapper[4824]: I1006 09:58:57.146116 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:57Z","lastTransitionTime":"2025-10-06T09:58:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:57 crc kubenswrapper[4824]: E1006 09:58:57.166788 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T09:58:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T09:58:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T09:58:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T09:58:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"740c2e7e-c94c-411f-b080-eb42c9aecdca\\\",\\\"systemUUID\\\":\\\"77448683-227c-4c6b-a334-eb212350758a\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:57Z is after 2025-08-24T17:21:41Z" Oct 06 09:58:57 crc kubenswrapper[4824]: E1006 09:58:57.167063 4824 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 06 09:58:57 crc kubenswrapper[4824]: I1006 09:58:57.169368 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:57 crc kubenswrapper[4824]: I1006 09:58:57.169484 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:57 crc kubenswrapper[4824]: I1006 09:58:57.169552 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:57 crc kubenswrapper[4824]: I1006 09:58:57.169594 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:57 crc kubenswrapper[4824]: I1006 09:58:57.169615 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:57Z","lastTransitionTime":"2025-10-06T09:58:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:57 crc kubenswrapper[4824]: I1006 09:58:57.273465 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:57 crc kubenswrapper[4824]: I1006 09:58:57.273562 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:57 crc kubenswrapper[4824]: I1006 09:58:57.273584 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:57 crc kubenswrapper[4824]: I1006 09:58:57.273611 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:57 crc kubenswrapper[4824]: I1006 09:58:57.273630 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:57Z","lastTransitionTime":"2025-10-06T09:58:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:57 crc kubenswrapper[4824]: I1006 09:58:57.376647 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:57 crc kubenswrapper[4824]: I1006 09:58:57.376720 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:57 crc kubenswrapper[4824]: I1006 09:58:57.376737 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:57 crc kubenswrapper[4824]: I1006 09:58:57.376765 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:57 crc kubenswrapper[4824]: I1006 09:58:57.376782 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:57Z","lastTransitionTime":"2025-10-06T09:58:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:57 crc kubenswrapper[4824]: I1006 09:58:57.480016 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:57 crc kubenswrapper[4824]: I1006 09:58:57.480075 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:57 crc kubenswrapper[4824]: I1006 09:58:57.480093 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:57 crc kubenswrapper[4824]: I1006 09:58:57.480117 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:57 crc kubenswrapper[4824]: I1006 09:58:57.480131 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:57Z","lastTransitionTime":"2025-10-06T09:58:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:57 crc kubenswrapper[4824]: I1006 09:58:57.583606 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:57 crc kubenswrapper[4824]: I1006 09:58:57.584290 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:57 crc kubenswrapper[4824]: I1006 09:58:57.584320 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:57 crc kubenswrapper[4824]: I1006 09:58:57.584352 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:57 crc kubenswrapper[4824]: I1006 09:58:57.584374 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:57Z","lastTransitionTime":"2025-10-06T09:58:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:57 crc kubenswrapper[4824]: I1006 09:58:57.688087 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:57 crc kubenswrapper[4824]: I1006 09:58:57.688164 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:57 crc kubenswrapper[4824]: I1006 09:58:57.688192 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:57 crc kubenswrapper[4824]: I1006 09:58:57.688225 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:57 crc kubenswrapper[4824]: I1006 09:58:57.688254 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:57Z","lastTransitionTime":"2025-10-06T09:58:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:57 crc kubenswrapper[4824]: I1006 09:58:57.791502 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:57 crc kubenswrapper[4824]: I1006 09:58:57.791609 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:57 crc kubenswrapper[4824]: I1006 09:58:57.791629 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:57 crc kubenswrapper[4824]: I1006 09:58:57.791662 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:57 crc kubenswrapper[4824]: I1006 09:58:57.791684 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:57Z","lastTransitionTime":"2025-10-06T09:58:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:57 crc kubenswrapper[4824]: I1006 09:58:57.895397 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:57 crc kubenswrapper[4824]: I1006 09:58:57.895465 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:57 crc kubenswrapper[4824]: I1006 09:58:57.895483 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:57 crc kubenswrapper[4824]: I1006 09:58:57.895510 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:57 crc kubenswrapper[4824]: I1006 09:58:57.895531 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:57Z","lastTransitionTime":"2025-10-06T09:58:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:57 crc kubenswrapper[4824]: I1006 09:58:57.998919 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:57 crc kubenswrapper[4824]: I1006 09:58:57.998991 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:57 crc kubenswrapper[4824]: I1006 09:58:57.999002 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:57 crc kubenswrapper[4824]: I1006 09:58:57.999022 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:57 crc kubenswrapper[4824]: I1006 09:58:57.999035 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:57Z","lastTransitionTime":"2025-10-06T09:58:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:58 crc kubenswrapper[4824]: I1006 09:58:58.102207 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:58 crc kubenswrapper[4824]: I1006 09:58:58.102268 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:58 crc kubenswrapper[4824]: I1006 09:58:58.102281 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:58 crc kubenswrapper[4824]: I1006 09:58:58.102301 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:58 crc kubenswrapper[4824]: I1006 09:58:58.102316 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:58Z","lastTransitionTime":"2025-10-06T09:58:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:58 crc kubenswrapper[4824]: I1006 09:58:58.205605 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:58 crc kubenswrapper[4824]: I1006 09:58:58.205673 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:58 crc kubenswrapper[4824]: I1006 09:58:58.205696 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:58 crc kubenswrapper[4824]: I1006 09:58:58.205737 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:58 crc kubenswrapper[4824]: I1006 09:58:58.205758 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:58Z","lastTransitionTime":"2025-10-06T09:58:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:58 crc kubenswrapper[4824]: I1006 09:58:58.273454 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 09:58:58 crc kubenswrapper[4824]: I1006 09:58:58.273552 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 09:58:58 crc kubenswrapper[4824]: I1006 09:58:58.273624 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gnw94" Oct 06 09:58:58 crc kubenswrapper[4824]: I1006 09:58:58.273872 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 09:58:58 crc kubenswrapper[4824]: E1006 09:58:58.273866 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 06 09:58:58 crc kubenswrapper[4824]: E1006 09:58:58.274088 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 06 09:58:58 crc kubenswrapper[4824]: E1006 09:58:58.274230 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 06 09:58:58 crc kubenswrapper[4824]: E1006 09:58:58.274409 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gnw94" podUID="7cf1d355-64cb-48a6-acfd-7d258b7afa10" Oct 06 09:58:58 crc kubenswrapper[4824]: I1006 09:58:58.308820 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:58 crc kubenswrapper[4824]: I1006 09:58:58.308874 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:58 crc kubenswrapper[4824]: I1006 09:58:58.308886 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:58 crc kubenswrapper[4824]: I1006 09:58:58.308908 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:58 crc kubenswrapper[4824]: I1006 09:58:58.308921 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:58Z","lastTransitionTime":"2025-10-06T09:58:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:58 crc kubenswrapper[4824]: I1006 09:58:58.412492 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:58 crc kubenswrapper[4824]: I1006 09:58:58.412562 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:58 crc kubenswrapper[4824]: I1006 09:58:58.412611 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:58 crc kubenswrapper[4824]: I1006 09:58:58.412649 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:58 crc kubenswrapper[4824]: I1006 09:58:58.412664 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:58Z","lastTransitionTime":"2025-10-06T09:58:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:58 crc kubenswrapper[4824]: I1006 09:58:58.515461 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:58 crc kubenswrapper[4824]: I1006 09:58:58.515526 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:58 crc kubenswrapper[4824]: I1006 09:58:58.515542 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:58 crc kubenswrapper[4824]: I1006 09:58:58.515560 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:58 crc kubenswrapper[4824]: I1006 09:58:58.515573 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:58Z","lastTransitionTime":"2025-10-06T09:58:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:58 crc kubenswrapper[4824]: I1006 09:58:58.619306 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:58 crc kubenswrapper[4824]: I1006 09:58:58.619362 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:58 crc kubenswrapper[4824]: I1006 09:58:58.619377 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:58 crc kubenswrapper[4824]: I1006 09:58:58.619394 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:58 crc kubenswrapper[4824]: I1006 09:58:58.619404 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:58Z","lastTransitionTime":"2025-10-06T09:58:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:58 crc kubenswrapper[4824]: I1006 09:58:58.722424 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:58 crc kubenswrapper[4824]: I1006 09:58:58.722477 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:58 crc kubenswrapper[4824]: I1006 09:58:58.722492 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:58 crc kubenswrapper[4824]: I1006 09:58:58.722518 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:58 crc kubenswrapper[4824]: I1006 09:58:58.722532 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:58Z","lastTransitionTime":"2025-10-06T09:58:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:58 crc kubenswrapper[4824]: I1006 09:58:58.825418 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:58 crc kubenswrapper[4824]: I1006 09:58:58.825485 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:58 crc kubenswrapper[4824]: I1006 09:58:58.825505 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:58 crc kubenswrapper[4824]: I1006 09:58:58.825536 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:58 crc kubenswrapper[4824]: I1006 09:58:58.825557 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:58Z","lastTransitionTime":"2025-10-06T09:58:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:58 crc kubenswrapper[4824]: I1006 09:58:58.928905 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:58 crc kubenswrapper[4824]: I1006 09:58:58.929015 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:58 crc kubenswrapper[4824]: I1006 09:58:58.929033 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:58 crc kubenswrapper[4824]: I1006 09:58:58.929063 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:58 crc kubenswrapper[4824]: I1006 09:58:58.929084 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:58Z","lastTransitionTime":"2025-10-06T09:58:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:59 crc kubenswrapper[4824]: I1006 09:58:59.032552 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:59 crc kubenswrapper[4824]: I1006 09:58:59.032620 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:59 crc kubenswrapper[4824]: I1006 09:58:59.032639 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:59 crc kubenswrapper[4824]: I1006 09:58:59.032668 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:59 crc kubenswrapper[4824]: I1006 09:58:59.032691 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:59Z","lastTransitionTime":"2025-10-06T09:58:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:59 crc kubenswrapper[4824]: I1006 09:58:59.135901 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:59 crc kubenswrapper[4824]: I1006 09:58:59.136039 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:59 crc kubenswrapper[4824]: I1006 09:58:59.136060 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:59 crc kubenswrapper[4824]: I1006 09:58:59.136134 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:59 crc kubenswrapper[4824]: I1006 09:58:59.136158 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:59Z","lastTransitionTime":"2025-10-06T09:58:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:59 crc kubenswrapper[4824]: I1006 09:58:59.240017 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:59 crc kubenswrapper[4824]: I1006 09:58:59.240098 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:59 crc kubenswrapper[4824]: I1006 09:58:59.240118 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:59 crc kubenswrapper[4824]: I1006 09:58:59.240150 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:59 crc kubenswrapper[4824]: I1006 09:58:59.240241 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:59Z","lastTransitionTime":"2025-10-06T09:58:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:59 crc kubenswrapper[4824]: I1006 09:58:59.343484 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:59 crc kubenswrapper[4824]: I1006 09:58:59.343556 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:59 crc kubenswrapper[4824]: I1006 09:58:59.343575 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:59 crc kubenswrapper[4824]: I1006 09:58:59.343603 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:59 crc kubenswrapper[4824]: I1006 09:58:59.343624 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:59Z","lastTransitionTime":"2025-10-06T09:58:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:59 crc kubenswrapper[4824]: I1006 09:58:59.446592 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:59 crc kubenswrapper[4824]: I1006 09:58:59.446701 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:59 crc kubenswrapper[4824]: I1006 09:58:59.446725 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:59 crc kubenswrapper[4824]: I1006 09:58:59.446759 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:59 crc kubenswrapper[4824]: I1006 09:58:59.446785 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:59Z","lastTransitionTime":"2025-10-06T09:58:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:59 crc kubenswrapper[4824]: I1006 09:58:59.550907 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:59 crc kubenswrapper[4824]: I1006 09:58:59.551013 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:59 crc kubenswrapper[4824]: I1006 09:58:59.551034 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:59 crc kubenswrapper[4824]: I1006 09:58:59.551065 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:59 crc kubenswrapper[4824]: I1006 09:58:59.551084 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:59Z","lastTransitionTime":"2025-10-06T09:58:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:59 crc kubenswrapper[4824]: I1006 09:58:59.654634 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:59 crc kubenswrapper[4824]: I1006 09:58:59.654702 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:59 crc kubenswrapper[4824]: I1006 09:58:59.654722 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:59 crc kubenswrapper[4824]: I1006 09:58:59.654750 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:59 crc kubenswrapper[4824]: I1006 09:58:59.654773 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:59Z","lastTransitionTime":"2025-10-06T09:58:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:59 crc kubenswrapper[4824]: I1006 09:58:59.757739 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:59 crc kubenswrapper[4824]: I1006 09:58:59.757797 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:59 crc kubenswrapper[4824]: I1006 09:58:59.757808 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:59 crc kubenswrapper[4824]: I1006 09:58:59.757825 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:59 crc kubenswrapper[4824]: I1006 09:58:59.757838 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:59Z","lastTransitionTime":"2025-10-06T09:58:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:59 crc kubenswrapper[4824]: I1006 09:58:59.861038 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:59 crc kubenswrapper[4824]: I1006 09:58:59.861097 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:59 crc kubenswrapper[4824]: I1006 09:58:59.861114 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:59 crc kubenswrapper[4824]: I1006 09:58:59.861135 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:59 crc kubenswrapper[4824]: I1006 09:58:59.861149 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:59Z","lastTransitionTime":"2025-10-06T09:58:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:58:59 crc kubenswrapper[4824]: I1006 09:58:59.964503 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:58:59 crc kubenswrapper[4824]: I1006 09:58:59.964568 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:58:59 crc kubenswrapper[4824]: I1006 09:58:59.964585 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:58:59 crc kubenswrapper[4824]: I1006 09:58:59.964611 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:58:59 crc kubenswrapper[4824]: I1006 09:58:59.964709 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:58:59Z","lastTransitionTime":"2025-10-06T09:58:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:00 crc kubenswrapper[4824]: I1006 09:59:00.068614 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:00 crc kubenswrapper[4824]: I1006 09:59:00.068689 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:00 crc kubenswrapper[4824]: I1006 09:59:00.068708 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:00 crc kubenswrapper[4824]: I1006 09:59:00.068740 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:00 crc kubenswrapper[4824]: I1006 09:59:00.068762 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:00Z","lastTransitionTime":"2025-10-06T09:59:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:00 crc kubenswrapper[4824]: I1006 09:59:00.172569 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:00 crc kubenswrapper[4824]: I1006 09:59:00.172652 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:00 crc kubenswrapper[4824]: I1006 09:59:00.172663 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:00 crc kubenswrapper[4824]: I1006 09:59:00.172683 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:00 crc kubenswrapper[4824]: I1006 09:59:00.172730 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:00Z","lastTransitionTime":"2025-10-06T09:59:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:00 crc kubenswrapper[4824]: I1006 09:59:00.273876 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 09:59:00 crc kubenswrapper[4824]: I1006 09:59:00.274626 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gnw94" Oct 06 09:59:00 crc kubenswrapper[4824]: I1006 09:59:00.274727 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 09:59:00 crc kubenswrapper[4824]: I1006 09:59:00.274747 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 09:59:00 crc kubenswrapper[4824]: E1006 09:59:00.274913 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 06 09:59:00 crc kubenswrapper[4824]: E1006 09:59:00.275341 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gnw94" podUID="7cf1d355-64cb-48a6-acfd-7d258b7afa10" Oct 06 09:59:00 crc kubenswrapper[4824]: E1006 09:59:00.275450 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 06 09:59:00 crc kubenswrapper[4824]: E1006 09:59:00.275623 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 06 09:59:00 crc kubenswrapper[4824]: I1006 09:59:00.275703 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:00 crc kubenswrapper[4824]: I1006 09:59:00.275738 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:00 crc kubenswrapper[4824]: I1006 09:59:00.275758 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:00 crc kubenswrapper[4824]: I1006 09:59:00.275787 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:00 crc kubenswrapper[4824]: I1006 09:59:00.275809 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:00Z","lastTransitionTime":"2025-10-06T09:59:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:00 crc kubenswrapper[4824]: I1006 09:59:00.276060 4824 scope.go:117] "RemoveContainer" containerID="91a256538722530a7679c5bededd29cff892d03964bd6c71f81d33989b0efdd9" Oct 06 09:59:00 crc kubenswrapper[4824]: I1006 09:59:00.380450 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:00 crc kubenswrapper[4824]: I1006 09:59:00.380521 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:00 crc kubenswrapper[4824]: I1006 09:59:00.380541 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:00 crc kubenswrapper[4824]: I1006 09:59:00.380569 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:00 crc kubenswrapper[4824]: I1006 09:59:00.380588 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:00Z","lastTransitionTime":"2025-10-06T09:59:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:00 crc kubenswrapper[4824]: I1006 09:59:00.483871 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:00 crc kubenswrapper[4824]: I1006 09:59:00.483924 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:00 crc kubenswrapper[4824]: I1006 09:59:00.483972 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:00 crc kubenswrapper[4824]: I1006 09:59:00.484048 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:00 crc kubenswrapper[4824]: I1006 09:59:00.484067 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:00Z","lastTransitionTime":"2025-10-06T09:59:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:00 crc kubenswrapper[4824]: I1006 09:59:00.588283 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:00 crc kubenswrapper[4824]: I1006 09:59:00.588336 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:00 crc kubenswrapper[4824]: I1006 09:59:00.588347 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:00 crc kubenswrapper[4824]: I1006 09:59:00.588365 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:00 crc kubenswrapper[4824]: I1006 09:59:00.588376 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:00Z","lastTransitionTime":"2025-10-06T09:59:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:00 crc kubenswrapper[4824]: I1006 09:59:00.691202 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:00 crc kubenswrapper[4824]: I1006 09:59:00.691265 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:00 crc kubenswrapper[4824]: I1006 09:59:00.691288 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:00 crc kubenswrapper[4824]: I1006 09:59:00.691313 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:00 crc kubenswrapper[4824]: I1006 09:59:00.691334 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:00Z","lastTransitionTime":"2025-10-06T09:59:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:00 crc kubenswrapper[4824]: I1006 09:59:00.728966 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-4vwms_2f511ece-25eb-465f-b85c-9e5f3f886c21/ovnkube-controller/1.log" Oct 06 09:59:00 crc kubenswrapper[4824]: I1006 09:59:00.733453 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" event={"ID":"2f511ece-25eb-465f-b85c-9e5f3f886c21","Type":"ContainerStarted","Data":"33caf222d92beb8a42c203fc61d69e804ec3a22d6889f8ead2ce783237d8b464"} Oct 06 09:59:00 crc kubenswrapper[4824]: I1006 09:59:00.734256 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" Oct 06 09:59:00 crc kubenswrapper[4824]: I1006 09:59:00.765735 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b665cda3-da4e-482e-8248-023abec382f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9235c742d8c45bf667f32979c2891fece50f4ee1bd0c7cf6a753ed2d2b9ab30e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b2a920dbc506df2426232990e31fbb60d255c00be54941826a0f094af4a92f0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d516f278f362d7c45e7149de82fa835997ac198f1165c04352910d51fb674d69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c820d8bea16cff0d463ce68d2f2acb72d7a7c73372757ff104e14fcec1025a05\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b8cdf54407cc95783919eba8e92f093bbc09eac5e908c4ba0685629948c71bd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-06T09:58:27Z\\\",\\\"message\\\":\\\"arting RequestHeaderAuthRequestController\\\\nI1006 09:58:27.156944 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1006 09:58:27.156783 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1006 09:58:27.157213 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-432254888/tls.crt::/tmp/serving-cert-432254888/tls.key\\\\\\\"\\\\nI1006 09:58:27.157407 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1006 09:58:27.157443 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1006 09:58:27.157410 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-432254888/tls.crt::/tmp/serving-cert-432254888/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759744705\\\\\\\\\\\\\\\" (2025-10-06 09:58:25 +0000 UTC to 2025-11-05 09:58:26 +0000 UTC (now=2025-10-06 09:58:27.157390076 +0000 UTC))\\\\\\\"\\\\nI1006 09:58:27.157632 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1006 09:58:27.158158 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759744707\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759744707\\\\\\\\\\\\\\\" (2025-10-06 08:58:26 +0000 UTC to 2026-10-06 08:58:26 +0000 UTC (now=2025-10-06 09:58:27.158090663 +0000 UTC))\\\\\\\"\\\\nI1006 09:58:27.158220 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1006 09:58:27.158270 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1006 09:58:27.158310 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nF1006 09:58:27.158617 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://413822c99ba2866521177cce4f1894aef17c160aebb2662da91824a04827d510\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e37e8c7400f7fd14a5c5390a5985e3d8d277e9c5ca60d5e9e9757d07d1778570\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e37e8c7400f7fd14a5c5390a5985e3d8d277e9c5ca60d5e9e9757d07d1778570\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:00Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:00 crc kubenswrapper[4824]: I1006 09:59:00.787951 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d739433-0a38-42fe-b884-b8433011967e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://464e279c5903770c34d80f31ed231b194d5628b39055c4058908039d10690059\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://efa1b62728466ef7aac0e745230348ce920eb9ec9b07fb2cd5a276150dac13fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fedc40f0217626c56d155fc7e78e0592556fef9464db2aaf09c5facc486aad8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78fac300cbe2ac52f5788306b8fc0011f6ba3a22c3546974e70998f2d394987a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:00Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:00 crc kubenswrapper[4824]: I1006 09:59:00.794581 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:00 crc kubenswrapper[4824]: I1006 09:59:00.794640 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:00 crc kubenswrapper[4824]: I1006 09:59:00.794660 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:00 crc kubenswrapper[4824]: I1006 09:59:00.794691 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:00 crc kubenswrapper[4824]: I1006 09:59:00.794710 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:00Z","lastTransitionTime":"2025-10-06T09:59:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:00 crc kubenswrapper[4824]: I1006 09:59:00.810852 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:00Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:00 crc kubenswrapper[4824]: I1006 09:59:00.827792 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-szn8q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b65abf9f-4e88-4571-960d-3ca997d9c344\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://664bf42e1324a9c9529f36f08d61e26999b3386bdd1fac77851fb7e7411e8934\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49s5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-szn8q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:00Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:00 crc kubenswrapper[4824]: I1006 09:59:00.853966 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f511ece-25eb-465f-b85c-9e5f3f886c21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10bfed3e523b7422ce1e981f98166eca78bd12dd4a7cf0634915e1f4ecc3de2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43199f86ad4a1f02bc57a926cd96b2c3c0e2ea46fb36a3f495668706be4c8041\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d6064ad3bcdafebb4c66e7d91f183cc787e9ce13660545c71a722b3ff04ce75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6213164d366243c2139d5971041d6f076cf38fd0cce4c0ca5495111963487ec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa6d59fac987f60a4340ab35d84aa059f2b7085a06362e7207a79702d17eee6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09638fbc9d163ad0b63fa57d24ea5298207ad1f55bab8671be311960be2627a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33caf222d92beb8a42c203fc61d69e804ec3a22d6889f8ead2ce783237d8b464\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91a256538722530a7679c5bededd29cff892d03964bd6c71f81d33989b0efdd9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-06T09:58:43Z\\\",\\\"message\\\":\\\" to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:43Z is after 2025-08-24T17:21:41Z]\\\\nI1006 09:58:43.725317 6283 obj_retry.go:365] Adding new object: *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI1006 09:58:43.725319 6283 services_controller.go:356] Processing sync for service openshift-marketplace/marketplace-operator-metrics for network=default\\\\nI1006 09:58:43.725320 6283 lb_config.go:1031] Cluster endpoints for openshift-machine-config-operator/machine-config-controller for network=default are: map[]\\\\nI1006 09:58:43.725336 6283 ovn.go:134] Ensuring zone local for Pod openshift-kube-controller-manager/kube-controller-manager-crc in node crc\\\\nI1006 09:58:43.725345 6283 obj_retry.go:386] Retry successful for *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc after 0 failed attempt(s)\\\\nI1006 09:58:43.725352 6283 default_network_controller.go:776] Recording success event on pod openshift-ku\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:42Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:59:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0644c7db28c7d26aa8d69cd5b9fbcff4d4c06b96df113137dea22a72d9fecd85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6da4c07be4d507a6f20bd8b42be6e6ea4712a956f57ef164036cfded7142a87e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6da4c07be4d507a6f20bd8b42be6e6ea4712a956f57ef164036cfded7142a87e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4vwms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:00Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:00 crc kubenswrapper[4824]: I1006 09:59:00.868999 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lszvg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0cf70ab4-4cfe-41c5-8db7-035451bafcfa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7b80b5b375b3b32132501ee0d661e06e199654c696ac76ad56e7ba96918cc77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fkbf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://519653c6b241cd89ef3ece003667d420a7a55cdf513f4577a69e14ade5fb4c76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fkbf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lszvg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:00Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:00 crc kubenswrapper[4824]: I1006 09:59:00.879522 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gnw94" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7cf1d355-64cb-48a6-acfd-7d258b7afa10\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-psbq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-psbq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:46Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gnw94\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:00Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:00 crc kubenswrapper[4824]: I1006 09:59:00.897176 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-7c22q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7eaaa68-d971-4a1c-b6c5-551ded8ff27f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f5150c8c6cfc65ecdcd0e7bd52199dd60e2472c96874f103b49919c81b5e5f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53336c7d3c509b4e022708998aedc5dee9fdaf7136ae3ee69b4181394c3a22fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53336c7d3c509b4e022708998aedc5dee9fdaf7136ae3ee69b4181394c3a22fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://652ab0a13a2bb421872edda3f7188dce8e296247052550c937f8447fea206756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://652ab0a13a2bb421872edda3f7188dce8e296247052550c937f8447fea206756\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3514dec0386bb8c016a7f555d7191ea420d5058b0ee98f97d965dd857c5376e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3514dec0386bb8c016a7f555d7191ea420d5058b0ee98f97d965dd857c5376e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b035e0bf34e3429400ac8942124a3e9834231899487d03d15609fe5624747a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b035e0bf34e3429400ac8942124a3e9834231899487d03d15609fe5624747a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6743e027a11864ad3f433d110e60451fb36d49791e03b22d437b70f89d056ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6743e027a11864ad3f433d110e60451fb36d49791e03b22d437b70f89d056ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8ea48b6886fd39a1c5e5b29ca1fffc2bd4552d61dc621eb2fba66fb464cf120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8ea48b6886fd39a1c5e5b29ca1fffc2bd4552d61dc621eb2fba66fb464cf120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-7c22q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:00Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:00 crc kubenswrapper[4824]: I1006 09:59:00.897924 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:00 crc kubenswrapper[4824]: I1006 09:59:00.897957 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:00 crc kubenswrapper[4824]: I1006 09:59:00.897972 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:00 crc kubenswrapper[4824]: I1006 09:59:00.898004 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:00 crc kubenswrapper[4824]: I1006 09:59:00.898015 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:00Z","lastTransitionTime":"2025-10-06T09:59:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:00 crc kubenswrapper[4824]: I1006 09:59:00.913455 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2035278cf272827b39a45aa41ecd5b1e41aafc1d9eb8bccd853660fb6e587159\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff3152fceadf36a55578f587623f37d748f90f2afeda5702a05504b372d399be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:00Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:00 crc kubenswrapper[4824]: I1006 09:59:00.922971 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xcgwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74317f75-9ebc-4327-b88c-598c2241240f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8def09a126f9bb910affad9b7fb32303b960add9fbfd5bb709d59b4aeba7d245\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v6sg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xcgwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:00Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:00 crc kubenswrapper[4824]: I1006 09:59:00.935904 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1140c8c-93f1-4fce-8c06-a69561ff8a6d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5462cb77b7d61aed7d4b4be2fbc6ba1451d267a66afd6dede2f259c15230e7c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gq49w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5512be36013baeca611ee92645a1d339da237b70d63aac43334beafe24513d44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gq49w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-khgzw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:00Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:00 crc kubenswrapper[4824]: I1006 09:59:00.950089 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f9011f6673310dffea3acd7bdc2a56f2af91a87641631490fdc63da610f047a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:00Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:00 crc kubenswrapper[4824]: I1006 09:59:00.963021 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:00Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:00 crc kubenswrapper[4824]: I1006 09:59:00.982469 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:00Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:01 crc kubenswrapper[4824]: I1006 09:59:01.000403 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:01 crc kubenswrapper[4824]: I1006 09:59:01.000490 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:01 crc kubenswrapper[4824]: I1006 09:59:01.000504 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:01 crc kubenswrapper[4824]: I1006 09:59:01.000527 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:01 crc kubenswrapper[4824]: I1006 09:59:01.000543 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:01Z","lastTransitionTime":"2025-10-06T09:59:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:01 crc kubenswrapper[4824]: I1006 09:59:01.003257 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab36ed34d6ca02b59043c46e151019620ce31733209d072c21394a79fe36be4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:01Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:01 crc kubenswrapper[4824]: I1006 09:59:01.020109 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5rx6l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0c320fc-94d9-4d82-81ee-ccccfd125efc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f8be7bad4e9644494dfb42380c7f5481e1039a1d27ad379c623515948d87392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njtbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5rx6l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:01Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:01 crc kubenswrapper[4824]: I1006 09:59:01.104473 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:01 crc kubenswrapper[4824]: I1006 09:59:01.104547 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:01 crc kubenswrapper[4824]: I1006 09:59:01.104565 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:01 crc kubenswrapper[4824]: I1006 09:59:01.104593 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:01 crc kubenswrapper[4824]: I1006 09:59:01.104613 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:01Z","lastTransitionTime":"2025-10-06T09:59:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:01 crc kubenswrapper[4824]: I1006 09:59:01.207744 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:01 crc kubenswrapper[4824]: I1006 09:59:01.207812 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:01 crc kubenswrapper[4824]: I1006 09:59:01.207830 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:01 crc kubenswrapper[4824]: I1006 09:59:01.207857 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:01 crc kubenswrapper[4824]: I1006 09:59:01.207879 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:01Z","lastTransitionTime":"2025-10-06T09:59:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:01 crc kubenswrapper[4824]: I1006 09:59:01.296500 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:01Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:01 crc kubenswrapper[4824]: I1006 09:59:01.311169 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:01 crc kubenswrapper[4824]: I1006 09:59:01.311231 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:01 crc kubenswrapper[4824]: I1006 09:59:01.311251 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:01 crc kubenswrapper[4824]: I1006 09:59:01.311280 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:01 crc kubenswrapper[4824]: I1006 09:59:01.311301 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:01Z","lastTransitionTime":"2025-10-06T09:59:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:01 crc kubenswrapper[4824]: I1006 09:59:01.318528 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-szn8q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b65abf9f-4e88-4571-960d-3ca997d9c344\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://664bf42e1324a9c9529f36f08d61e26999b3386bdd1fac77851fb7e7411e8934\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49s5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-szn8q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:01Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:01 crc kubenswrapper[4824]: I1006 09:59:01.352177 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f511ece-25eb-465f-b85c-9e5f3f886c21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10bfed3e523b7422ce1e981f98166eca78bd12dd4a7cf0634915e1f4ecc3de2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43199f86ad4a1f02bc57a926cd96b2c3c0e2ea46fb36a3f495668706be4c8041\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d6064ad3bcdafebb4c66e7d91f183cc787e9ce13660545c71a722b3ff04ce75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6213164d366243c2139d5971041d6f076cf38fd0cce4c0ca5495111963487ec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa6d59fac987f60a4340ab35d84aa059f2b7085a06362e7207a79702d17eee6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09638fbc9d163ad0b63fa57d24ea5298207ad1f55bab8671be311960be2627a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33caf222d92beb8a42c203fc61d69e804ec3a22d6889f8ead2ce783237d8b464\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91a256538722530a7679c5bededd29cff892d03964bd6c71f81d33989b0efdd9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-06T09:58:43Z\\\",\\\"message\\\":\\\" to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:43Z is after 2025-08-24T17:21:41Z]\\\\nI1006 09:58:43.725317 6283 obj_retry.go:365] Adding new object: *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI1006 09:58:43.725319 6283 services_controller.go:356] Processing sync for service openshift-marketplace/marketplace-operator-metrics for network=default\\\\nI1006 09:58:43.725320 6283 lb_config.go:1031] Cluster endpoints for openshift-machine-config-operator/machine-config-controller for network=default are: map[]\\\\nI1006 09:58:43.725336 6283 ovn.go:134] Ensuring zone local for Pod openshift-kube-controller-manager/kube-controller-manager-crc in node crc\\\\nI1006 09:58:43.725345 6283 obj_retry.go:386] Retry successful for *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc after 0 failed attempt(s)\\\\nI1006 09:58:43.725352 6283 default_network_controller.go:776] Recording success event on pod openshift-ku\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:42Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:59:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0644c7db28c7d26aa8d69cd5b9fbcff4d4c06b96df113137dea22a72d9fecd85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6da4c07be4d507a6f20bd8b42be6e6ea4712a956f57ef164036cfded7142a87e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6da4c07be4d507a6f20bd8b42be6e6ea4712a956f57ef164036cfded7142a87e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4vwms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:01Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:01 crc kubenswrapper[4824]: I1006 09:59:01.374666 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lszvg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0cf70ab4-4cfe-41c5-8db7-035451bafcfa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7b80b5b375b3b32132501ee0d661e06e199654c696ac76ad56e7ba96918cc77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fkbf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://519653c6b241cd89ef3ece003667d420a7a55cdf513f4577a69e14ade5fb4c76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fkbf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lszvg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:01Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:01 crc kubenswrapper[4824]: I1006 09:59:01.391524 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gnw94" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7cf1d355-64cb-48a6-acfd-7d258b7afa10\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-psbq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-psbq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:46Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gnw94\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:01Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:01 crc kubenswrapper[4824]: I1006 09:59:01.414776 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b665cda3-da4e-482e-8248-023abec382f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9235c742d8c45bf667f32979c2891fece50f4ee1bd0c7cf6a753ed2d2b9ab30e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b2a920dbc506df2426232990e31fbb60d255c00be54941826a0f094af4a92f0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d516f278f362d7c45e7149de82fa835997ac198f1165c04352910d51fb674d69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c820d8bea16cff0d463ce68d2f2acb72d7a7c73372757ff104e14fcec1025a05\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b8cdf54407cc95783919eba8e92f093bbc09eac5e908c4ba0685629948c71bd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-06T09:58:27Z\\\",\\\"message\\\":\\\"arting RequestHeaderAuthRequestController\\\\nI1006 09:58:27.156944 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1006 09:58:27.156783 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1006 09:58:27.157213 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-432254888/tls.crt::/tmp/serving-cert-432254888/tls.key\\\\\\\"\\\\nI1006 09:58:27.157407 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1006 09:58:27.157443 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1006 09:58:27.157410 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-432254888/tls.crt::/tmp/serving-cert-432254888/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759744705\\\\\\\\\\\\\\\" (2025-10-06 09:58:25 +0000 UTC to 2025-11-05 09:58:26 +0000 UTC (now=2025-10-06 09:58:27.157390076 +0000 UTC))\\\\\\\"\\\\nI1006 09:58:27.157632 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1006 09:58:27.158158 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759744707\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759744707\\\\\\\\\\\\\\\" (2025-10-06 08:58:26 +0000 UTC to 2026-10-06 08:58:26 +0000 UTC (now=2025-10-06 09:58:27.158090663 +0000 UTC))\\\\\\\"\\\\nI1006 09:58:27.158220 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1006 09:58:27.158270 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1006 09:58:27.158310 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nF1006 09:58:27.158617 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://413822c99ba2866521177cce4f1894aef17c160aebb2662da91824a04827d510\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e37e8c7400f7fd14a5c5390a5985e3d8d277e9c5ca60d5e9e9757d07d1778570\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e37e8c7400f7fd14a5c5390a5985e3d8d277e9c5ca60d5e9e9757d07d1778570\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:01Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:01 crc kubenswrapper[4824]: I1006 09:59:01.415647 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:01 crc kubenswrapper[4824]: I1006 09:59:01.415708 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:01 crc kubenswrapper[4824]: I1006 09:59:01.415719 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:01 crc kubenswrapper[4824]: I1006 09:59:01.415739 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:01 crc kubenswrapper[4824]: I1006 09:59:01.415752 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:01Z","lastTransitionTime":"2025-10-06T09:59:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:01 crc kubenswrapper[4824]: I1006 09:59:01.434954 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d739433-0a38-42fe-b884-b8433011967e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://464e279c5903770c34d80f31ed231b194d5628b39055c4058908039d10690059\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://efa1b62728466ef7aac0e745230348ce920eb9ec9b07fb2cd5a276150dac13fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fedc40f0217626c56d155fc7e78e0592556fef9464db2aaf09c5facc486aad8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78fac300cbe2ac52f5788306b8fc0011f6ba3a22c3546974e70998f2d394987a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:01Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:01 crc kubenswrapper[4824]: I1006 09:59:01.460115 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-7c22q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7eaaa68-d971-4a1c-b6c5-551ded8ff27f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f5150c8c6cfc65ecdcd0e7bd52199dd60e2472c96874f103b49919c81b5e5f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53336c7d3c509b4e022708998aedc5dee9fdaf7136ae3ee69b4181394c3a22fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53336c7d3c509b4e022708998aedc5dee9fdaf7136ae3ee69b4181394c3a22fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://652ab0a13a2bb421872edda3f7188dce8e296247052550c937f8447fea206756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://652ab0a13a2bb421872edda3f7188dce8e296247052550c937f8447fea206756\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3514dec0386bb8c016a7f555d7191ea420d5058b0ee98f97d965dd857c5376e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3514dec0386bb8c016a7f555d7191ea420d5058b0ee98f97d965dd857c5376e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b035e0bf34e3429400ac8942124a3e9834231899487d03d15609fe5624747a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b035e0bf34e3429400ac8942124a3e9834231899487d03d15609fe5624747a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6743e027a11864ad3f433d110e60451fb36d49791e03b22d437b70f89d056ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6743e027a11864ad3f433d110e60451fb36d49791e03b22d437b70f89d056ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8ea48b6886fd39a1c5e5b29ca1fffc2bd4552d61dc621eb2fba66fb464cf120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8ea48b6886fd39a1c5e5b29ca1fffc2bd4552d61dc621eb2fba66fb464cf120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-7c22q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:01Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:01 crc kubenswrapper[4824]: I1006 09:59:01.477723 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1140c8c-93f1-4fce-8c06-a69561ff8a6d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5462cb77b7d61aed7d4b4be2fbc6ba1451d267a66afd6dede2f259c15230e7c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gq49w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5512be36013baeca611ee92645a1d339da237b70d63aac43334beafe24513d44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gq49w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-khgzw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:01Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:01 crc kubenswrapper[4824]: I1006 09:59:01.498734 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2035278cf272827b39a45aa41ecd5b1e41aafc1d9eb8bccd853660fb6e587159\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff3152fceadf36a55578f587623f37d748f90f2afeda5702a05504b372d399be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:01Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:01 crc kubenswrapper[4824]: I1006 09:59:01.515332 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xcgwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74317f75-9ebc-4327-b88c-598c2241240f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8def09a126f9bb910affad9b7fb32303b960add9fbfd5bb709d59b4aeba7d245\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v6sg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xcgwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:01Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:01 crc kubenswrapper[4824]: I1006 09:59:01.519139 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:01 crc kubenswrapper[4824]: I1006 09:59:01.519250 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:01 crc kubenswrapper[4824]: I1006 09:59:01.519272 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:01 crc kubenswrapper[4824]: I1006 09:59:01.519301 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:01 crc kubenswrapper[4824]: I1006 09:59:01.519322 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:01Z","lastTransitionTime":"2025-10-06T09:59:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:01 crc kubenswrapper[4824]: I1006 09:59:01.541872 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f9011f6673310dffea3acd7bdc2a56f2af91a87641631490fdc63da610f047a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:01Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:01 crc kubenswrapper[4824]: I1006 09:59:01.558106 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:01Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:01 crc kubenswrapper[4824]: I1006 09:59:01.576315 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:01Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:01 crc kubenswrapper[4824]: I1006 09:59:01.590499 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab36ed34d6ca02b59043c46e151019620ce31733209d072c21394a79fe36be4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:01Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:01 crc kubenswrapper[4824]: I1006 09:59:01.608271 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5rx6l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0c320fc-94d9-4d82-81ee-ccccfd125efc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f8be7bad4e9644494dfb42380c7f5481e1039a1d27ad379c623515948d87392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njtbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5rx6l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:01Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:01 crc kubenswrapper[4824]: I1006 09:59:01.621919 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:01 crc kubenswrapper[4824]: I1006 09:59:01.622018 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:01 crc kubenswrapper[4824]: I1006 09:59:01.622041 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:01 crc kubenswrapper[4824]: I1006 09:59:01.622073 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:01 crc kubenswrapper[4824]: I1006 09:59:01.622092 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:01Z","lastTransitionTime":"2025-10-06T09:59:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:01 crc kubenswrapper[4824]: I1006 09:59:01.725848 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:01 crc kubenswrapper[4824]: I1006 09:59:01.725952 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:01 crc kubenswrapper[4824]: I1006 09:59:01.725972 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:01 crc kubenswrapper[4824]: I1006 09:59:01.726043 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:01 crc kubenswrapper[4824]: I1006 09:59:01.726069 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:01Z","lastTransitionTime":"2025-10-06T09:59:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:01 crc kubenswrapper[4824]: I1006 09:59:01.741915 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-4vwms_2f511ece-25eb-465f-b85c-9e5f3f886c21/ovnkube-controller/2.log" Oct 06 09:59:01 crc kubenswrapper[4824]: I1006 09:59:01.743288 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-4vwms_2f511ece-25eb-465f-b85c-9e5f3f886c21/ovnkube-controller/1.log" Oct 06 09:59:01 crc kubenswrapper[4824]: I1006 09:59:01.748129 4824 generic.go:334] "Generic (PLEG): container finished" podID="2f511ece-25eb-465f-b85c-9e5f3f886c21" containerID="33caf222d92beb8a42c203fc61d69e804ec3a22d6889f8ead2ce783237d8b464" exitCode=1 Oct 06 09:59:01 crc kubenswrapper[4824]: I1006 09:59:01.748200 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" event={"ID":"2f511ece-25eb-465f-b85c-9e5f3f886c21","Type":"ContainerDied","Data":"33caf222d92beb8a42c203fc61d69e804ec3a22d6889f8ead2ce783237d8b464"} Oct 06 09:59:01 crc kubenswrapper[4824]: I1006 09:59:01.748362 4824 scope.go:117] "RemoveContainer" containerID="91a256538722530a7679c5bededd29cff892d03964bd6c71f81d33989b0efdd9" Oct 06 09:59:01 crc kubenswrapper[4824]: I1006 09:59:01.749494 4824 scope.go:117] "RemoveContainer" containerID="33caf222d92beb8a42c203fc61d69e804ec3a22d6889f8ead2ce783237d8b464" Oct 06 09:59:01 crc kubenswrapper[4824]: E1006 09:59:01.749855 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-4vwms_openshift-ovn-kubernetes(2f511ece-25eb-465f-b85c-9e5f3f886c21)\"" pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" podUID="2f511ece-25eb-465f-b85c-9e5f3f886c21" Oct 06 09:59:01 crc kubenswrapper[4824]: I1006 09:59:01.773407 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-7c22q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7eaaa68-d971-4a1c-b6c5-551ded8ff27f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f5150c8c6cfc65ecdcd0e7bd52199dd60e2472c96874f103b49919c81b5e5f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53336c7d3c509b4e022708998aedc5dee9fdaf7136ae3ee69b4181394c3a22fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53336c7d3c509b4e022708998aedc5dee9fdaf7136ae3ee69b4181394c3a22fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://652ab0a13a2bb421872edda3f7188dce8e296247052550c937f8447fea206756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://652ab0a13a2bb421872edda3f7188dce8e296247052550c937f8447fea206756\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3514dec0386bb8c016a7f555d7191ea420d5058b0ee98f97d965dd857c5376e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3514dec0386bb8c016a7f555d7191ea420d5058b0ee98f97d965dd857c5376e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b035e0bf34e3429400ac8942124a3e9834231899487d03d15609fe5624747a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b035e0bf34e3429400ac8942124a3e9834231899487d03d15609fe5624747a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6743e027a11864ad3f433d110e60451fb36d49791e03b22d437b70f89d056ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6743e027a11864ad3f433d110e60451fb36d49791e03b22d437b70f89d056ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8ea48b6886fd39a1c5e5b29ca1fffc2bd4552d61dc621eb2fba66fb464cf120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8ea48b6886fd39a1c5e5b29ca1fffc2bd4552d61dc621eb2fba66fb464cf120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-7c22q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:01Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:01 crc kubenswrapper[4824]: I1006 09:59:01.803935 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2035278cf272827b39a45aa41ecd5b1e41aafc1d9eb8bccd853660fb6e587159\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff3152fceadf36a55578f587623f37d748f90f2afeda5702a05504b372d399be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:01Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:01 crc kubenswrapper[4824]: I1006 09:59:01.822328 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xcgwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74317f75-9ebc-4327-b88c-598c2241240f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8def09a126f9bb910affad9b7fb32303b960add9fbfd5bb709d59b4aeba7d245\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v6sg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xcgwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:01Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:01 crc kubenswrapper[4824]: I1006 09:59:01.829122 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:01 crc kubenswrapper[4824]: I1006 09:59:01.829170 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:01 crc kubenswrapper[4824]: I1006 09:59:01.829183 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:01 crc kubenswrapper[4824]: I1006 09:59:01.829226 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:01 crc kubenswrapper[4824]: I1006 09:59:01.829242 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:01Z","lastTransitionTime":"2025-10-06T09:59:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:01 crc kubenswrapper[4824]: I1006 09:59:01.839411 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1140c8c-93f1-4fce-8c06-a69561ff8a6d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5462cb77b7d61aed7d4b4be2fbc6ba1451d267a66afd6dede2f259c15230e7c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gq49w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5512be36013baeca611ee92645a1d339da237b70d63aac43334beafe24513d44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gq49w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-khgzw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:01Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:01 crc kubenswrapper[4824]: I1006 09:59:01.860276 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:01Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:01 crc kubenswrapper[4824]: I1006 09:59:01.880567 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab36ed34d6ca02b59043c46e151019620ce31733209d072c21394a79fe36be4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:01Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:01 crc kubenswrapper[4824]: I1006 09:59:01.895970 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5rx6l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0c320fc-94d9-4d82-81ee-ccccfd125efc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f8be7bad4e9644494dfb42380c7f5481e1039a1d27ad379c623515948d87392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njtbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5rx6l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:01Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:01 crc kubenswrapper[4824]: I1006 09:59:01.916224 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f9011f6673310dffea3acd7bdc2a56f2af91a87641631490fdc63da610f047a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:01Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:01 crc kubenswrapper[4824]: I1006 09:59:01.932476 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:01 crc kubenswrapper[4824]: I1006 09:59:01.932570 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:01 crc kubenswrapper[4824]: I1006 09:59:01.932589 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:01 crc kubenswrapper[4824]: I1006 09:59:01.932619 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:01 crc kubenswrapper[4824]: I1006 09:59:01.932641 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:01Z","lastTransitionTime":"2025-10-06T09:59:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:01 crc kubenswrapper[4824]: I1006 09:59:01.935691 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:01Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:01 crc kubenswrapper[4824]: I1006 09:59:01.960619 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f511ece-25eb-465f-b85c-9e5f3f886c21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10bfed3e523b7422ce1e981f98166eca78bd12dd4a7cf0634915e1f4ecc3de2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43199f86ad4a1f02bc57a926cd96b2c3c0e2ea46fb36a3f495668706be4c8041\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d6064ad3bcdafebb4c66e7d91f183cc787e9ce13660545c71a722b3ff04ce75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6213164d366243c2139d5971041d6f076cf38fd0cce4c0ca5495111963487ec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa6d59fac987f60a4340ab35d84aa059f2b7085a06362e7207a79702d17eee6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09638fbc9d163ad0b63fa57d24ea5298207ad1f55bab8671be311960be2627a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33caf222d92beb8a42c203fc61d69e804ec3a22d6889f8ead2ce783237d8b464\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91a256538722530a7679c5bededd29cff892d03964bd6c71f81d33989b0efdd9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-06T09:58:43Z\\\",\\\"message\\\":\\\" to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:58:43Z is after 2025-08-24T17:21:41Z]\\\\nI1006 09:58:43.725317 6283 obj_retry.go:365] Adding new object: *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI1006 09:58:43.725319 6283 services_controller.go:356] Processing sync for service openshift-marketplace/marketplace-operator-metrics for network=default\\\\nI1006 09:58:43.725320 6283 lb_config.go:1031] Cluster endpoints for openshift-machine-config-operator/machine-config-controller for network=default are: map[]\\\\nI1006 09:58:43.725336 6283 ovn.go:134] Ensuring zone local for Pod openshift-kube-controller-manager/kube-controller-manager-crc in node crc\\\\nI1006 09:58:43.725345 6283 obj_retry.go:386] Retry successful for *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc after 0 failed attempt(s)\\\\nI1006 09:58:43.725352 6283 default_network_controller.go:776] Recording success event on pod openshift-ku\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:42Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33caf222d92beb8a42c203fc61d69e804ec3a22d6889f8ead2ce783237d8b464\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-06T09:59:01Z\\\",\\\"message\\\":\\\"nt-go/informers/factory.go:160\\\\nI1006 09:59:01.318023 6499 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1006 09:59:01.317995 6499 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1006 09:59:01.318223 6499 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1006 09:59:01.318391 6499 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1006 09:59:01.319128 6499 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1006 09:59:01.319521 6499 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1006 09:59:01.319562 6499 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1006 09:59:01.319610 6499 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1006 09:59:01.319617 6499 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1006 09:59:01.319629 6499 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1006 09:59:01.319666 6499 factory.go:656] Stopping watch factory\\\\nI1006 09:59:01.319698 6499 ovnkube.go:599] Stopped ovnkube\\\\nI1006 09:59:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T09:59:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0644c7db28c7d26aa8d69cd5b9fbcff4d4c06b96df113137dea22a72d9fecd85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6da4c07be4d507a6f20bd8b42be6e6ea4712a956f57ef164036cfded7142a87e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6da4c07be4d507a6f20bd8b42be6e6ea4712a956f57ef164036cfded7142a87e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4vwms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:01Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:01 crc kubenswrapper[4824]: I1006 09:59:01.979340 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lszvg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0cf70ab4-4cfe-41c5-8db7-035451bafcfa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7b80b5b375b3b32132501ee0d661e06e199654c696ac76ad56e7ba96918cc77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fkbf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://519653c6b241cd89ef3ece003667d420a7a55cdf513f4577a69e14ade5fb4c76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fkbf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lszvg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:01Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:01 crc kubenswrapper[4824]: I1006 09:59:01.997964 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gnw94" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7cf1d355-64cb-48a6-acfd-7d258b7afa10\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-psbq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-psbq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:46Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gnw94\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:01Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:02 crc kubenswrapper[4824]: I1006 09:59:02.008793 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7cf1d355-64cb-48a6-acfd-7d258b7afa10-metrics-certs\") pod \"network-metrics-daemon-gnw94\" (UID: \"7cf1d355-64cb-48a6-acfd-7d258b7afa10\") " pod="openshift-multus/network-metrics-daemon-gnw94" Oct 06 09:59:02 crc kubenswrapper[4824]: E1006 09:59:02.009078 4824 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 06 09:59:02 crc kubenswrapper[4824]: E1006 09:59:02.009205 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7cf1d355-64cb-48a6-acfd-7d258b7afa10-metrics-certs podName:7cf1d355-64cb-48a6-acfd-7d258b7afa10 nodeName:}" failed. No retries permitted until 2025-10-06 09:59:18.009172992 +0000 UTC m=+67.373595893 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/7cf1d355-64cb-48a6-acfd-7d258b7afa10-metrics-certs") pod "network-metrics-daemon-gnw94" (UID: "7cf1d355-64cb-48a6-acfd-7d258b7afa10") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 06 09:59:02 crc kubenswrapper[4824]: I1006 09:59:02.023183 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b665cda3-da4e-482e-8248-023abec382f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9235c742d8c45bf667f32979c2891fece50f4ee1bd0c7cf6a753ed2d2b9ab30e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b2a920dbc506df2426232990e31fbb60d255c00be54941826a0f094af4a92f0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d516f278f362d7c45e7149de82fa835997ac198f1165c04352910d51fb674d69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c820d8bea16cff0d463ce68d2f2acb72d7a7c73372757ff104e14fcec1025a05\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b8cdf54407cc95783919eba8e92f093bbc09eac5e908c4ba0685629948c71bd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-06T09:58:27Z\\\",\\\"message\\\":\\\"arting RequestHeaderAuthRequestController\\\\nI1006 09:58:27.156944 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1006 09:58:27.156783 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1006 09:58:27.157213 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-432254888/tls.crt::/tmp/serving-cert-432254888/tls.key\\\\\\\"\\\\nI1006 09:58:27.157407 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1006 09:58:27.157443 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1006 09:58:27.157410 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-432254888/tls.crt::/tmp/serving-cert-432254888/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759744705\\\\\\\\\\\\\\\" (2025-10-06 09:58:25 +0000 UTC to 2025-11-05 09:58:26 +0000 UTC (now=2025-10-06 09:58:27.157390076 +0000 UTC))\\\\\\\"\\\\nI1006 09:58:27.157632 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1006 09:58:27.158158 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759744707\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759744707\\\\\\\\\\\\\\\" (2025-10-06 08:58:26 +0000 UTC to 2026-10-06 08:58:26 +0000 UTC (now=2025-10-06 09:58:27.158090663 +0000 UTC))\\\\\\\"\\\\nI1006 09:58:27.158220 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1006 09:58:27.158270 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1006 09:58:27.158310 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nF1006 09:58:27.158617 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://413822c99ba2866521177cce4f1894aef17c160aebb2662da91824a04827d510\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e37e8c7400f7fd14a5c5390a5985e3d8d277e9c5ca60d5e9e9757d07d1778570\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e37e8c7400f7fd14a5c5390a5985e3d8d277e9c5ca60d5e9e9757d07d1778570\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:02Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:02 crc kubenswrapper[4824]: I1006 09:59:02.036160 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:02 crc kubenswrapper[4824]: I1006 09:59:02.036219 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:02 crc kubenswrapper[4824]: I1006 09:59:02.036238 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:02 crc kubenswrapper[4824]: I1006 09:59:02.036269 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:02 crc kubenswrapper[4824]: I1006 09:59:02.036290 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:02Z","lastTransitionTime":"2025-10-06T09:59:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:02 crc kubenswrapper[4824]: I1006 09:59:02.044603 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d739433-0a38-42fe-b884-b8433011967e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://464e279c5903770c34d80f31ed231b194d5628b39055c4058908039d10690059\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://efa1b62728466ef7aac0e745230348ce920eb9ec9b07fb2cd5a276150dac13fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fedc40f0217626c56d155fc7e78e0592556fef9464db2aaf09c5facc486aad8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78fac300cbe2ac52f5788306b8fc0011f6ba3a22c3546974e70998f2d394987a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:02Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:02 crc kubenswrapper[4824]: I1006 09:59:02.067242 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:02Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:02 crc kubenswrapper[4824]: I1006 09:59:02.089445 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-szn8q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b65abf9f-4e88-4571-960d-3ca997d9c344\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://664bf42e1324a9c9529f36f08d61e26999b3386bdd1fac77851fb7e7411e8934\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49s5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-szn8q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:02Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:02 crc kubenswrapper[4824]: I1006 09:59:02.139216 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:02 crc kubenswrapper[4824]: I1006 09:59:02.139308 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:02 crc kubenswrapper[4824]: I1006 09:59:02.139328 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:02 crc kubenswrapper[4824]: I1006 09:59:02.139356 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:02 crc kubenswrapper[4824]: I1006 09:59:02.139376 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:02Z","lastTransitionTime":"2025-10-06T09:59:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:02 crc kubenswrapper[4824]: I1006 09:59:02.211529 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 09:59:02 crc kubenswrapper[4824]: I1006 09:59:02.211732 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 09:59:02 crc kubenswrapper[4824]: E1006 09:59:02.211781 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-06 09:59:34.211739563 +0000 UTC m=+83.576162454 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 09:59:02 crc kubenswrapper[4824]: E1006 09:59:02.211867 4824 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 06 09:59:02 crc kubenswrapper[4824]: I1006 09:59:02.211939 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 09:59:02 crc kubenswrapper[4824]: E1006 09:59:02.211964 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-06 09:59:34.211941339 +0000 UTC m=+83.576364230 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 06 09:59:02 crc kubenswrapper[4824]: E1006 09:59:02.212169 4824 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 06 09:59:02 crc kubenswrapper[4824]: E1006 09:59:02.212243 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-06 09:59:34.212229166 +0000 UTC m=+83.576652057 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 06 09:59:02 crc kubenswrapper[4824]: I1006 09:59:02.242622 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:02 crc kubenswrapper[4824]: I1006 09:59:02.242714 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:02 crc kubenswrapper[4824]: I1006 09:59:02.242737 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:02 crc kubenswrapper[4824]: I1006 09:59:02.242768 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:02 crc kubenswrapper[4824]: I1006 09:59:02.242789 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:02Z","lastTransitionTime":"2025-10-06T09:59:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:02 crc kubenswrapper[4824]: I1006 09:59:02.273279 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gnw94" Oct 06 09:59:02 crc kubenswrapper[4824]: I1006 09:59:02.273304 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 09:59:02 crc kubenswrapper[4824]: I1006 09:59:02.273369 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 09:59:02 crc kubenswrapper[4824]: E1006 09:59:02.273553 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gnw94" podUID="7cf1d355-64cb-48a6-acfd-7d258b7afa10" Oct 06 09:59:02 crc kubenswrapper[4824]: I1006 09:59:02.273584 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 09:59:02 crc kubenswrapper[4824]: E1006 09:59:02.273747 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 06 09:59:02 crc kubenswrapper[4824]: E1006 09:59:02.273947 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 06 09:59:02 crc kubenswrapper[4824]: E1006 09:59:02.274289 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 06 09:59:02 crc kubenswrapper[4824]: I1006 09:59:02.313102 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 09:59:02 crc kubenswrapper[4824]: I1006 09:59:02.313191 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 09:59:02 crc kubenswrapper[4824]: E1006 09:59:02.313318 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 06 09:59:02 crc kubenswrapper[4824]: E1006 09:59:02.313356 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 06 09:59:02 crc kubenswrapper[4824]: E1006 09:59:02.313379 4824 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 06 09:59:02 crc kubenswrapper[4824]: E1006 09:59:02.313468 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-06 09:59:34.313440314 +0000 UTC m=+83.677863215 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 06 09:59:02 crc kubenswrapper[4824]: E1006 09:59:02.313463 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 06 09:59:02 crc kubenswrapper[4824]: E1006 09:59:02.313517 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 06 09:59:02 crc kubenswrapper[4824]: E1006 09:59:02.313541 4824 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 06 09:59:02 crc kubenswrapper[4824]: E1006 09:59:02.313617 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-06 09:59:34.313592428 +0000 UTC m=+83.678015329 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 06 09:59:02 crc kubenswrapper[4824]: I1006 09:59:02.346192 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:02 crc kubenswrapper[4824]: I1006 09:59:02.346251 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:02 crc kubenswrapper[4824]: I1006 09:59:02.346268 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:02 crc kubenswrapper[4824]: I1006 09:59:02.346294 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:02 crc kubenswrapper[4824]: I1006 09:59:02.346315 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:02Z","lastTransitionTime":"2025-10-06T09:59:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:02 crc kubenswrapper[4824]: I1006 09:59:02.451911 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:02 crc kubenswrapper[4824]: I1006 09:59:02.452016 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:02 crc kubenswrapper[4824]: I1006 09:59:02.452035 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:02 crc kubenswrapper[4824]: I1006 09:59:02.452062 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:02 crc kubenswrapper[4824]: I1006 09:59:02.452082 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:02Z","lastTransitionTime":"2025-10-06T09:59:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:02 crc kubenswrapper[4824]: I1006 09:59:02.555579 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:02 crc kubenswrapper[4824]: I1006 09:59:02.555654 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:02 crc kubenswrapper[4824]: I1006 09:59:02.555668 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:02 crc kubenswrapper[4824]: I1006 09:59:02.555693 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:02 crc kubenswrapper[4824]: I1006 09:59:02.555709 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:02Z","lastTransitionTime":"2025-10-06T09:59:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:02 crc kubenswrapper[4824]: I1006 09:59:02.658831 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:02 crc kubenswrapper[4824]: I1006 09:59:02.658883 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:02 crc kubenswrapper[4824]: I1006 09:59:02.658897 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:02 crc kubenswrapper[4824]: I1006 09:59:02.658919 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:02 crc kubenswrapper[4824]: I1006 09:59:02.658933 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:02Z","lastTransitionTime":"2025-10-06T09:59:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:02 crc kubenswrapper[4824]: I1006 09:59:02.754357 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-4vwms_2f511ece-25eb-465f-b85c-9e5f3f886c21/ovnkube-controller/2.log" Oct 06 09:59:02 crc kubenswrapper[4824]: I1006 09:59:02.759113 4824 scope.go:117] "RemoveContainer" containerID="33caf222d92beb8a42c203fc61d69e804ec3a22d6889f8ead2ce783237d8b464" Oct 06 09:59:02 crc kubenswrapper[4824]: E1006 09:59:02.759470 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-4vwms_openshift-ovn-kubernetes(2f511ece-25eb-465f-b85c-9e5f3f886c21)\"" pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" podUID="2f511ece-25eb-465f-b85c-9e5f3f886c21" Oct 06 09:59:02 crc kubenswrapper[4824]: I1006 09:59:02.761427 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:02 crc kubenswrapper[4824]: I1006 09:59:02.761473 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:02 crc kubenswrapper[4824]: I1006 09:59:02.761489 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:02 crc kubenswrapper[4824]: I1006 09:59:02.761508 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:02 crc kubenswrapper[4824]: I1006 09:59:02.761522 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:02Z","lastTransitionTime":"2025-10-06T09:59:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:02 crc kubenswrapper[4824]: I1006 09:59:02.777960 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gnw94" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7cf1d355-64cb-48a6-acfd-7d258b7afa10\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-psbq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-psbq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:46Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gnw94\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:02Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:02 crc kubenswrapper[4824]: I1006 09:59:02.803114 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b665cda3-da4e-482e-8248-023abec382f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9235c742d8c45bf667f32979c2891fece50f4ee1bd0c7cf6a753ed2d2b9ab30e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b2a920dbc506df2426232990e31fbb60d255c00be54941826a0f094af4a92f0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d516f278f362d7c45e7149de82fa835997ac198f1165c04352910d51fb674d69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c820d8bea16cff0d463ce68d2f2acb72d7a7c73372757ff104e14fcec1025a05\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b8cdf54407cc95783919eba8e92f093bbc09eac5e908c4ba0685629948c71bd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-06T09:58:27Z\\\",\\\"message\\\":\\\"arting RequestHeaderAuthRequestController\\\\nI1006 09:58:27.156944 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1006 09:58:27.156783 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1006 09:58:27.157213 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-432254888/tls.crt::/tmp/serving-cert-432254888/tls.key\\\\\\\"\\\\nI1006 09:58:27.157407 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1006 09:58:27.157443 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1006 09:58:27.157410 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-432254888/tls.crt::/tmp/serving-cert-432254888/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759744705\\\\\\\\\\\\\\\" (2025-10-06 09:58:25 +0000 UTC to 2025-11-05 09:58:26 +0000 UTC (now=2025-10-06 09:58:27.157390076 +0000 UTC))\\\\\\\"\\\\nI1006 09:58:27.157632 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1006 09:58:27.158158 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759744707\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759744707\\\\\\\\\\\\\\\" (2025-10-06 08:58:26 +0000 UTC to 2026-10-06 08:58:26 +0000 UTC (now=2025-10-06 09:58:27.158090663 +0000 UTC))\\\\\\\"\\\\nI1006 09:58:27.158220 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1006 09:58:27.158270 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1006 09:58:27.158310 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nF1006 09:58:27.158617 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://413822c99ba2866521177cce4f1894aef17c160aebb2662da91824a04827d510\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e37e8c7400f7fd14a5c5390a5985e3d8d277e9c5ca60d5e9e9757d07d1778570\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e37e8c7400f7fd14a5c5390a5985e3d8d277e9c5ca60d5e9e9757d07d1778570\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:02Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:02 crc kubenswrapper[4824]: I1006 09:59:02.825514 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d739433-0a38-42fe-b884-b8433011967e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://464e279c5903770c34d80f31ed231b194d5628b39055c4058908039d10690059\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://efa1b62728466ef7aac0e745230348ce920eb9ec9b07fb2cd5a276150dac13fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fedc40f0217626c56d155fc7e78e0592556fef9464db2aaf09c5facc486aad8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78fac300cbe2ac52f5788306b8fc0011f6ba3a22c3546974e70998f2d394987a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:02Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:02 crc kubenswrapper[4824]: I1006 09:59:02.843286 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:02Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:02 crc kubenswrapper[4824]: I1006 09:59:02.864762 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:02 crc kubenswrapper[4824]: I1006 09:59:02.864821 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:02 crc kubenswrapper[4824]: I1006 09:59:02.864833 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:02 crc kubenswrapper[4824]: I1006 09:59:02.864853 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:02 crc kubenswrapper[4824]: I1006 09:59:02.864867 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:02Z","lastTransitionTime":"2025-10-06T09:59:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:02 crc kubenswrapper[4824]: I1006 09:59:02.868590 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-szn8q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b65abf9f-4e88-4571-960d-3ca997d9c344\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://664bf42e1324a9c9529f36f08d61e26999b3386bdd1fac77851fb7e7411e8934\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49s5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-szn8q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:02Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:02 crc kubenswrapper[4824]: I1006 09:59:02.899289 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f511ece-25eb-465f-b85c-9e5f3f886c21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10bfed3e523b7422ce1e981f98166eca78bd12dd4a7cf0634915e1f4ecc3de2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43199f86ad4a1f02bc57a926cd96b2c3c0e2ea46fb36a3f495668706be4c8041\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d6064ad3bcdafebb4c66e7d91f183cc787e9ce13660545c71a722b3ff04ce75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6213164d366243c2139d5971041d6f076cf38fd0cce4c0ca5495111963487ec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa6d59fac987f60a4340ab35d84aa059f2b7085a06362e7207a79702d17eee6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09638fbc9d163ad0b63fa57d24ea5298207ad1f55bab8671be311960be2627a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33caf222d92beb8a42c203fc61d69e804ec3a22d6889f8ead2ce783237d8b464\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33caf222d92beb8a42c203fc61d69e804ec3a22d6889f8ead2ce783237d8b464\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-06T09:59:01Z\\\",\\\"message\\\":\\\"nt-go/informers/factory.go:160\\\\nI1006 09:59:01.318023 6499 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1006 09:59:01.317995 6499 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1006 09:59:01.318223 6499 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1006 09:59:01.318391 6499 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1006 09:59:01.319128 6499 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1006 09:59:01.319521 6499 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1006 09:59:01.319562 6499 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1006 09:59:01.319610 6499 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1006 09:59:01.319617 6499 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1006 09:59:01.319629 6499 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1006 09:59:01.319666 6499 factory.go:656] Stopping watch factory\\\\nI1006 09:59:01.319698 6499 ovnkube.go:599] Stopped ovnkube\\\\nI1006 09:59:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T09:59:00Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-4vwms_openshift-ovn-kubernetes(2f511ece-25eb-465f-b85c-9e5f3f886c21)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0644c7db28c7d26aa8d69cd5b9fbcff4d4c06b96df113137dea22a72d9fecd85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6da4c07be4d507a6f20bd8b42be6e6ea4712a956f57ef164036cfded7142a87e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6da4c07be4d507a6f20bd8b42be6e6ea4712a956f57ef164036cfded7142a87e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4vwms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:02Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:02 crc kubenswrapper[4824]: I1006 09:59:02.916591 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lszvg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0cf70ab4-4cfe-41c5-8db7-035451bafcfa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7b80b5b375b3b32132501ee0d661e06e199654c696ac76ad56e7ba96918cc77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fkbf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://519653c6b241cd89ef3ece003667d420a7a55cdf513f4577a69e14ade5fb4c76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fkbf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lszvg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:02Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:02 crc kubenswrapper[4824]: I1006 09:59:02.942329 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-7c22q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7eaaa68-d971-4a1c-b6c5-551ded8ff27f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f5150c8c6cfc65ecdcd0e7bd52199dd60e2472c96874f103b49919c81b5e5f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53336c7d3c509b4e022708998aedc5dee9fdaf7136ae3ee69b4181394c3a22fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53336c7d3c509b4e022708998aedc5dee9fdaf7136ae3ee69b4181394c3a22fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://652ab0a13a2bb421872edda3f7188dce8e296247052550c937f8447fea206756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://652ab0a13a2bb421872edda3f7188dce8e296247052550c937f8447fea206756\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3514dec0386bb8c016a7f555d7191ea420d5058b0ee98f97d965dd857c5376e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3514dec0386bb8c016a7f555d7191ea420d5058b0ee98f97d965dd857c5376e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b035e0bf34e3429400ac8942124a3e9834231899487d03d15609fe5624747a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b035e0bf34e3429400ac8942124a3e9834231899487d03d15609fe5624747a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6743e027a11864ad3f433d110e60451fb36d49791e03b22d437b70f89d056ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6743e027a11864ad3f433d110e60451fb36d49791e03b22d437b70f89d056ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8ea48b6886fd39a1c5e5b29ca1fffc2bd4552d61dc621eb2fba66fb464cf120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8ea48b6886fd39a1c5e5b29ca1fffc2bd4552d61dc621eb2fba66fb464cf120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-7c22q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:02Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:02 crc kubenswrapper[4824]: I1006 09:59:02.964488 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2035278cf272827b39a45aa41ecd5b1e41aafc1d9eb8bccd853660fb6e587159\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff3152fceadf36a55578f587623f37d748f90f2afeda5702a05504b372d399be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:02Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:02 crc kubenswrapper[4824]: I1006 09:59:02.969764 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:02 crc kubenswrapper[4824]: I1006 09:59:02.969863 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:02 crc kubenswrapper[4824]: I1006 09:59:02.969953 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:02 crc kubenswrapper[4824]: I1006 09:59:02.970014 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:02 crc kubenswrapper[4824]: I1006 09:59:02.970052 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:02Z","lastTransitionTime":"2025-10-06T09:59:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:02 crc kubenswrapper[4824]: I1006 09:59:02.983663 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xcgwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74317f75-9ebc-4327-b88c-598c2241240f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8def09a126f9bb910affad9b7fb32303b960add9fbfd5bb709d59b4aeba7d245\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v6sg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xcgwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:02Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:03 crc kubenswrapper[4824]: I1006 09:59:03.000134 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1140c8c-93f1-4fce-8c06-a69561ff8a6d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5462cb77b7d61aed7d4b4be2fbc6ba1451d267a66afd6dede2f259c15230e7c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gq49w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5512be36013baeca611ee92645a1d339da237b70d63aac43334beafe24513d44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gq49w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-khgzw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:02Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:03 crc kubenswrapper[4824]: I1006 09:59:03.014511 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5rx6l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0c320fc-94d9-4d82-81ee-ccccfd125efc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f8be7bad4e9644494dfb42380c7f5481e1039a1d27ad379c623515948d87392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njtbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5rx6l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:03Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:03 crc kubenswrapper[4824]: I1006 09:59:03.034499 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f9011f6673310dffea3acd7bdc2a56f2af91a87641631490fdc63da610f047a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:03Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:03 crc kubenswrapper[4824]: I1006 09:59:03.054006 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:03Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:03 crc kubenswrapper[4824]: I1006 09:59:03.072650 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:03Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:03 crc kubenswrapper[4824]: I1006 09:59:03.075123 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:03 crc kubenswrapper[4824]: I1006 09:59:03.075169 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:03 crc kubenswrapper[4824]: I1006 09:59:03.075180 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:03 crc kubenswrapper[4824]: I1006 09:59:03.075207 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:03 crc kubenswrapper[4824]: I1006 09:59:03.075219 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:03Z","lastTransitionTime":"2025-10-06T09:59:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:03 crc kubenswrapper[4824]: I1006 09:59:03.089800 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab36ed34d6ca02b59043c46e151019620ce31733209d072c21394a79fe36be4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:03Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:03 crc kubenswrapper[4824]: I1006 09:59:03.178478 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:03 crc kubenswrapper[4824]: I1006 09:59:03.178542 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:03 crc kubenswrapper[4824]: I1006 09:59:03.178560 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:03 crc kubenswrapper[4824]: I1006 09:59:03.178588 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:03 crc kubenswrapper[4824]: I1006 09:59:03.178609 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:03Z","lastTransitionTime":"2025-10-06T09:59:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:03 crc kubenswrapper[4824]: I1006 09:59:03.281412 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:03 crc kubenswrapper[4824]: I1006 09:59:03.281870 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:03 crc kubenswrapper[4824]: I1006 09:59:03.281888 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:03 crc kubenswrapper[4824]: I1006 09:59:03.281911 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:03 crc kubenswrapper[4824]: I1006 09:59:03.281931 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:03Z","lastTransitionTime":"2025-10-06T09:59:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:03 crc kubenswrapper[4824]: I1006 09:59:03.384247 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:03 crc kubenswrapper[4824]: I1006 09:59:03.384308 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:03 crc kubenswrapper[4824]: I1006 09:59:03.384323 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:03 crc kubenswrapper[4824]: I1006 09:59:03.384340 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:03 crc kubenswrapper[4824]: I1006 09:59:03.384352 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:03Z","lastTransitionTime":"2025-10-06T09:59:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:03 crc kubenswrapper[4824]: I1006 09:59:03.488570 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:03 crc kubenswrapper[4824]: I1006 09:59:03.488631 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:03 crc kubenswrapper[4824]: I1006 09:59:03.488648 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:03 crc kubenswrapper[4824]: I1006 09:59:03.488674 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:03 crc kubenswrapper[4824]: I1006 09:59:03.488691 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:03Z","lastTransitionTime":"2025-10-06T09:59:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:03 crc kubenswrapper[4824]: I1006 09:59:03.592467 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:03 crc kubenswrapper[4824]: I1006 09:59:03.592538 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:03 crc kubenswrapper[4824]: I1006 09:59:03.592558 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:03 crc kubenswrapper[4824]: I1006 09:59:03.592587 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:03 crc kubenswrapper[4824]: I1006 09:59:03.592606 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:03Z","lastTransitionTime":"2025-10-06T09:59:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:03 crc kubenswrapper[4824]: I1006 09:59:03.622278 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 06 09:59:03 crc kubenswrapper[4824]: I1006 09:59:03.638747 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Oct 06 09:59:03 crc kubenswrapper[4824]: I1006 09:59:03.639389 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lszvg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0cf70ab4-4cfe-41c5-8db7-035451bafcfa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7b80b5b375b3b32132501ee0d661e06e199654c696ac76ad56e7ba96918cc77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fkbf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://519653c6b241cd89ef3ece003667d420a7a55cdf513f4577a69e14ade5fb4c76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fkbf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lszvg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:03Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:03 crc kubenswrapper[4824]: I1006 09:59:03.657104 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gnw94" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7cf1d355-64cb-48a6-acfd-7d258b7afa10\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-psbq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-psbq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:46Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gnw94\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:03Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:03 crc kubenswrapper[4824]: I1006 09:59:03.679362 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b665cda3-da4e-482e-8248-023abec382f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9235c742d8c45bf667f32979c2891fece50f4ee1bd0c7cf6a753ed2d2b9ab30e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b2a920dbc506df2426232990e31fbb60d255c00be54941826a0f094af4a92f0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d516f278f362d7c45e7149de82fa835997ac198f1165c04352910d51fb674d69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c820d8bea16cff0d463ce68d2f2acb72d7a7c73372757ff104e14fcec1025a05\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b8cdf54407cc95783919eba8e92f093bbc09eac5e908c4ba0685629948c71bd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-06T09:58:27Z\\\",\\\"message\\\":\\\"arting RequestHeaderAuthRequestController\\\\nI1006 09:58:27.156944 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1006 09:58:27.156783 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1006 09:58:27.157213 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-432254888/tls.crt::/tmp/serving-cert-432254888/tls.key\\\\\\\"\\\\nI1006 09:58:27.157407 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1006 09:58:27.157443 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1006 09:58:27.157410 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-432254888/tls.crt::/tmp/serving-cert-432254888/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759744705\\\\\\\\\\\\\\\" (2025-10-06 09:58:25 +0000 UTC to 2025-11-05 09:58:26 +0000 UTC (now=2025-10-06 09:58:27.157390076 +0000 UTC))\\\\\\\"\\\\nI1006 09:58:27.157632 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1006 09:58:27.158158 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759744707\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759744707\\\\\\\\\\\\\\\" (2025-10-06 08:58:26 +0000 UTC to 2026-10-06 08:58:26 +0000 UTC (now=2025-10-06 09:58:27.158090663 +0000 UTC))\\\\\\\"\\\\nI1006 09:58:27.158220 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1006 09:58:27.158270 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1006 09:58:27.158310 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nF1006 09:58:27.158617 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://413822c99ba2866521177cce4f1894aef17c160aebb2662da91824a04827d510\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e37e8c7400f7fd14a5c5390a5985e3d8d277e9c5ca60d5e9e9757d07d1778570\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e37e8c7400f7fd14a5c5390a5985e3d8d277e9c5ca60d5e9e9757d07d1778570\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:03Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:03 crc kubenswrapper[4824]: I1006 09:59:03.695018 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:03 crc kubenswrapper[4824]: I1006 09:59:03.695067 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:03 crc kubenswrapper[4824]: I1006 09:59:03.695075 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:03 crc kubenswrapper[4824]: I1006 09:59:03.695094 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:03 crc kubenswrapper[4824]: I1006 09:59:03.695106 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:03Z","lastTransitionTime":"2025-10-06T09:59:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:03 crc kubenswrapper[4824]: I1006 09:59:03.700411 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d739433-0a38-42fe-b884-b8433011967e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://464e279c5903770c34d80f31ed231b194d5628b39055c4058908039d10690059\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://efa1b62728466ef7aac0e745230348ce920eb9ec9b07fb2cd5a276150dac13fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fedc40f0217626c56d155fc7e78e0592556fef9464db2aaf09c5facc486aad8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78fac300cbe2ac52f5788306b8fc0011f6ba3a22c3546974e70998f2d394987a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:03Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:03 crc kubenswrapper[4824]: I1006 09:59:03.717527 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:03Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:03 crc kubenswrapper[4824]: I1006 09:59:03.735489 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-szn8q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b65abf9f-4e88-4571-960d-3ca997d9c344\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://664bf42e1324a9c9529f36f08d61e26999b3386bdd1fac77851fb7e7411e8934\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49s5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-szn8q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:03Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:03 crc kubenswrapper[4824]: I1006 09:59:03.768039 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f511ece-25eb-465f-b85c-9e5f3f886c21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10bfed3e523b7422ce1e981f98166eca78bd12dd4a7cf0634915e1f4ecc3de2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43199f86ad4a1f02bc57a926cd96b2c3c0e2ea46fb36a3f495668706be4c8041\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d6064ad3bcdafebb4c66e7d91f183cc787e9ce13660545c71a722b3ff04ce75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6213164d366243c2139d5971041d6f076cf38fd0cce4c0ca5495111963487ec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa6d59fac987f60a4340ab35d84aa059f2b7085a06362e7207a79702d17eee6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09638fbc9d163ad0b63fa57d24ea5298207ad1f55bab8671be311960be2627a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33caf222d92beb8a42c203fc61d69e804ec3a22d6889f8ead2ce783237d8b464\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33caf222d92beb8a42c203fc61d69e804ec3a22d6889f8ead2ce783237d8b464\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-06T09:59:01Z\\\",\\\"message\\\":\\\"nt-go/informers/factory.go:160\\\\nI1006 09:59:01.318023 6499 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1006 09:59:01.317995 6499 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1006 09:59:01.318223 6499 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1006 09:59:01.318391 6499 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1006 09:59:01.319128 6499 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1006 09:59:01.319521 6499 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1006 09:59:01.319562 6499 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1006 09:59:01.319610 6499 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1006 09:59:01.319617 6499 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1006 09:59:01.319629 6499 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1006 09:59:01.319666 6499 factory.go:656] Stopping watch factory\\\\nI1006 09:59:01.319698 6499 ovnkube.go:599] Stopped ovnkube\\\\nI1006 09:59:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T09:59:00Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-4vwms_openshift-ovn-kubernetes(2f511ece-25eb-465f-b85c-9e5f3f886c21)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0644c7db28c7d26aa8d69cd5b9fbcff4d4c06b96df113137dea22a72d9fecd85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6da4c07be4d507a6f20bd8b42be6e6ea4712a956f57ef164036cfded7142a87e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6da4c07be4d507a6f20bd8b42be6e6ea4712a956f57ef164036cfded7142a87e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4vwms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:03Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:03 crc kubenswrapper[4824]: I1006 09:59:03.792353 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-7c22q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7eaaa68-d971-4a1c-b6c5-551ded8ff27f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f5150c8c6cfc65ecdcd0e7bd52199dd60e2472c96874f103b49919c81b5e5f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53336c7d3c509b4e022708998aedc5dee9fdaf7136ae3ee69b4181394c3a22fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53336c7d3c509b4e022708998aedc5dee9fdaf7136ae3ee69b4181394c3a22fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://652ab0a13a2bb421872edda3f7188dce8e296247052550c937f8447fea206756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://652ab0a13a2bb421872edda3f7188dce8e296247052550c937f8447fea206756\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3514dec0386bb8c016a7f555d7191ea420d5058b0ee98f97d965dd857c5376e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3514dec0386bb8c016a7f555d7191ea420d5058b0ee98f97d965dd857c5376e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b035e0bf34e3429400ac8942124a3e9834231899487d03d15609fe5624747a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b035e0bf34e3429400ac8942124a3e9834231899487d03d15609fe5624747a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6743e027a11864ad3f433d110e60451fb36d49791e03b22d437b70f89d056ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6743e027a11864ad3f433d110e60451fb36d49791e03b22d437b70f89d056ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8ea48b6886fd39a1c5e5b29ca1fffc2bd4552d61dc621eb2fba66fb464cf120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8ea48b6886fd39a1c5e5b29ca1fffc2bd4552d61dc621eb2fba66fb464cf120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-7c22q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:03Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:03 crc kubenswrapper[4824]: I1006 09:59:03.798162 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:03 crc kubenswrapper[4824]: I1006 09:59:03.798222 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:03 crc kubenswrapper[4824]: I1006 09:59:03.798242 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:03 crc kubenswrapper[4824]: I1006 09:59:03.798267 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:03 crc kubenswrapper[4824]: I1006 09:59:03.798289 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:03Z","lastTransitionTime":"2025-10-06T09:59:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:03 crc kubenswrapper[4824]: I1006 09:59:03.814572 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2035278cf272827b39a45aa41ecd5b1e41aafc1d9eb8bccd853660fb6e587159\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff3152fceadf36a55578f587623f37d748f90f2afeda5702a05504b372d399be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:03Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:03 crc kubenswrapper[4824]: I1006 09:59:03.830558 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xcgwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74317f75-9ebc-4327-b88c-598c2241240f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8def09a126f9bb910affad9b7fb32303b960add9fbfd5bb709d59b4aeba7d245\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v6sg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xcgwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:03Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:03 crc kubenswrapper[4824]: I1006 09:59:03.847079 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1140c8c-93f1-4fce-8c06-a69561ff8a6d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5462cb77b7d61aed7d4b4be2fbc6ba1451d267a66afd6dede2f259c15230e7c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gq49w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5512be36013baeca611ee92645a1d339da237b70d63aac43334beafe24513d44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gq49w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-khgzw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:03Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:03 crc kubenswrapper[4824]: I1006 09:59:03.864438 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab36ed34d6ca02b59043c46e151019620ce31733209d072c21394a79fe36be4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:03Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:03 crc kubenswrapper[4824]: I1006 09:59:03.881141 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5rx6l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0c320fc-94d9-4d82-81ee-ccccfd125efc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f8be7bad4e9644494dfb42380c7f5481e1039a1d27ad379c623515948d87392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njtbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5rx6l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:03Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:03 crc kubenswrapper[4824]: I1006 09:59:03.902145 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:03 crc kubenswrapper[4824]: I1006 09:59:03.902191 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:03 crc kubenswrapper[4824]: I1006 09:59:03.902208 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:03 crc kubenswrapper[4824]: I1006 09:59:03.902239 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:03 crc kubenswrapper[4824]: I1006 09:59:03.902259 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:03Z","lastTransitionTime":"2025-10-06T09:59:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:03 crc kubenswrapper[4824]: I1006 09:59:03.902448 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f9011f6673310dffea3acd7bdc2a56f2af91a87641631490fdc63da610f047a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:03Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:03 crc kubenswrapper[4824]: I1006 09:59:03.924818 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:03Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:03 crc kubenswrapper[4824]: I1006 09:59:03.947125 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:03Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:04 crc kubenswrapper[4824]: I1006 09:59:04.006063 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:04 crc kubenswrapper[4824]: I1006 09:59:04.006117 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:04 crc kubenswrapper[4824]: I1006 09:59:04.006136 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:04 crc kubenswrapper[4824]: I1006 09:59:04.006160 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:04 crc kubenswrapper[4824]: I1006 09:59:04.006179 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:04Z","lastTransitionTime":"2025-10-06T09:59:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:04 crc kubenswrapper[4824]: I1006 09:59:04.109490 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:04 crc kubenswrapper[4824]: I1006 09:59:04.109533 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:04 crc kubenswrapper[4824]: I1006 09:59:04.109543 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:04 crc kubenswrapper[4824]: I1006 09:59:04.109561 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:04 crc kubenswrapper[4824]: I1006 09:59:04.109574 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:04Z","lastTransitionTime":"2025-10-06T09:59:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:04 crc kubenswrapper[4824]: I1006 09:59:04.212699 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:04 crc kubenswrapper[4824]: I1006 09:59:04.212753 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:04 crc kubenswrapper[4824]: I1006 09:59:04.212763 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:04 crc kubenswrapper[4824]: I1006 09:59:04.212781 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:04 crc kubenswrapper[4824]: I1006 09:59:04.212792 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:04Z","lastTransitionTime":"2025-10-06T09:59:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:04 crc kubenswrapper[4824]: I1006 09:59:04.273840 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 09:59:04 crc kubenswrapper[4824]: I1006 09:59:04.273916 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 09:59:04 crc kubenswrapper[4824]: I1006 09:59:04.273969 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 09:59:04 crc kubenswrapper[4824]: E1006 09:59:04.274111 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 06 09:59:04 crc kubenswrapper[4824]: E1006 09:59:04.274248 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 06 09:59:04 crc kubenswrapper[4824]: I1006 09:59:04.273883 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gnw94" Oct 06 09:59:04 crc kubenswrapper[4824]: E1006 09:59:04.274418 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 06 09:59:04 crc kubenswrapper[4824]: E1006 09:59:04.274579 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gnw94" podUID="7cf1d355-64cb-48a6-acfd-7d258b7afa10" Oct 06 09:59:04 crc kubenswrapper[4824]: I1006 09:59:04.315833 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:04 crc kubenswrapper[4824]: I1006 09:59:04.315908 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:04 crc kubenswrapper[4824]: I1006 09:59:04.315929 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:04 crc kubenswrapper[4824]: I1006 09:59:04.315958 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:04 crc kubenswrapper[4824]: I1006 09:59:04.316004 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:04Z","lastTransitionTime":"2025-10-06T09:59:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:04 crc kubenswrapper[4824]: I1006 09:59:04.419999 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:04 crc kubenswrapper[4824]: I1006 09:59:04.420066 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:04 crc kubenswrapper[4824]: I1006 09:59:04.420077 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:04 crc kubenswrapper[4824]: I1006 09:59:04.420100 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:04 crc kubenswrapper[4824]: I1006 09:59:04.420114 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:04Z","lastTransitionTime":"2025-10-06T09:59:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:04 crc kubenswrapper[4824]: I1006 09:59:04.523177 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:04 crc kubenswrapper[4824]: I1006 09:59:04.523805 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:04 crc kubenswrapper[4824]: I1006 09:59:04.524021 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:04 crc kubenswrapper[4824]: I1006 09:59:04.524204 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:04 crc kubenswrapper[4824]: I1006 09:59:04.524358 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:04Z","lastTransitionTime":"2025-10-06T09:59:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:04 crc kubenswrapper[4824]: I1006 09:59:04.628518 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:04 crc kubenswrapper[4824]: I1006 09:59:04.629100 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:04 crc kubenswrapper[4824]: I1006 09:59:04.629329 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:04 crc kubenswrapper[4824]: I1006 09:59:04.629735 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:04 crc kubenswrapper[4824]: I1006 09:59:04.629894 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:04Z","lastTransitionTime":"2025-10-06T09:59:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:04 crc kubenswrapper[4824]: I1006 09:59:04.733616 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:04 crc kubenswrapper[4824]: I1006 09:59:04.733685 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:04 crc kubenswrapper[4824]: I1006 09:59:04.733705 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:04 crc kubenswrapper[4824]: I1006 09:59:04.733734 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:04 crc kubenswrapper[4824]: I1006 09:59:04.733754 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:04Z","lastTransitionTime":"2025-10-06T09:59:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:04 crc kubenswrapper[4824]: I1006 09:59:04.837155 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:04 crc kubenswrapper[4824]: I1006 09:59:04.837218 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:04 crc kubenswrapper[4824]: I1006 09:59:04.837230 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:04 crc kubenswrapper[4824]: I1006 09:59:04.837247 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:04 crc kubenswrapper[4824]: I1006 09:59:04.837260 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:04Z","lastTransitionTime":"2025-10-06T09:59:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:04 crc kubenswrapper[4824]: I1006 09:59:04.940913 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:04 crc kubenswrapper[4824]: I1006 09:59:04.941006 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:04 crc kubenswrapper[4824]: I1006 09:59:04.941020 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:04 crc kubenswrapper[4824]: I1006 09:59:04.941044 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:04 crc kubenswrapper[4824]: I1006 09:59:04.941061 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:04Z","lastTransitionTime":"2025-10-06T09:59:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:05 crc kubenswrapper[4824]: I1006 09:59:05.044152 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:05 crc kubenswrapper[4824]: I1006 09:59:05.044215 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:05 crc kubenswrapper[4824]: I1006 09:59:05.044237 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:05 crc kubenswrapper[4824]: I1006 09:59:05.044263 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:05 crc kubenswrapper[4824]: I1006 09:59:05.044282 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:05Z","lastTransitionTime":"2025-10-06T09:59:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:05 crc kubenswrapper[4824]: I1006 09:59:05.148229 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:05 crc kubenswrapper[4824]: I1006 09:59:05.148302 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:05 crc kubenswrapper[4824]: I1006 09:59:05.148325 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:05 crc kubenswrapper[4824]: I1006 09:59:05.148358 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:05 crc kubenswrapper[4824]: I1006 09:59:05.148385 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:05Z","lastTransitionTime":"2025-10-06T09:59:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:05 crc kubenswrapper[4824]: I1006 09:59:05.251538 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:05 crc kubenswrapper[4824]: I1006 09:59:05.251592 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:05 crc kubenswrapper[4824]: I1006 09:59:05.251606 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:05 crc kubenswrapper[4824]: I1006 09:59:05.251626 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:05 crc kubenswrapper[4824]: I1006 09:59:05.251641 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:05Z","lastTransitionTime":"2025-10-06T09:59:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:05 crc kubenswrapper[4824]: I1006 09:59:05.354528 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:05 crc kubenswrapper[4824]: I1006 09:59:05.354602 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:05 crc kubenswrapper[4824]: I1006 09:59:05.354622 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:05 crc kubenswrapper[4824]: I1006 09:59:05.354654 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:05 crc kubenswrapper[4824]: I1006 09:59:05.354677 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:05Z","lastTransitionTime":"2025-10-06T09:59:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:05 crc kubenswrapper[4824]: I1006 09:59:05.457962 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:05 crc kubenswrapper[4824]: I1006 09:59:05.458037 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:05 crc kubenswrapper[4824]: I1006 09:59:05.458051 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:05 crc kubenswrapper[4824]: I1006 09:59:05.458069 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:05 crc kubenswrapper[4824]: I1006 09:59:05.458082 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:05Z","lastTransitionTime":"2025-10-06T09:59:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:05 crc kubenswrapper[4824]: I1006 09:59:05.560654 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:05 crc kubenswrapper[4824]: I1006 09:59:05.560731 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:05 crc kubenswrapper[4824]: I1006 09:59:05.560752 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:05 crc kubenswrapper[4824]: I1006 09:59:05.560779 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:05 crc kubenswrapper[4824]: I1006 09:59:05.560798 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:05Z","lastTransitionTime":"2025-10-06T09:59:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:05 crc kubenswrapper[4824]: I1006 09:59:05.664632 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:05 crc kubenswrapper[4824]: I1006 09:59:05.664699 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:05 crc kubenswrapper[4824]: I1006 09:59:05.664717 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:05 crc kubenswrapper[4824]: I1006 09:59:05.664744 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:05 crc kubenswrapper[4824]: I1006 09:59:05.664765 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:05Z","lastTransitionTime":"2025-10-06T09:59:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:05 crc kubenswrapper[4824]: I1006 09:59:05.768433 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:05 crc kubenswrapper[4824]: I1006 09:59:05.768504 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:05 crc kubenswrapper[4824]: I1006 09:59:05.768521 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:05 crc kubenswrapper[4824]: I1006 09:59:05.768547 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:05 crc kubenswrapper[4824]: I1006 09:59:05.768570 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:05Z","lastTransitionTime":"2025-10-06T09:59:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:05 crc kubenswrapper[4824]: I1006 09:59:05.872250 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:05 crc kubenswrapper[4824]: I1006 09:59:05.872335 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:05 crc kubenswrapper[4824]: I1006 09:59:05.872360 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:05 crc kubenswrapper[4824]: I1006 09:59:05.872397 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:05 crc kubenswrapper[4824]: I1006 09:59:05.872422 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:05Z","lastTransitionTime":"2025-10-06T09:59:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:05 crc kubenswrapper[4824]: I1006 09:59:05.976155 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:05 crc kubenswrapper[4824]: I1006 09:59:05.976244 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:05 crc kubenswrapper[4824]: I1006 09:59:05.976262 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:05 crc kubenswrapper[4824]: I1006 09:59:05.976293 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:05 crc kubenswrapper[4824]: I1006 09:59:05.976312 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:05Z","lastTransitionTime":"2025-10-06T09:59:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:06 crc kubenswrapper[4824]: I1006 09:59:06.078963 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:06 crc kubenswrapper[4824]: I1006 09:59:06.079025 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:06 crc kubenswrapper[4824]: I1006 09:59:06.079036 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:06 crc kubenswrapper[4824]: I1006 09:59:06.079054 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:06 crc kubenswrapper[4824]: I1006 09:59:06.079068 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:06Z","lastTransitionTime":"2025-10-06T09:59:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:06 crc kubenswrapper[4824]: I1006 09:59:06.183944 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:06 crc kubenswrapper[4824]: I1006 09:59:06.184052 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:06 crc kubenswrapper[4824]: I1006 09:59:06.184073 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:06 crc kubenswrapper[4824]: I1006 09:59:06.184103 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:06 crc kubenswrapper[4824]: I1006 09:59:06.184123 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:06Z","lastTransitionTime":"2025-10-06T09:59:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:06 crc kubenswrapper[4824]: I1006 09:59:06.273754 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gnw94" Oct 06 09:59:06 crc kubenswrapper[4824]: I1006 09:59:06.273861 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 09:59:06 crc kubenswrapper[4824]: I1006 09:59:06.273779 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 09:59:06 crc kubenswrapper[4824]: I1006 09:59:06.273929 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 09:59:06 crc kubenswrapper[4824]: E1006 09:59:06.274032 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gnw94" podUID="7cf1d355-64cb-48a6-acfd-7d258b7afa10" Oct 06 09:59:06 crc kubenswrapper[4824]: E1006 09:59:06.274184 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 06 09:59:06 crc kubenswrapper[4824]: E1006 09:59:06.274358 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 06 09:59:06 crc kubenswrapper[4824]: E1006 09:59:06.274559 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 06 09:59:06 crc kubenswrapper[4824]: I1006 09:59:06.287580 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:06 crc kubenswrapper[4824]: I1006 09:59:06.287633 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:06 crc kubenswrapper[4824]: I1006 09:59:06.287651 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:06 crc kubenswrapper[4824]: I1006 09:59:06.287679 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:06 crc kubenswrapper[4824]: I1006 09:59:06.287698 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:06Z","lastTransitionTime":"2025-10-06T09:59:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:06 crc kubenswrapper[4824]: I1006 09:59:06.391575 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:06 crc kubenswrapper[4824]: I1006 09:59:06.391665 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:06 crc kubenswrapper[4824]: I1006 09:59:06.391685 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:06 crc kubenswrapper[4824]: I1006 09:59:06.391715 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:06 crc kubenswrapper[4824]: I1006 09:59:06.391736 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:06Z","lastTransitionTime":"2025-10-06T09:59:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:06 crc kubenswrapper[4824]: I1006 09:59:06.495875 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:06 crc kubenswrapper[4824]: I1006 09:59:06.495944 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:06 crc kubenswrapper[4824]: I1006 09:59:06.495965 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:06 crc kubenswrapper[4824]: I1006 09:59:06.496035 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:06 crc kubenswrapper[4824]: I1006 09:59:06.496059 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:06Z","lastTransitionTime":"2025-10-06T09:59:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:06 crc kubenswrapper[4824]: I1006 09:59:06.599492 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:06 crc kubenswrapper[4824]: I1006 09:59:06.599556 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:06 crc kubenswrapper[4824]: I1006 09:59:06.599578 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:06 crc kubenswrapper[4824]: I1006 09:59:06.599610 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:06 crc kubenswrapper[4824]: I1006 09:59:06.599637 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:06Z","lastTransitionTime":"2025-10-06T09:59:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:06 crc kubenswrapper[4824]: I1006 09:59:06.703212 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:06 crc kubenswrapper[4824]: I1006 09:59:06.703284 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:06 crc kubenswrapper[4824]: I1006 09:59:06.703302 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:06 crc kubenswrapper[4824]: I1006 09:59:06.703330 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:06 crc kubenswrapper[4824]: I1006 09:59:06.703351 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:06Z","lastTransitionTime":"2025-10-06T09:59:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:06 crc kubenswrapper[4824]: I1006 09:59:06.806787 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:06 crc kubenswrapper[4824]: I1006 09:59:06.806839 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:06 crc kubenswrapper[4824]: I1006 09:59:06.806856 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:06 crc kubenswrapper[4824]: I1006 09:59:06.806883 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:06 crc kubenswrapper[4824]: I1006 09:59:06.806902 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:06Z","lastTransitionTime":"2025-10-06T09:59:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:06 crc kubenswrapper[4824]: I1006 09:59:06.910043 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:06 crc kubenswrapper[4824]: I1006 09:59:06.910109 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:06 crc kubenswrapper[4824]: I1006 09:59:06.910126 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:06 crc kubenswrapper[4824]: I1006 09:59:06.910149 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:06 crc kubenswrapper[4824]: I1006 09:59:06.910167 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:06Z","lastTransitionTime":"2025-10-06T09:59:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:07 crc kubenswrapper[4824]: I1006 09:59:07.013847 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:07 crc kubenswrapper[4824]: I1006 09:59:07.013917 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:07 crc kubenswrapper[4824]: I1006 09:59:07.013931 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:07 crc kubenswrapper[4824]: I1006 09:59:07.013954 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:07 crc kubenswrapper[4824]: I1006 09:59:07.013968 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:07Z","lastTransitionTime":"2025-10-06T09:59:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:07 crc kubenswrapper[4824]: I1006 09:59:07.117826 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:07 crc kubenswrapper[4824]: I1006 09:59:07.117893 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:07 crc kubenswrapper[4824]: I1006 09:59:07.117912 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:07 crc kubenswrapper[4824]: I1006 09:59:07.117937 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:07 crc kubenswrapper[4824]: I1006 09:59:07.117957 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:07Z","lastTransitionTime":"2025-10-06T09:59:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:07 crc kubenswrapper[4824]: I1006 09:59:07.221441 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:07 crc kubenswrapper[4824]: I1006 09:59:07.221519 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:07 crc kubenswrapper[4824]: I1006 09:59:07.221542 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:07 crc kubenswrapper[4824]: I1006 09:59:07.221572 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:07 crc kubenswrapper[4824]: I1006 09:59:07.221593 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:07Z","lastTransitionTime":"2025-10-06T09:59:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:07 crc kubenswrapper[4824]: I1006 09:59:07.325212 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:07 crc kubenswrapper[4824]: I1006 09:59:07.325299 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:07 crc kubenswrapper[4824]: I1006 09:59:07.325330 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:07 crc kubenswrapper[4824]: I1006 09:59:07.325368 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:07 crc kubenswrapper[4824]: I1006 09:59:07.325394 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:07Z","lastTransitionTime":"2025-10-06T09:59:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:07 crc kubenswrapper[4824]: I1006 09:59:07.429927 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:07 crc kubenswrapper[4824]: I1006 09:59:07.430041 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:07 crc kubenswrapper[4824]: I1006 09:59:07.430063 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:07 crc kubenswrapper[4824]: I1006 09:59:07.430092 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:07 crc kubenswrapper[4824]: I1006 09:59:07.430110 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:07Z","lastTransitionTime":"2025-10-06T09:59:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:07 crc kubenswrapper[4824]: I1006 09:59:07.503548 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:07 crc kubenswrapper[4824]: I1006 09:59:07.503619 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:07 crc kubenswrapper[4824]: I1006 09:59:07.503637 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:07 crc kubenswrapper[4824]: I1006 09:59:07.503664 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:07 crc kubenswrapper[4824]: I1006 09:59:07.503683 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:07Z","lastTransitionTime":"2025-10-06T09:59:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:07 crc kubenswrapper[4824]: E1006 09:59:07.526822 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T09:59:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T09:59:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T09:59:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T09:59:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T09:59:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T09:59:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T09:59:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T09:59:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"740c2e7e-c94c-411f-b080-eb42c9aecdca\\\",\\\"systemUUID\\\":\\\"77448683-227c-4c6b-a334-eb212350758a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:07Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:07 crc kubenswrapper[4824]: I1006 09:59:07.536385 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:07 crc kubenswrapper[4824]: I1006 09:59:07.536460 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:07 crc kubenswrapper[4824]: I1006 09:59:07.536476 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:07 crc kubenswrapper[4824]: I1006 09:59:07.536503 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:07 crc kubenswrapper[4824]: I1006 09:59:07.536522 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:07Z","lastTransitionTime":"2025-10-06T09:59:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:07 crc kubenswrapper[4824]: E1006 09:59:07.558327 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T09:59:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T09:59:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T09:59:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T09:59:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T09:59:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T09:59:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T09:59:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T09:59:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"740c2e7e-c94c-411f-b080-eb42c9aecdca\\\",\\\"systemUUID\\\":\\\"77448683-227c-4c6b-a334-eb212350758a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:07Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:07 crc kubenswrapper[4824]: I1006 09:59:07.563842 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:07 crc kubenswrapper[4824]: I1006 09:59:07.563895 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:07 crc kubenswrapper[4824]: I1006 09:59:07.563916 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:07 crc kubenswrapper[4824]: I1006 09:59:07.563940 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:07 crc kubenswrapper[4824]: I1006 09:59:07.563959 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:07Z","lastTransitionTime":"2025-10-06T09:59:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:07 crc kubenswrapper[4824]: E1006 09:59:07.584306 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T09:59:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T09:59:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T09:59:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T09:59:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T09:59:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T09:59:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T09:59:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T09:59:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"740c2e7e-c94c-411f-b080-eb42c9aecdca\\\",\\\"systemUUID\\\":\\\"77448683-227c-4c6b-a334-eb212350758a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:07Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:07 crc kubenswrapper[4824]: I1006 09:59:07.589593 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:07 crc kubenswrapper[4824]: I1006 09:59:07.589655 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:07 crc kubenswrapper[4824]: I1006 09:59:07.589680 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:07 crc kubenswrapper[4824]: I1006 09:59:07.589710 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:07 crc kubenswrapper[4824]: I1006 09:59:07.589733 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:07Z","lastTransitionTime":"2025-10-06T09:59:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:07 crc kubenswrapper[4824]: E1006 09:59:07.611190 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T09:59:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T09:59:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T09:59:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T09:59:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T09:59:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T09:59:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T09:59:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T09:59:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"740c2e7e-c94c-411f-b080-eb42c9aecdca\\\",\\\"systemUUID\\\":\\\"77448683-227c-4c6b-a334-eb212350758a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:07Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:07 crc kubenswrapper[4824]: I1006 09:59:07.616663 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:07 crc kubenswrapper[4824]: I1006 09:59:07.616718 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:07 crc kubenswrapper[4824]: I1006 09:59:07.616737 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:07 crc kubenswrapper[4824]: I1006 09:59:07.616764 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:07 crc kubenswrapper[4824]: I1006 09:59:07.616782 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:07Z","lastTransitionTime":"2025-10-06T09:59:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:07 crc kubenswrapper[4824]: E1006 09:59:07.639133 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T09:59:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T09:59:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T09:59:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T09:59:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T09:59:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T09:59:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T09:59:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T09:59:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"740c2e7e-c94c-411f-b080-eb42c9aecdca\\\",\\\"systemUUID\\\":\\\"77448683-227c-4c6b-a334-eb212350758a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:07Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:07 crc kubenswrapper[4824]: E1006 09:59:07.639379 4824 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 06 09:59:07 crc kubenswrapper[4824]: I1006 09:59:07.641487 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:07 crc kubenswrapper[4824]: I1006 09:59:07.641558 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:07 crc kubenswrapper[4824]: I1006 09:59:07.641582 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:07 crc kubenswrapper[4824]: I1006 09:59:07.641616 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:07 crc kubenswrapper[4824]: I1006 09:59:07.641645 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:07Z","lastTransitionTime":"2025-10-06T09:59:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:07 crc kubenswrapper[4824]: I1006 09:59:07.745015 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:07 crc kubenswrapper[4824]: I1006 09:59:07.745066 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:07 crc kubenswrapper[4824]: I1006 09:59:07.745085 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:07 crc kubenswrapper[4824]: I1006 09:59:07.745110 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:07 crc kubenswrapper[4824]: I1006 09:59:07.745129 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:07Z","lastTransitionTime":"2025-10-06T09:59:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:07 crc kubenswrapper[4824]: I1006 09:59:07.848159 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:07 crc kubenswrapper[4824]: I1006 09:59:07.848191 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:07 crc kubenswrapper[4824]: I1006 09:59:07.848201 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:07 crc kubenswrapper[4824]: I1006 09:59:07.848217 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:07 crc kubenswrapper[4824]: I1006 09:59:07.848232 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:07Z","lastTransitionTime":"2025-10-06T09:59:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:07 crc kubenswrapper[4824]: I1006 09:59:07.952018 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:07 crc kubenswrapper[4824]: I1006 09:59:07.952075 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:07 crc kubenswrapper[4824]: I1006 09:59:07.952092 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:07 crc kubenswrapper[4824]: I1006 09:59:07.952120 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:07 crc kubenswrapper[4824]: I1006 09:59:07.952140 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:07Z","lastTransitionTime":"2025-10-06T09:59:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:08 crc kubenswrapper[4824]: I1006 09:59:08.054248 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:08 crc kubenswrapper[4824]: I1006 09:59:08.054304 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:08 crc kubenswrapper[4824]: I1006 09:59:08.054322 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:08 crc kubenswrapper[4824]: I1006 09:59:08.054347 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:08 crc kubenswrapper[4824]: I1006 09:59:08.054366 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:08Z","lastTransitionTime":"2025-10-06T09:59:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:08 crc kubenswrapper[4824]: I1006 09:59:08.156651 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:08 crc kubenswrapper[4824]: I1006 09:59:08.156700 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:08 crc kubenswrapper[4824]: I1006 09:59:08.156710 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:08 crc kubenswrapper[4824]: I1006 09:59:08.156721 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:08 crc kubenswrapper[4824]: I1006 09:59:08.156730 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:08Z","lastTransitionTime":"2025-10-06T09:59:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:08 crc kubenswrapper[4824]: I1006 09:59:08.259674 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:08 crc kubenswrapper[4824]: I1006 09:59:08.259747 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:08 crc kubenswrapper[4824]: I1006 09:59:08.259760 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:08 crc kubenswrapper[4824]: I1006 09:59:08.259783 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:08 crc kubenswrapper[4824]: I1006 09:59:08.259798 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:08Z","lastTransitionTime":"2025-10-06T09:59:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:08 crc kubenswrapper[4824]: I1006 09:59:08.274210 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 09:59:08 crc kubenswrapper[4824]: I1006 09:59:08.274348 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 09:59:08 crc kubenswrapper[4824]: I1006 09:59:08.274406 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 09:59:08 crc kubenswrapper[4824]: E1006 09:59:08.274429 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 06 09:59:08 crc kubenswrapper[4824]: I1006 09:59:08.274425 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gnw94" Oct 06 09:59:08 crc kubenswrapper[4824]: E1006 09:59:08.274570 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 06 09:59:08 crc kubenswrapper[4824]: E1006 09:59:08.274880 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 06 09:59:08 crc kubenswrapper[4824]: E1006 09:59:08.275024 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gnw94" podUID="7cf1d355-64cb-48a6-acfd-7d258b7afa10" Oct 06 09:59:08 crc kubenswrapper[4824]: I1006 09:59:08.363494 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:08 crc kubenswrapper[4824]: I1006 09:59:08.363555 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:08 crc kubenswrapper[4824]: I1006 09:59:08.363567 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:08 crc kubenswrapper[4824]: I1006 09:59:08.363586 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:08 crc kubenswrapper[4824]: I1006 09:59:08.363597 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:08Z","lastTransitionTime":"2025-10-06T09:59:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:08 crc kubenswrapper[4824]: I1006 09:59:08.472882 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:08 crc kubenswrapper[4824]: I1006 09:59:08.472948 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:08 crc kubenswrapper[4824]: I1006 09:59:08.472963 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:08 crc kubenswrapper[4824]: I1006 09:59:08.473009 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:08 crc kubenswrapper[4824]: I1006 09:59:08.473023 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:08Z","lastTransitionTime":"2025-10-06T09:59:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:08 crc kubenswrapper[4824]: I1006 09:59:08.577149 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:08 crc kubenswrapper[4824]: I1006 09:59:08.577219 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:08 crc kubenswrapper[4824]: I1006 09:59:08.577238 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:08 crc kubenswrapper[4824]: I1006 09:59:08.577269 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:08 crc kubenswrapper[4824]: I1006 09:59:08.577292 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:08Z","lastTransitionTime":"2025-10-06T09:59:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:08 crc kubenswrapper[4824]: I1006 09:59:08.680844 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:08 crc kubenswrapper[4824]: I1006 09:59:08.680902 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:08 crc kubenswrapper[4824]: I1006 09:59:08.680919 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:08 crc kubenswrapper[4824]: I1006 09:59:08.680946 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:08 crc kubenswrapper[4824]: I1006 09:59:08.680965 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:08Z","lastTransitionTime":"2025-10-06T09:59:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:08 crc kubenswrapper[4824]: I1006 09:59:08.783692 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:08 crc kubenswrapper[4824]: I1006 09:59:08.783782 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:08 crc kubenswrapper[4824]: I1006 09:59:08.783807 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:08 crc kubenswrapper[4824]: I1006 09:59:08.783844 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:08 crc kubenswrapper[4824]: I1006 09:59:08.783868 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:08Z","lastTransitionTime":"2025-10-06T09:59:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:08 crc kubenswrapper[4824]: I1006 09:59:08.887754 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:08 crc kubenswrapper[4824]: I1006 09:59:08.887815 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:08 crc kubenswrapper[4824]: I1006 09:59:08.887834 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:08 crc kubenswrapper[4824]: I1006 09:59:08.887864 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:08 crc kubenswrapper[4824]: I1006 09:59:08.887885 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:08Z","lastTransitionTime":"2025-10-06T09:59:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:09 crc kubenswrapper[4824]: I1006 09:59:09.005061 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:09 crc kubenswrapper[4824]: I1006 09:59:09.005551 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:09 crc kubenswrapper[4824]: I1006 09:59:09.005713 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:09 crc kubenswrapper[4824]: I1006 09:59:09.005873 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:09 crc kubenswrapper[4824]: I1006 09:59:09.006060 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:09Z","lastTransitionTime":"2025-10-06T09:59:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:09 crc kubenswrapper[4824]: I1006 09:59:09.109565 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:09 crc kubenswrapper[4824]: I1006 09:59:09.109957 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:09 crc kubenswrapper[4824]: I1006 09:59:09.110190 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:09 crc kubenswrapper[4824]: I1006 09:59:09.110381 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:09 crc kubenswrapper[4824]: I1006 09:59:09.110533 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:09Z","lastTransitionTime":"2025-10-06T09:59:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:09 crc kubenswrapper[4824]: I1006 09:59:09.214238 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:09 crc kubenswrapper[4824]: I1006 09:59:09.214757 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:09 crc kubenswrapper[4824]: I1006 09:59:09.215015 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:09 crc kubenswrapper[4824]: I1006 09:59:09.215241 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:09 crc kubenswrapper[4824]: I1006 09:59:09.215410 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:09Z","lastTransitionTime":"2025-10-06T09:59:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:09 crc kubenswrapper[4824]: I1006 09:59:09.318059 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:09 crc kubenswrapper[4824]: I1006 09:59:09.318135 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:09 crc kubenswrapper[4824]: I1006 09:59:09.318162 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:09 crc kubenswrapper[4824]: I1006 09:59:09.318192 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:09 crc kubenswrapper[4824]: I1006 09:59:09.318216 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:09Z","lastTransitionTime":"2025-10-06T09:59:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:09 crc kubenswrapper[4824]: I1006 09:59:09.421335 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:09 crc kubenswrapper[4824]: I1006 09:59:09.421387 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:09 crc kubenswrapper[4824]: I1006 09:59:09.421398 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:09 crc kubenswrapper[4824]: I1006 09:59:09.421419 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:09 crc kubenswrapper[4824]: I1006 09:59:09.421437 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:09Z","lastTransitionTime":"2025-10-06T09:59:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:09 crc kubenswrapper[4824]: I1006 09:59:09.525402 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:09 crc kubenswrapper[4824]: I1006 09:59:09.525472 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:09 crc kubenswrapper[4824]: I1006 09:59:09.525485 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:09 crc kubenswrapper[4824]: I1006 09:59:09.525509 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:09 crc kubenswrapper[4824]: I1006 09:59:09.525523 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:09Z","lastTransitionTime":"2025-10-06T09:59:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:09 crc kubenswrapper[4824]: I1006 09:59:09.629116 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:09 crc kubenswrapper[4824]: I1006 09:59:09.629156 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:09 crc kubenswrapper[4824]: I1006 09:59:09.629167 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:09 crc kubenswrapper[4824]: I1006 09:59:09.629184 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:09 crc kubenswrapper[4824]: I1006 09:59:09.629196 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:09Z","lastTransitionTime":"2025-10-06T09:59:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:09 crc kubenswrapper[4824]: I1006 09:59:09.733263 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:09 crc kubenswrapper[4824]: I1006 09:59:09.733328 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:09 crc kubenswrapper[4824]: I1006 09:59:09.733353 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:09 crc kubenswrapper[4824]: I1006 09:59:09.733384 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:09 crc kubenswrapper[4824]: I1006 09:59:09.733407 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:09Z","lastTransitionTime":"2025-10-06T09:59:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:09 crc kubenswrapper[4824]: I1006 09:59:09.836307 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:09 crc kubenswrapper[4824]: I1006 09:59:09.836370 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:09 crc kubenswrapper[4824]: I1006 09:59:09.836386 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:09 crc kubenswrapper[4824]: I1006 09:59:09.836411 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:09 crc kubenswrapper[4824]: I1006 09:59:09.836429 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:09Z","lastTransitionTime":"2025-10-06T09:59:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:09 crc kubenswrapper[4824]: I1006 09:59:09.940024 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:09 crc kubenswrapper[4824]: I1006 09:59:09.940090 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:09 crc kubenswrapper[4824]: I1006 09:59:09.940107 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:09 crc kubenswrapper[4824]: I1006 09:59:09.940130 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:09 crc kubenswrapper[4824]: I1006 09:59:09.940147 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:09Z","lastTransitionTime":"2025-10-06T09:59:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:10 crc kubenswrapper[4824]: I1006 09:59:10.043586 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:10 crc kubenswrapper[4824]: I1006 09:59:10.043645 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:10 crc kubenswrapper[4824]: I1006 09:59:10.043658 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:10 crc kubenswrapper[4824]: I1006 09:59:10.043678 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:10 crc kubenswrapper[4824]: I1006 09:59:10.043694 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:10Z","lastTransitionTime":"2025-10-06T09:59:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:10 crc kubenswrapper[4824]: I1006 09:59:10.147218 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:10 crc kubenswrapper[4824]: I1006 09:59:10.147603 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:10 crc kubenswrapper[4824]: I1006 09:59:10.147693 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:10 crc kubenswrapper[4824]: I1006 09:59:10.147793 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:10 crc kubenswrapper[4824]: I1006 09:59:10.148184 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:10Z","lastTransitionTime":"2025-10-06T09:59:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:10 crc kubenswrapper[4824]: I1006 09:59:10.252242 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:10 crc kubenswrapper[4824]: I1006 09:59:10.252303 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:10 crc kubenswrapper[4824]: I1006 09:59:10.252322 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:10 crc kubenswrapper[4824]: I1006 09:59:10.252348 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:10 crc kubenswrapper[4824]: I1006 09:59:10.252369 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:10Z","lastTransitionTime":"2025-10-06T09:59:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:10 crc kubenswrapper[4824]: I1006 09:59:10.273862 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 09:59:10 crc kubenswrapper[4824]: I1006 09:59:10.273948 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gnw94" Oct 06 09:59:10 crc kubenswrapper[4824]: E1006 09:59:10.274078 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 06 09:59:10 crc kubenswrapper[4824]: I1006 09:59:10.274101 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 09:59:10 crc kubenswrapper[4824]: E1006 09:59:10.274222 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gnw94" podUID="7cf1d355-64cb-48a6-acfd-7d258b7afa10" Oct 06 09:59:10 crc kubenswrapper[4824]: E1006 09:59:10.274382 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 06 09:59:10 crc kubenswrapper[4824]: I1006 09:59:10.274559 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 09:59:10 crc kubenswrapper[4824]: E1006 09:59:10.274835 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 06 09:59:10 crc kubenswrapper[4824]: I1006 09:59:10.355909 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:10 crc kubenswrapper[4824]: I1006 09:59:10.355970 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:10 crc kubenswrapper[4824]: I1006 09:59:10.356022 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:10 crc kubenswrapper[4824]: I1006 09:59:10.356051 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:10 crc kubenswrapper[4824]: I1006 09:59:10.356072 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:10Z","lastTransitionTime":"2025-10-06T09:59:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:10 crc kubenswrapper[4824]: I1006 09:59:10.459909 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:10 crc kubenswrapper[4824]: I1006 09:59:10.460002 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:10 crc kubenswrapper[4824]: I1006 09:59:10.460021 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:10 crc kubenswrapper[4824]: I1006 09:59:10.460047 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:10 crc kubenswrapper[4824]: I1006 09:59:10.460066 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:10Z","lastTransitionTime":"2025-10-06T09:59:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:10 crc kubenswrapper[4824]: I1006 09:59:10.563411 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:10 crc kubenswrapper[4824]: I1006 09:59:10.563491 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:10 crc kubenswrapper[4824]: I1006 09:59:10.563513 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:10 crc kubenswrapper[4824]: I1006 09:59:10.563546 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:10 crc kubenswrapper[4824]: I1006 09:59:10.563567 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:10Z","lastTransitionTime":"2025-10-06T09:59:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:10 crc kubenswrapper[4824]: I1006 09:59:10.667254 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:10 crc kubenswrapper[4824]: I1006 09:59:10.667303 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:10 crc kubenswrapper[4824]: I1006 09:59:10.667315 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:10 crc kubenswrapper[4824]: I1006 09:59:10.667333 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:10 crc kubenswrapper[4824]: I1006 09:59:10.667343 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:10Z","lastTransitionTime":"2025-10-06T09:59:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:10 crc kubenswrapper[4824]: I1006 09:59:10.770513 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:10 crc kubenswrapper[4824]: I1006 09:59:10.770600 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:10 crc kubenswrapper[4824]: I1006 09:59:10.770621 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:10 crc kubenswrapper[4824]: I1006 09:59:10.770655 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:10 crc kubenswrapper[4824]: I1006 09:59:10.770677 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:10Z","lastTransitionTime":"2025-10-06T09:59:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:10 crc kubenswrapper[4824]: I1006 09:59:10.873930 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:10 crc kubenswrapper[4824]: I1006 09:59:10.874027 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:10 crc kubenswrapper[4824]: I1006 09:59:10.874048 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:10 crc kubenswrapper[4824]: I1006 09:59:10.874077 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:10 crc kubenswrapper[4824]: I1006 09:59:10.874102 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:10Z","lastTransitionTime":"2025-10-06T09:59:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:10 crc kubenswrapper[4824]: I1006 09:59:10.977692 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:10 crc kubenswrapper[4824]: I1006 09:59:10.977765 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:10 crc kubenswrapper[4824]: I1006 09:59:10.977784 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:10 crc kubenswrapper[4824]: I1006 09:59:10.977812 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:10 crc kubenswrapper[4824]: I1006 09:59:10.977831 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:10Z","lastTransitionTime":"2025-10-06T09:59:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:11 crc kubenswrapper[4824]: I1006 09:59:11.081961 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:11 crc kubenswrapper[4824]: I1006 09:59:11.082460 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:11 crc kubenswrapper[4824]: I1006 09:59:11.082597 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:11 crc kubenswrapper[4824]: I1006 09:59:11.082781 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:11 crc kubenswrapper[4824]: I1006 09:59:11.082951 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:11Z","lastTransitionTime":"2025-10-06T09:59:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:11 crc kubenswrapper[4824]: I1006 09:59:11.187281 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:11 crc kubenswrapper[4824]: I1006 09:59:11.187355 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:11 crc kubenswrapper[4824]: I1006 09:59:11.187378 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:11 crc kubenswrapper[4824]: I1006 09:59:11.187409 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:11 crc kubenswrapper[4824]: I1006 09:59:11.187429 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:11Z","lastTransitionTime":"2025-10-06T09:59:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:11 crc kubenswrapper[4824]: I1006 09:59:11.291212 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:11 crc kubenswrapper[4824]: I1006 09:59:11.291298 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:11 crc kubenswrapper[4824]: I1006 09:59:11.291322 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:11 crc kubenswrapper[4824]: I1006 09:59:11.291358 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:11 crc kubenswrapper[4824]: I1006 09:59:11.291381 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:11Z","lastTransitionTime":"2025-10-06T09:59:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:11 crc kubenswrapper[4824]: I1006 09:59:11.298318 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d739433-0a38-42fe-b884-b8433011967e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://464e279c5903770c34d80f31ed231b194d5628b39055c4058908039d10690059\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://efa1b62728466ef7aac0e745230348ce920eb9ec9b07fb2cd5a276150dac13fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fedc40f0217626c56d155fc7e78e0592556fef9464db2aaf09c5facc486aad8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78fac300cbe2ac52f5788306b8fc0011f6ba3a22c3546974e70998f2d394987a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:11Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:11 crc kubenswrapper[4824]: I1006 09:59:11.320122 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:11Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:11 crc kubenswrapper[4824]: I1006 09:59:11.343553 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-szn8q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b65abf9f-4e88-4571-960d-3ca997d9c344\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://664bf42e1324a9c9529f36f08d61e26999b3386bdd1fac77851fb7e7411e8934\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49s5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-szn8q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:11Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:11 crc kubenswrapper[4824]: I1006 09:59:11.380391 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f511ece-25eb-465f-b85c-9e5f3f886c21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10bfed3e523b7422ce1e981f98166eca78bd12dd4a7cf0634915e1f4ecc3de2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43199f86ad4a1f02bc57a926cd96b2c3c0e2ea46fb36a3f495668706be4c8041\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d6064ad3bcdafebb4c66e7d91f183cc787e9ce13660545c71a722b3ff04ce75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6213164d366243c2139d5971041d6f076cf38fd0cce4c0ca5495111963487ec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa6d59fac987f60a4340ab35d84aa059f2b7085a06362e7207a79702d17eee6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09638fbc9d163ad0b63fa57d24ea5298207ad1f55bab8671be311960be2627a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33caf222d92beb8a42c203fc61d69e804ec3a22d6889f8ead2ce783237d8b464\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33caf222d92beb8a42c203fc61d69e804ec3a22d6889f8ead2ce783237d8b464\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-06T09:59:01Z\\\",\\\"message\\\":\\\"nt-go/informers/factory.go:160\\\\nI1006 09:59:01.318023 6499 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1006 09:59:01.317995 6499 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1006 09:59:01.318223 6499 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1006 09:59:01.318391 6499 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1006 09:59:01.319128 6499 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1006 09:59:01.319521 6499 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1006 09:59:01.319562 6499 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1006 09:59:01.319610 6499 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1006 09:59:01.319617 6499 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1006 09:59:01.319629 6499 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1006 09:59:01.319666 6499 factory.go:656] Stopping watch factory\\\\nI1006 09:59:01.319698 6499 ovnkube.go:599] Stopped ovnkube\\\\nI1006 09:59:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T09:59:00Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-4vwms_openshift-ovn-kubernetes(2f511ece-25eb-465f-b85c-9e5f3f886c21)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0644c7db28c7d26aa8d69cd5b9fbcff4d4c06b96df113137dea22a72d9fecd85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6da4c07be4d507a6f20bd8b42be6e6ea4712a956f57ef164036cfded7142a87e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6da4c07be4d507a6f20bd8b42be6e6ea4712a956f57ef164036cfded7142a87e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4vwms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:11Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:11 crc kubenswrapper[4824]: I1006 09:59:11.393767 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:11 crc kubenswrapper[4824]: I1006 09:59:11.393841 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:11 crc kubenswrapper[4824]: I1006 09:59:11.393853 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:11 crc kubenswrapper[4824]: I1006 09:59:11.393872 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:11 crc kubenswrapper[4824]: I1006 09:59:11.393887 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:11Z","lastTransitionTime":"2025-10-06T09:59:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:11 crc kubenswrapper[4824]: I1006 09:59:11.399711 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lszvg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0cf70ab4-4cfe-41c5-8db7-035451bafcfa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7b80b5b375b3b32132501ee0d661e06e199654c696ac76ad56e7ba96918cc77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fkbf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://519653c6b241cd89ef3ece003667d420a7a55cdf513f4577a69e14ade5fb4c76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fkbf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lszvg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:11Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:11 crc kubenswrapper[4824]: I1006 09:59:11.417553 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gnw94" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7cf1d355-64cb-48a6-acfd-7d258b7afa10\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-psbq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-psbq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:46Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gnw94\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:11Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:11 crc kubenswrapper[4824]: I1006 09:59:11.443086 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b665cda3-da4e-482e-8248-023abec382f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9235c742d8c45bf667f32979c2891fece50f4ee1bd0c7cf6a753ed2d2b9ab30e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b2a920dbc506df2426232990e31fbb60d255c00be54941826a0f094af4a92f0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d516f278f362d7c45e7149de82fa835997ac198f1165c04352910d51fb674d69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c820d8bea16cff0d463ce68d2f2acb72d7a7c73372757ff104e14fcec1025a05\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b8cdf54407cc95783919eba8e92f093bbc09eac5e908c4ba0685629948c71bd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-06T09:58:27Z\\\",\\\"message\\\":\\\"arting RequestHeaderAuthRequestController\\\\nI1006 09:58:27.156944 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1006 09:58:27.156783 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1006 09:58:27.157213 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-432254888/tls.crt::/tmp/serving-cert-432254888/tls.key\\\\\\\"\\\\nI1006 09:58:27.157407 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1006 09:58:27.157443 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1006 09:58:27.157410 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-432254888/tls.crt::/tmp/serving-cert-432254888/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759744705\\\\\\\\\\\\\\\" (2025-10-06 09:58:25 +0000 UTC to 2025-11-05 09:58:26 +0000 UTC (now=2025-10-06 09:58:27.157390076 +0000 UTC))\\\\\\\"\\\\nI1006 09:58:27.157632 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1006 09:58:27.158158 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759744707\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759744707\\\\\\\\\\\\\\\" (2025-10-06 08:58:26 +0000 UTC to 2026-10-06 08:58:26 +0000 UTC (now=2025-10-06 09:58:27.158090663 +0000 UTC))\\\\\\\"\\\\nI1006 09:58:27.158220 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1006 09:58:27.158270 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1006 09:58:27.158310 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nF1006 09:58:27.158617 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://413822c99ba2866521177cce4f1894aef17c160aebb2662da91824a04827d510\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e37e8c7400f7fd14a5c5390a5985e3d8d277e9c5ca60d5e9e9757d07d1778570\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e37e8c7400f7fd14a5c5390a5985e3d8d277e9c5ca60d5e9e9757d07d1778570\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:11Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:11 crc kubenswrapper[4824]: I1006 09:59:11.469213 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-7c22q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7eaaa68-d971-4a1c-b6c5-551ded8ff27f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f5150c8c6cfc65ecdcd0e7bd52199dd60e2472c96874f103b49919c81b5e5f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53336c7d3c509b4e022708998aedc5dee9fdaf7136ae3ee69b4181394c3a22fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53336c7d3c509b4e022708998aedc5dee9fdaf7136ae3ee69b4181394c3a22fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://652ab0a13a2bb421872edda3f7188dce8e296247052550c937f8447fea206756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://652ab0a13a2bb421872edda3f7188dce8e296247052550c937f8447fea206756\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3514dec0386bb8c016a7f555d7191ea420d5058b0ee98f97d965dd857c5376e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3514dec0386bb8c016a7f555d7191ea420d5058b0ee98f97d965dd857c5376e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b035e0bf34e3429400ac8942124a3e9834231899487d03d15609fe5624747a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b035e0bf34e3429400ac8942124a3e9834231899487d03d15609fe5624747a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6743e027a11864ad3f433d110e60451fb36d49791e03b22d437b70f89d056ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6743e027a11864ad3f433d110e60451fb36d49791e03b22d437b70f89d056ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8ea48b6886fd39a1c5e5b29ca1fffc2bd4552d61dc621eb2fba66fb464cf120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8ea48b6886fd39a1c5e5b29ca1fffc2bd4552d61dc621eb2fba66fb464cf120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-7c22q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:11Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:11 crc kubenswrapper[4824]: I1006 09:59:11.491324 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b67b743d-7d7e-4a7f-b648-c721f50bb3a9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:59:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:59:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b5bb30b2f9b1a35bf1dfd2274c7e3d280757902687d9ae79b710fbd5f07f398\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0658bdad936b07337e1eb58a4d62c1d7fc98fb6b3b115d9be4c1d5d79ecbc8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://268983a8eccc2af9242e8f5ebea3722e30cd66086272ec6f681ff96ed9a0e7fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6f4a5101060399a25f1d98b0495745ab0a679b88cc6bee5ea1bf0caff7b6767\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6f4a5101060399a25f1d98b0495745ab0a679b88cc6bee5ea1bf0caff7b6767\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:11Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:11Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:11 crc kubenswrapper[4824]: I1006 09:59:11.496958 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:11 crc kubenswrapper[4824]: I1006 09:59:11.497014 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:11 crc kubenswrapper[4824]: I1006 09:59:11.497027 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:11 crc kubenswrapper[4824]: I1006 09:59:11.497046 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:11 crc kubenswrapper[4824]: I1006 09:59:11.497074 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:11Z","lastTransitionTime":"2025-10-06T09:59:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:11 crc kubenswrapper[4824]: I1006 09:59:11.508277 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xcgwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74317f75-9ebc-4327-b88c-598c2241240f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8def09a126f9bb910affad9b7fb32303b960add9fbfd5bb709d59b4aeba7d245\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v6sg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xcgwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:11Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:11 crc kubenswrapper[4824]: I1006 09:59:11.528533 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1140c8c-93f1-4fce-8c06-a69561ff8a6d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5462cb77b7d61aed7d4b4be2fbc6ba1451d267a66afd6dede2f259c15230e7c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gq49w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5512be36013baeca611ee92645a1d339da237b70d63aac43334beafe24513d44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gq49w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-khgzw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:11Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:11 crc kubenswrapper[4824]: I1006 09:59:11.551899 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2035278cf272827b39a45aa41ecd5b1e41aafc1d9eb8bccd853660fb6e587159\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff3152fceadf36a55578f587623f37d748f90f2afeda5702a05504b372d399be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:11Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:11 crc kubenswrapper[4824]: I1006 09:59:11.579560 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f9011f6673310dffea3acd7bdc2a56f2af91a87641631490fdc63da610f047a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:11Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:11 crc kubenswrapper[4824]: I1006 09:59:11.600443 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:11 crc kubenswrapper[4824]: I1006 09:59:11.600490 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:11 crc kubenswrapper[4824]: I1006 09:59:11.600501 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:11 crc kubenswrapper[4824]: I1006 09:59:11.600525 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:11 crc kubenswrapper[4824]: I1006 09:59:11.600541 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:11Z","lastTransitionTime":"2025-10-06T09:59:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:11 crc kubenswrapper[4824]: I1006 09:59:11.600929 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:11Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:11 crc kubenswrapper[4824]: I1006 09:59:11.627190 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:11Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:11 crc kubenswrapper[4824]: I1006 09:59:11.651180 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab36ed34d6ca02b59043c46e151019620ce31733209d072c21394a79fe36be4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:11Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:11 crc kubenswrapper[4824]: I1006 09:59:11.669331 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5rx6l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0c320fc-94d9-4d82-81ee-ccccfd125efc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f8be7bad4e9644494dfb42380c7f5481e1039a1d27ad379c623515948d87392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njtbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5rx6l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:11Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:11 crc kubenswrapper[4824]: I1006 09:59:11.704074 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:11 crc kubenswrapper[4824]: I1006 09:59:11.704147 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:11 crc kubenswrapper[4824]: I1006 09:59:11.704167 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:11 crc kubenswrapper[4824]: I1006 09:59:11.704194 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:11 crc kubenswrapper[4824]: I1006 09:59:11.704215 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:11Z","lastTransitionTime":"2025-10-06T09:59:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:11 crc kubenswrapper[4824]: I1006 09:59:11.807052 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:11 crc kubenswrapper[4824]: I1006 09:59:11.807134 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:11 crc kubenswrapper[4824]: I1006 09:59:11.807148 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:11 crc kubenswrapper[4824]: I1006 09:59:11.807169 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:11 crc kubenswrapper[4824]: I1006 09:59:11.807207 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:11Z","lastTransitionTime":"2025-10-06T09:59:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:11 crc kubenswrapper[4824]: I1006 09:59:11.910968 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:11 crc kubenswrapper[4824]: I1006 09:59:11.911059 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:11 crc kubenswrapper[4824]: I1006 09:59:11.911076 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:11 crc kubenswrapper[4824]: I1006 09:59:11.911102 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:11 crc kubenswrapper[4824]: I1006 09:59:11.911121 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:11Z","lastTransitionTime":"2025-10-06T09:59:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:12 crc kubenswrapper[4824]: I1006 09:59:12.014911 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:12 crc kubenswrapper[4824]: I1006 09:59:12.014973 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:12 crc kubenswrapper[4824]: I1006 09:59:12.015018 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:12 crc kubenswrapper[4824]: I1006 09:59:12.015044 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:12 crc kubenswrapper[4824]: I1006 09:59:12.015065 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:12Z","lastTransitionTime":"2025-10-06T09:59:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:12 crc kubenswrapper[4824]: I1006 09:59:12.118852 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:12 crc kubenswrapper[4824]: I1006 09:59:12.118919 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:12 crc kubenswrapper[4824]: I1006 09:59:12.118937 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:12 crc kubenswrapper[4824]: I1006 09:59:12.119009 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:12 crc kubenswrapper[4824]: I1006 09:59:12.119028 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:12Z","lastTransitionTime":"2025-10-06T09:59:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:12 crc kubenswrapper[4824]: I1006 09:59:12.222690 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:12 crc kubenswrapper[4824]: I1006 09:59:12.222795 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:12 crc kubenswrapper[4824]: I1006 09:59:12.222814 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:12 crc kubenswrapper[4824]: I1006 09:59:12.222842 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:12 crc kubenswrapper[4824]: I1006 09:59:12.222862 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:12Z","lastTransitionTime":"2025-10-06T09:59:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:12 crc kubenswrapper[4824]: I1006 09:59:12.274056 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 09:59:12 crc kubenswrapper[4824]: I1006 09:59:12.274179 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gnw94" Oct 06 09:59:12 crc kubenswrapper[4824]: I1006 09:59:12.274248 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 09:59:12 crc kubenswrapper[4824]: I1006 09:59:12.274296 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 09:59:12 crc kubenswrapper[4824]: E1006 09:59:12.274360 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 06 09:59:12 crc kubenswrapper[4824]: E1006 09:59:12.274461 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 06 09:59:12 crc kubenswrapper[4824]: E1006 09:59:12.274663 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gnw94" podUID="7cf1d355-64cb-48a6-acfd-7d258b7afa10" Oct 06 09:59:12 crc kubenswrapper[4824]: E1006 09:59:12.274812 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 06 09:59:12 crc kubenswrapper[4824]: I1006 09:59:12.325761 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:12 crc kubenswrapper[4824]: I1006 09:59:12.325822 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:12 crc kubenswrapper[4824]: I1006 09:59:12.325839 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:12 crc kubenswrapper[4824]: I1006 09:59:12.325865 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:12 crc kubenswrapper[4824]: I1006 09:59:12.325884 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:12Z","lastTransitionTime":"2025-10-06T09:59:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:12 crc kubenswrapper[4824]: I1006 09:59:12.430407 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:12 crc kubenswrapper[4824]: I1006 09:59:12.430483 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:12 crc kubenswrapper[4824]: I1006 09:59:12.430500 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:12 crc kubenswrapper[4824]: I1006 09:59:12.430526 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:12 crc kubenswrapper[4824]: I1006 09:59:12.430551 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:12Z","lastTransitionTime":"2025-10-06T09:59:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:12 crc kubenswrapper[4824]: I1006 09:59:12.533851 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:12 crc kubenswrapper[4824]: I1006 09:59:12.533901 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:12 crc kubenswrapper[4824]: I1006 09:59:12.533911 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:12 crc kubenswrapper[4824]: I1006 09:59:12.533933 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:12 crc kubenswrapper[4824]: I1006 09:59:12.533944 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:12Z","lastTransitionTime":"2025-10-06T09:59:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:12 crc kubenswrapper[4824]: I1006 09:59:12.637924 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:12 crc kubenswrapper[4824]: I1006 09:59:12.638035 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:12 crc kubenswrapper[4824]: I1006 09:59:12.638051 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:12 crc kubenswrapper[4824]: I1006 09:59:12.638076 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:12 crc kubenswrapper[4824]: I1006 09:59:12.638100 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:12Z","lastTransitionTime":"2025-10-06T09:59:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:12 crc kubenswrapper[4824]: I1006 09:59:12.748625 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:12 crc kubenswrapper[4824]: I1006 09:59:12.748700 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:12 crc kubenswrapper[4824]: I1006 09:59:12.748718 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:12 crc kubenswrapper[4824]: I1006 09:59:12.748743 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:12 crc kubenswrapper[4824]: I1006 09:59:12.748760 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:12Z","lastTransitionTime":"2025-10-06T09:59:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:12 crc kubenswrapper[4824]: I1006 09:59:12.851830 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:12 crc kubenswrapper[4824]: I1006 09:59:12.852237 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:12 crc kubenswrapper[4824]: I1006 09:59:12.852410 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:12 crc kubenswrapper[4824]: I1006 09:59:12.852547 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:12 crc kubenswrapper[4824]: I1006 09:59:12.852665 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:12Z","lastTransitionTime":"2025-10-06T09:59:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:12 crc kubenswrapper[4824]: I1006 09:59:12.956220 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:12 crc kubenswrapper[4824]: I1006 09:59:12.956258 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:12 crc kubenswrapper[4824]: I1006 09:59:12.956269 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:12 crc kubenswrapper[4824]: I1006 09:59:12.956285 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:12 crc kubenswrapper[4824]: I1006 09:59:12.956297 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:12Z","lastTransitionTime":"2025-10-06T09:59:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:13 crc kubenswrapper[4824]: I1006 09:59:13.059328 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:13 crc kubenswrapper[4824]: I1006 09:59:13.059917 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:13 crc kubenswrapper[4824]: I1006 09:59:13.059930 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:13 crc kubenswrapper[4824]: I1006 09:59:13.059953 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:13 crc kubenswrapper[4824]: I1006 09:59:13.059967 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:13Z","lastTransitionTime":"2025-10-06T09:59:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:13 crc kubenswrapper[4824]: I1006 09:59:13.163671 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:13 crc kubenswrapper[4824]: I1006 09:59:13.163731 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:13 crc kubenswrapper[4824]: I1006 09:59:13.163744 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:13 crc kubenswrapper[4824]: I1006 09:59:13.163762 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:13 crc kubenswrapper[4824]: I1006 09:59:13.163774 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:13Z","lastTransitionTime":"2025-10-06T09:59:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:13 crc kubenswrapper[4824]: I1006 09:59:13.267077 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:13 crc kubenswrapper[4824]: I1006 09:59:13.267155 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:13 crc kubenswrapper[4824]: I1006 09:59:13.267173 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:13 crc kubenswrapper[4824]: I1006 09:59:13.267202 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:13 crc kubenswrapper[4824]: I1006 09:59:13.267232 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:13Z","lastTransitionTime":"2025-10-06T09:59:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:13 crc kubenswrapper[4824]: I1006 09:59:13.370605 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:13 crc kubenswrapper[4824]: I1006 09:59:13.370663 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:13 crc kubenswrapper[4824]: I1006 09:59:13.370684 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:13 crc kubenswrapper[4824]: I1006 09:59:13.370711 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:13 crc kubenswrapper[4824]: I1006 09:59:13.370729 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:13Z","lastTransitionTime":"2025-10-06T09:59:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:13 crc kubenswrapper[4824]: I1006 09:59:13.473079 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:13 crc kubenswrapper[4824]: I1006 09:59:13.473134 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:13 crc kubenswrapper[4824]: I1006 09:59:13.473144 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:13 crc kubenswrapper[4824]: I1006 09:59:13.473162 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:13 crc kubenswrapper[4824]: I1006 09:59:13.473175 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:13Z","lastTransitionTime":"2025-10-06T09:59:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:13 crc kubenswrapper[4824]: I1006 09:59:13.576620 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:13 crc kubenswrapper[4824]: I1006 09:59:13.576669 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:13 crc kubenswrapper[4824]: I1006 09:59:13.576679 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:13 crc kubenswrapper[4824]: I1006 09:59:13.576694 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:13 crc kubenswrapper[4824]: I1006 09:59:13.576705 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:13Z","lastTransitionTime":"2025-10-06T09:59:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:13 crc kubenswrapper[4824]: I1006 09:59:13.679817 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:13 crc kubenswrapper[4824]: I1006 09:59:13.679863 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:13 crc kubenswrapper[4824]: I1006 09:59:13.679872 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:13 crc kubenswrapper[4824]: I1006 09:59:13.679890 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:13 crc kubenswrapper[4824]: I1006 09:59:13.679901 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:13Z","lastTransitionTime":"2025-10-06T09:59:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:13 crc kubenswrapper[4824]: I1006 09:59:13.783763 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:13 crc kubenswrapper[4824]: I1006 09:59:13.783814 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:13 crc kubenswrapper[4824]: I1006 09:59:13.783831 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:13 crc kubenswrapper[4824]: I1006 09:59:13.783856 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:13 crc kubenswrapper[4824]: I1006 09:59:13.783873 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:13Z","lastTransitionTime":"2025-10-06T09:59:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:13 crc kubenswrapper[4824]: I1006 09:59:13.887920 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:13 crc kubenswrapper[4824]: I1006 09:59:13.888035 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:13 crc kubenswrapper[4824]: I1006 09:59:13.888066 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:13 crc kubenswrapper[4824]: I1006 09:59:13.888093 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:13 crc kubenswrapper[4824]: I1006 09:59:13.888111 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:13Z","lastTransitionTime":"2025-10-06T09:59:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:13 crc kubenswrapper[4824]: I1006 09:59:13.991817 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:13 crc kubenswrapper[4824]: I1006 09:59:13.991885 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:13 crc kubenswrapper[4824]: I1006 09:59:13.991909 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:13 crc kubenswrapper[4824]: I1006 09:59:13.991935 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:13 crc kubenswrapper[4824]: I1006 09:59:13.991956 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:13Z","lastTransitionTime":"2025-10-06T09:59:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:14 crc kubenswrapper[4824]: I1006 09:59:14.095667 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:14 crc kubenswrapper[4824]: I1006 09:59:14.095735 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:14 crc kubenswrapper[4824]: I1006 09:59:14.095754 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:14 crc kubenswrapper[4824]: I1006 09:59:14.095781 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:14 crc kubenswrapper[4824]: I1006 09:59:14.095803 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:14Z","lastTransitionTime":"2025-10-06T09:59:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:14 crc kubenswrapper[4824]: I1006 09:59:14.198853 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:14 crc kubenswrapper[4824]: I1006 09:59:14.198955 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:14 crc kubenswrapper[4824]: I1006 09:59:14.198973 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:14 crc kubenswrapper[4824]: I1006 09:59:14.199029 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:14 crc kubenswrapper[4824]: I1006 09:59:14.199048 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:14Z","lastTransitionTime":"2025-10-06T09:59:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:14 crc kubenswrapper[4824]: I1006 09:59:14.273601 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 09:59:14 crc kubenswrapper[4824]: I1006 09:59:14.273655 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 09:59:14 crc kubenswrapper[4824]: I1006 09:59:14.273711 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 09:59:14 crc kubenswrapper[4824]: E1006 09:59:14.273734 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 06 09:59:14 crc kubenswrapper[4824]: E1006 09:59:14.273860 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 06 09:59:14 crc kubenswrapper[4824]: E1006 09:59:14.274016 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 06 09:59:14 crc kubenswrapper[4824]: I1006 09:59:14.274318 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gnw94" Oct 06 09:59:14 crc kubenswrapper[4824]: E1006 09:59:14.274444 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gnw94" podUID="7cf1d355-64cb-48a6-acfd-7d258b7afa10" Oct 06 09:59:14 crc kubenswrapper[4824]: I1006 09:59:14.274606 4824 scope.go:117] "RemoveContainer" containerID="33caf222d92beb8a42c203fc61d69e804ec3a22d6889f8ead2ce783237d8b464" Oct 06 09:59:14 crc kubenswrapper[4824]: E1006 09:59:14.274750 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-4vwms_openshift-ovn-kubernetes(2f511ece-25eb-465f-b85c-9e5f3f886c21)\"" pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" podUID="2f511ece-25eb-465f-b85c-9e5f3f886c21" Oct 06 09:59:14 crc kubenswrapper[4824]: I1006 09:59:14.302046 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:14 crc kubenswrapper[4824]: I1006 09:59:14.302103 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:14 crc kubenswrapper[4824]: I1006 09:59:14.302121 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:14 crc kubenswrapper[4824]: I1006 09:59:14.302146 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:14 crc kubenswrapper[4824]: I1006 09:59:14.302169 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:14Z","lastTransitionTime":"2025-10-06T09:59:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:14 crc kubenswrapper[4824]: I1006 09:59:14.405138 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:14 crc kubenswrapper[4824]: I1006 09:59:14.405203 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:14 crc kubenswrapper[4824]: I1006 09:59:14.405265 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:14 crc kubenswrapper[4824]: I1006 09:59:14.405295 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:14 crc kubenswrapper[4824]: I1006 09:59:14.405313 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:14Z","lastTransitionTime":"2025-10-06T09:59:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:14 crc kubenswrapper[4824]: I1006 09:59:14.507700 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:14 crc kubenswrapper[4824]: I1006 09:59:14.507765 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:14 crc kubenswrapper[4824]: I1006 09:59:14.507782 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:14 crc kubenswrapper[4824]: I1006 09:59:14.507807 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:14 crc kubenswrapper[4824]: I1006 09:59:14.507826 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:14Z","lastTransitionTime":"2025-10-06T09:59:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:14 crc kubenswrapper[4824]: I1006 09:59:14.612427 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:14 crc kubenswrapper[4824]: I1006 09:59:14.612496 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:14 crc kubenswrapper[4824]: I1006 09:59:14.612514 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:14 crc kubenswrapper[4824]: I1006 09:59:14.612544 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:14 crc kubenswrapper[4824]: I1006 09:59:14.612571 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:14Z","lastTransitionTime":"2025-10-06T09:59:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:14 crc kubenswrapper[4824]: I1006 09:59:14.716189 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:14 crc kubenswrapper[4824]: I1006 09:59:14.716238 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:14 crc kubenswrapper[4824]: I1006 09:59:14.716254 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:14 crc kubenswrapper[4824]: I1006 09:59:14.716280 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:14 crc kubenswrapper[4824]: I1006 09:59:14.716297 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:14Z","lastTransitionTime":"2025-10-06T09:59:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:14 crc kubenswrapper[4824]: I1006 09:59:14.819251 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:14 crc kubenswrapper[4824]: I1006 09:59:14.819323 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:14 crc kubenswrapper[4824]: I1006 09:59:14.819342 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:14 crc kubenswrapper[4824]: I1006 09:59:14.819368 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:14 crc kubenswrapper[4824]: I1006 09:59:14.819386 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:14Z","lastTransitionTime":"2025-10-06T09:59:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:14 crc kubenswrapper[4824]: I1006 09:59:14.923337 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:14 crc kubenswrapper[4824]: I1006 09:59:14.923399 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:14 crc kubenswrapper[4824]: I1006 09:59:14.923412 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:14 crc kubenswrapper[4824]: I1006 09:59:14.923433 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:14 crc kubenswrapper[4824]: I1006 09:59:14.923446 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:14Z","lastTransitionTime":"2025-10-06T09:59:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:15 crc kubenswrapper[4824]: I1006 09:59:15.027592 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:15 crc kubenswrapper[4824]: I1006 09:59:15.028144 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:15 crc kubenswrapper[4824]: I1006 09:59:15.028352 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:15 crc kubenswrapper[4824]: I1006 09:59:15.028563 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:15 crc kubenswrapper[4824]: I1006 09:59:15.028729 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:15Z","lastTransitionTime":"2025-10-06T09:59:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:15 crc kubenswrapper[4824]: I1006 09:59:15.131566 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:15 crc kubenswrapper[4824]: I1006 09:59:15.131612 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:15 crc kubenswrapper[4824]: I1006 09:59:15.131624 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:15 crc kubenswrapper[4824]: I1006 09:59:15.131640 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:15 crc kubenswrapper[4824]: I1006 09:59:15.131652 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:15Z","lastTransitionTime":"2025-10-06T09:59:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:15 crc kubenswrapper[4824]: I1006 09:59:15.235151 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:15 crc kubenswrapper[4824]: I1006 09:59:15.235227 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:15 crc kubenswrapper[4824]: I1006 09:59:15.235245 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:15 crc kubenswrapper[4824]: I1006 09:59:15.235272 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:15 crc kubenswrapper[4824]: I1006 09:59:15.235289 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:15Z","lastTransitionTime":"2025-10-06T09:59:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:15 crc kubenswrapper[4824]: I1006 09:59:15.339341 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:15 crc kubenswrapper[4824]: I1006 09:59:15.339740 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:15 crc kubenswrapper[4824]: I1006 09:59:15.340030 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:15 crc kubenswrapper[4824]: I1006 09:59:15.340295 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:15 crc kubenswrapper[4824]: I1006 09:59:15.340669 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:15Z","lastTransitionTime":"2025-10-06T09:59:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:15 crc kubenswrapper[4824]: I1006 09:59:15.444470 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:15 crc kubenswrapper[4824]: I1006 09:59:15.444883 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:15 crc kubenswrapper[4824]: I1006 09:59:15.445127 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:15 crc kubenswrapper[4824]: I1006 09:59:15.445392 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:15 crc kubenswrapper[4824]: I1006 09:59:15.445633 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:15Z","lastTransitionTime":"2025-10-06T09:59:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:15 crc kubenswrapper[4824]: I1006 09:59:15.548783 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:15 crc kubenswrapper[4824]: I1006 09:59:15.548841 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:15 crc kubenswrapper[4824]: I1006 09:59:15.548858 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:15 crc kubenswrapper[4824]: I1006 09:59:15.548880 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:15 crc kubenswrapper[4824]: I1006 09:59:15.548895 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:15Z","lastTransitionTime":"2025-10-06T09:59:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:15 crc kubenswrapper[4824]: I1006 09:59:15.651903 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:15 crc kubenswrapper[4824]: I1006 09:59:15.651973 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:15 crc kubenswrapper[4824]: I1006 09:59:15.652029 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:15 crc kubenswrapper[4824]: I1006 09:59:15.652058 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:15 crc kubenswrapper[4824]: I1006 09:59:15.652079 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:15Z","lastTransitionTime":"2025-10-06T09:59:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:15 crc kubenswrapper[4824]: I1006 09:59:15.755125 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:15 crc kubenswrapper[4824]: I1006 09:59:15.756035 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:15 crc kubenswrapper[4824]: I1006 09:59:15.756216 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:15 crc kubenswrapper[4824]: I1006 09:59:15.756375 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:15 crc kubenswrapper[4824]: I1006 09:59:15.756549 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:15Z","lastTransitionTime":"2025-10-06T09:59:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:15 crc kubenswrapper[4824]: I1006 09:59:15.859334 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:15 crc kubenswrapper[4824]: I1006 09:59:15.859672 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:15 crc kubenswrapper[4824]: I1006 09:59:15.859757 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:15 crc kubenswrapper[4824]: I1006 09:59:15.859841 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:15 crc kubenswrapper[4824]: I1006 09:59:15.859924 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:15Z","lastTransitionTime":"2025-10-06T09:59:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:15 crc kubenswrapper[4824]: I1006 09:59:15.962487 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:15 crc kubenswrapper[4824]: I1006 09:59:15.962548 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:15 crc kubenswrapper[4824]: I1006 09:59:15.962565 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:15 crc kubenswrapper[4824]: I1006 09:59:15.962589 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:15 crc kubenswrapper[4824]: I1006 09:59:15.962606 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:15Z","lastTransitionTime":"2025-10-06T09:59:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:16 crc kubenswrapper[4824]: I1006 09:59:16.065285 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:16 crc kubenswrapper[4824]: I1006 09:59:16.065347 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:16 crc kubenswrapper[4824]: I1006 09:59:16.065364 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:16 crc kubenswrapper[4824]: I1006 09:59:16.065390 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:16 crc kubenswrapper[4824]: I1006 09:59:16.065407 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:16Z","lastTransitionTime":"2025-10-06T09:59:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:16 crc kubenswrapper[4824]: I1006 09:59:16.168094 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:16 crc kubenswrapper[4824]: I1006 09:59:16.168597 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:16 crc kubenswrapper[4824]: I1006 09:59:16.168784 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:16 crc kubenswrapper[4824]: I1006 09:59:16.169213 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:16 crc kubenswrapper[4824]: I1006 09:59:16.169579 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:16Z","lastTransitionTime":"2025-10-06T09:59:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:16 crc kubenswrapper[4824]: I1006 09:59:16.272879 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:16 crc kubenswrapper[4824]: I1006 09:59:16.273201 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:16 crc kubenswrapper[4824]: I1006 09:59:16.273330 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 09:59:16 crc kubenswrapper[4824]: I1006 09:59:16.273277 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 09:59:16 crc kubenswrapper[4824]: E1006 09:59:16.273534 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 06 09:59:16 crc kubenswrapper[4824]: I1006 09:59:16.273334 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:16 crc kubenswrapper[4824]: I1006 09:59:16.273620 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:16 crc kubenswrapper[4824]: I1006 09:59:16.273248 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gnw94" Oct 06 09:59:16 crc kubenswrapper[4824]: E1006 09:59:16.273703 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 06 09:59:16 crc kubenswrapper[4824]: I1006 09:59:16.273285 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 09:59:16 crc kubenswrapper[4824]: E1006 09:59:16.274065 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gnw94" podUID="7cf1d355-64cb-48a6-acfd-7d258b7afa10" Oct 06 09:59:16 crc kubenswrapper[4824]: E1006 09:59:16.274176 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 06 09:59:16 crc kubenswrapper[4824]: I1006 09:59:16.273650 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:16Z","lastTransitionTime":"2025-10-06T09:59:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:16 crc kubenswrapper[4824]: I1006 09:59:16.378486 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:16 crc kubenswrapper[4824]: I1006 09:59:16.378571 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:16 crc kubenswrapper[4824]: I1006 09:59:16.378596 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:16 crc kubenswrapper[4824]: I1006 09:59:16.378622 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:16 crc kubenswrapper[4824]: I1006 09:59:16.378640 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:16Z","lastTransitionTime":"2025-10-06T09:59:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:16 crc kubenswrapper[4824]: I1006 09:59:16.481814 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:16 crc kubenswrapper[4824]: I1006 09:59:16.482288 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:16 crc kubenswrapper[4824]: I1006 09:59:16.482440 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:16 crc kubenswrapper[4824]: I1006 09:59:16.482611 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:16 crc kubenswrapper[4824]: I1006 09:59:16.482751 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:16Z","lastTransitionTime":"2025-10-06T09:59:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:16 crc kubenswrapper[4824]: I1006 09:59:16.586874 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:16 crc kubenswrapper[4824]: I1006 09:59:16.587027 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:16 crc kubenswrapper[4824]: I1006 09:59:16.587061 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:16 crc kubenswrapper[4824]: I1006 09:59:16.587106 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:16 crc kubenswrapper[4824]: I1006 09:59:16.587132 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:16Z","lastTransitionTime":"2025-10-06T09:59:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:16 crc kubenswrapper[4824]: I1006 09:59:16.690788 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:16 crc kubenswrapper[4824]: I1006 09:59:16.690858 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:16 crc kubenswrapper[4824]: I1006 09:59:16.690877 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:16 crc kubenswrapper[4824]: I1006 09:59:16.690905 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:16 crc kubenswrapper[4824]: I1006 09:59:16.690933 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:16Z","lastTransitionTime":"2025-10-06T09:59:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:16 crc kubenswrapper[4824]: I1006 09:59:16.794041 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:16 crc kubenswrapper[4824]: I1006 09:59:16.794131 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:16 crc kubenswrapper[4824]: I1006 09:59:16.794149 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:16 crc kubenswrapper[4824]: I1006 09:59:16.794661 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:16 crc kubenswrapper[4824]: I1006 09:59:16.794729 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:16Z","lastTransitionTime":"2025-10-06T09:59:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:16 crc kubenswrapper[4824]: I1006 09:59:16.898798 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:16 crc kubenswrapper[4824]: I1006 09:59:16.898879 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:16 crc kubenswrapper[4824]: I1006 09:59:16.898898 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:16 crc kubenswrapper[4824]: I1006 09:59:16.898926 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:16 crc kubenswrapper[4824]: I1006 09:59:16.898947 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:16Z","lastTransitionTime":"2025-10-06T09:59:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:17 crc kubenswrapper[4824]: I1006 09:59:17.001524 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:17 crc kubenswrapper[4824]: I1006 09:59:17.001884 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:17 crc kubenswrapper[4824]: I1006 09:59:17.001971 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:17 crc kubenswrapper[4824]: I1006 09:59:17.002099 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:17 crc kubenswrapper[4824]: I1006 09:59:17.002206 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:17Z","lastTransitionTime":"2025-10-06T09:59:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:17 crc kubenswrapper[4824]: I1006 09:59:17.105524 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:17 crc kubenswrapper[4824]: I1006 09:59:17.105582 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:17 crc kubenswrapper[4824]: I1006 09:59:17.105600 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:17 crc kubenswrapper[4824]: I1006 09:59:17.105626 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:17 crc kubenswrapper[4824]: I1006 09:59:17.105644 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:17Z","lastTransitionTime":"2025-10-06T09:59:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:17 crc kubenswrapper[4824]: I1006 09:59:17.208413 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:17 crc kubenswrapper[4824]: I1006 09:59:17.208492 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:17 crc kubenswrapper[4824]: I1006 09:59:17.208504 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:17 crc kubenswrapper[4824]: I1006 09:59:17.208527 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:17 crc kubenswrapper[4824]: I1006 09:59:17.208540 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:17Z","lastTransitionTime":"2025-10-06T09:59:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:17 crc kubenswrapper[4824]: I1006 09:59:17.311020 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:17 crc kubenswrapper[4824]: I1006 09:59:17.311337 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:17 crc kubenswrapper[4824]: I1006 09:59:17.311409 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:17 crc kubenswrapper[4824]: I1006 09:59:17.311490 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:17 crc kubenswrapper[4824]: I1006 09:59:17.311546 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:17Z","lastTransitionTime":"2025-10-06T09:59:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:17 crc kubenswrapper[4824]: I1006 09:59:17.414787 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:17 crc kubenswrapper[4824]: I1006 09:59:17.414850 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:17 crc kubenswrapper[4824]: I1006 09:59:17.414872 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:17 crc kubenswrapper[4824]: I1006 09:59:17.414897 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:17 crc kubenswrapper[4824]: I1006 09:59:17.414914 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:17Z","lastTransitionTime":"2025-10-06T09:59:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:17 crc kubenswrapper[4824]: I1006 09:59:17.519138 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:17 crc kubenswrapper[4824]: I1006 09:59:17.519197 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:17 crc kubenswrapper[4824]: I1006 09:59:17.519213 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:17 crc kubenswrapper[4824]: I1006 09:59:17.519237 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:17 crc kubenswrapper[4824]: I1006 09:59:17.519254 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:17Z","lastTransitionTime":"2025-10-06T09:59:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:17 crc kubenswrapper[4824]: I1006 09:59:17.622876 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:17 crc kubenswrapper[4824]: I1006 09:59:17.622936 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:17 crc kubenswrapper[4824]: I1006 09:59:17.622954 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:17 crc kubenswrapper[4824]: I1006 09:59:17.623021 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:17 crc kubenswrapper[4824]: I1006 09:59:17.623052 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:17Z","lastTransitionTime":"2025-10-06T09:59:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:17 crc kubenswrapper[4824]: I1006 09:59:17.726889 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:17 crc kubenswrapper[4824]: I1006 09:59:17.726933 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:17 crc kubenswrapper[4824]: I1006 09:59:17.726945 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:17 crc kubenswrapper[4824]: I1006 09:59:17.726966 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:17 crc kubenswrapper[4824]: I1006 09:59:17.727000 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:17Z","lastTransitionTime":"2025-10-06T09:59:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:17 crc kubenswrapper[4824]: I1006 09:59:17.788830 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:17 crc kubenswrapper[4824]: I1006 09:59:17.790223 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:17 crc kubenswrapper[4824]: I1006 09:59:17.790390 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:17 crc kubenswrapper[4824]: I1006 09:59:17.790584 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:17 crc kubenswrapper[4824]: I1006 09:59:17.790714 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:17Z","lastTransitionTime":"2025-10-06T09:59:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:17 crc kubenswrapper[4824]: E1006 09:59:17.812472 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T09:59:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T09:59:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T09:59:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T09:59:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T09:59:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T09:59:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T09:59:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T09:59:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"740c2e7e-c94c-411f-b080-eb42c9aecdca\\\",\\\"systemUUID\\\":\\\"77448683-227c-4c6b-a334-eb212350758a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:17Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:17 crc kubenswrapper[4824]: I1006 09:59:17.820169 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:17 crc kubenswrapper[4824]: I1006 09:59:17.820212 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:17 crc kubenswrapper[4824]: I1006 09:59:17.820231 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:17 crc kubenswrapper[4824]: I1006 09:59:17.820258 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:17 crc kubenswrapper[4824]: I1006 09:59:17.820277 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:17Z","lastTransitionTime":"2025-10-06T09:59:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:17 crc kubenswrapper[4824]: E1006 09:59:17.837949 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T09:59:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T09:59:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T09:59:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T09:59:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T09:59:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T09:59:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T09:59:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T09:59:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"740c2e7e-c94c-411f-b080-eb42c9aecdca\\\",\\\"systemUUID\\\":\\\"77448683-227c-4c6b-a334-eb212350758a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:17Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:17 crc kubenswrapper[4824]: I1006 09:59:17.843229 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:17 crc kubenswrapper[4824]: I1006 09:59:17.843302 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:17 crc kubenswrapper[4824]: I1006 09:59:17.843322 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:17 crc kubenswrapper[4824]: I1006 09:59:17.843351 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:17 crc kubenswrapper[4824]: I1006 09:59:17.843374 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:17Z","lastTransitionTime":"2025-10-06T09:59:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:17 crc kubenswrapper[4824]: E1006 09:59:17.856603 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T09:59:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T09:59:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T09:59:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T09:59:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T09:59:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T09:59:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T09:59:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T09:59:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"740c2e7e-c94c-411f-b080-eb42c9aecdca\\\",\\\"systemUUID\\\":\\\"77448683-227c-4c6b-a334-eb212350758a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:17Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:17 crc kubenswrapper[4824]: I1006 09:59:17.861062 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:17 crc kubenswrapper[4824]: I1006 09:59:17.861161 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:17 crc kubenswrapper[4824]: I1006 09:59:17.861178 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:17 crc kubenswrapper[4824]: I1006 09:59:17.861199 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:17 crc kubenswrapper[4824]: I1006 09:59:17.861214 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:17Z","lastTransitionTime":"2025-10-06T09:59:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:17 crc kubenswrapper[4824]: E1006 09:59:17.876132 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T09:59:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T09:59:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T09:59:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T09:59:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T09:59:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T09:59:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T09:59:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T09:59:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"740c2e7e-c94c-411f-b080-eb42c9aecdca\\\",\\\"systemUUID\\\":\\\"77448683-227c-4c6b-a334-eb212350758a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:17Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:17 crc kubenswrapper[4824]: I1006 09:59:17.880944 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:17 crc kubenswrapper[4824]: I1006 09:59:17.881020 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:17 crc kubenswrapper[4824]: I1006 09:59:17.881033 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:17 crc kubenswrapper[4824]: I1006 09:59:17.881050 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:17 crc kubenswrapper[4824]: I1006 09:59:17.881104 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:17Z","lastTransitionTime":"2025-10-06T09:59:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:17 crc kubenswrapper[4824]: E1006 09:59:17.895142 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T09:59:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T09:59:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T09:59:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T09:59:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T09:59:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T09:59:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T09:59:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T09:59:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"740c2e7e-c94c-411f-b080-eb42c9aecdca\\\",\\\"systemUUID\\\":\\\"77448683-227c-4c6b-a334-eb212350758a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:17Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:17 crc kubenswrapper[4824]: E1006 09:59:17.895389 4824 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 06 09:59:17 crc kubenswrapper[4824]: I1006 09:59:17.897172 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:17 crc kubenswrapper[4824]: I1006 09:59:17.897215 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:17 crc kubenswrapper[4824]: I1006 09:59:17.897232 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:17 crc kubenswrapper[4824]: I1006 09:59:17.897258 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:17 crc kubenswrapper[4824]: I1006 09:59:17.897279 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:17Z","lastTransitionTime":"2025-10-06T09:59:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:18 crc kubenswrapper[4824]: I1006 09:59:18.000819 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:18 crc kubenswrapper[4824]: I1006 09:59:18.000898 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:18 crc kubenswrapper[4824]: I1006 09:59:18.000922 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:18 crc kubenswrapper[4824]: I1006 09:59:18.000953 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:18 crc kubenswrapper[4824]: I1006 09:59:18.001009 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:18Z","lastTransitionTime":"2025-10-06T09:59:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:18 crc kubenswrapper[4824]: I1006 09:59:18.103962 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:18 crc kubenswrapper[4824]: I1006 09:59:18.104040 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:18 crc kubenswrapper[4824]: I1006 09:59:18.104052 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:18 crc kubenswrapper[4824]: I1006 09:59:18.104066 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:18 crc kubenswrapper[4824]: I1006 09:59:18.104077 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:18Z","lastTransitionTime":"2025-10-06T09:59:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:18 crc kubenswrapper[4824]: I1006 09:59:18.109565 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7cf1d355-64cb-48a6-acfd-7d258b7afa10-metrics-certs\") pod \"network-metrics-daemon-gnw94\" (UID: \"7cf1d355-64cb-48a6-acfd-7d258b7afa10\") " pod="openshift-multus/network-metrics-daemon-gnw94" Oct 06 09:59:18 crc kubenswrapper[4824]: E1006 09:59:18.109780 4824 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 06 09:59:18 crc kubenswrapper[4824]: E1006 09:59:18.109910 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7cf1d355-64cb-48a6-acfd-7d258b7afa10-metrics-certs podName:7cf1d355-64cb-48a6-acfd-7d258b7afa10 nodeName:}" failed. No retries permitted until 2025-10-06 09:59:50.109876397 +0000 UTC m=+99.474299288 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/7cf1d355-64cb-48a6-acfd-7d258b7afa10-metrics-certs") pod "network-metrics-daemon-gnw94" (UID: "7cf1d355-64cb-48a6-acfd-7d258b7afa10") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 06 09:59:18 crc kubenswrapper[4824]: I1006 09:59:18.206640 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:18 crc kubenswrapper[4824]: I1006 09:59:18.206743 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:18 crc kubenswrapper[4824]: I1006 09:59:18.206764 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:18 crc kubenswrapper[4824]: I1006 09:59:18.206785 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:18 crc kubenswrapper[4824]: I1006 09:59:18.206804 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:18Z","lastTransitionTime":"2025-10-06T09:59:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:18 crc kubenswrapper[4824]: I1006 09:59:18.274053 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 09:59:18 crc kubenswrapper[4824]: I1006 09:59:18.274079 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 09:59:18 crc kubenswrapper[4824]: I1006 09:59:18.274148 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 09:59:18 crc kubenswrapper[4824]: I1006 09:59:18.274181 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gnw94" Oct 06 09:59:18 crc kubenswrapper[4824]: E1006 09:59:18.274250 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 06 09:59:18 crc kubenswrapper[4824]: E1006 09:59:18.274401 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gnw94" podUID="7cf1d355-64cb-48a6-acfd-7d258b7afa10" Oct 06 09:59:18 crc kubenswrapper[4824]: E1006 09:59:18.274550 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 06 09:59:18 crc kubenswrapper[4824]: E1006 09:59:18.274645 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 06 09:59:18 crc kubenswrapper[4824]: I1006 09:59:18.310063 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:18 crc kubenswrapper[4824]: I1006 09:59:18.310094 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:18 crc kubenswrapper[4824]: I1006 09:59:18.310102 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:18 crc kubenswrapper[4824]: I1006 09:59:18.310117 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:18 crc kubenswrapper[4824]: I1006 09:59:18.310129 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:18Z","lastTransitionTime":"2025-10-06T09:59:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:18 crc kubenswrapper[4824]: I1006 09:59:18.413139 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:18 crc kubenswrapper[4824]: I1006 09:59:18.413189 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:18 crc kubenswrapper[4824]: I1006 09:59:18.413206 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:18 crc kubenswrapper[4824]: I1006 09:59:18.413229 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:18 crc kubenswrapper[4824]: I1006 09:59:18.413247 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:18Z","lastTransitionTime":"2025-10-06T09:59:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:18 crc kubenswrapper[4824]: I1006 09:59:18.516164 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:18 crc kubenswrapper[4824]: I1006 09:59:18.516218 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:18 crc kubenswrapper[4824]: I1006 09:59:18.516236 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:18 crc kubenswrapper[4824]: I1006 09:59:18.516261 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:18 crc kubenswrapper[4824]: I1006 09:59:18.516281 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:18Z","lastTransitionTime":"2025-10-06T09:59:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:18 crc kubenswrapper[4824]: I1006 09:59:18.619648 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:18 crc kubenswrapper[4824]: I1006 09:59:18.619723 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:18 crc kubenswrapper[4824]: I1006 09:59:18.619746 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:18 crc kubenswrapper[4824]: I1006 09:59:18.619779 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:18 crc kubenswrapper[4824]: I1006 09:59:18.619802 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:18Z","lastTransitionTime":"2025-10-06T09:59:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:18 crc kubenswrapper[4824]: I1006 09:59:18.722569 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:18 crc kubenswrapper[4824]: I1006 09:59:18.722630 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:18 crc kubenswrapper[4824]: I1006 09:59:18.722641 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:18 crc kubenswrapper[4824]: I1006 09:59:18.722662 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:18 crc kubenswrapper[4824]: I1006 09:59:18.722675 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:18Z","lastTransitionTime":"2025-10-06T09:59:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:18 crc kubenswrapper[4824]: I1006 09:59:18.825355 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:18 crc kubenswrapper[4824]: I1006 09:59:18.825405 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:18 crc kubenswrapper[4824]: I1006 09:59:18.825416 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:18 crc kubenswrapper[4824]: I1006 09:59:18.825432 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:18 crc kubenswrapper[4824]: I1006 09:59:18.825444 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:18Z","lastTransitionTime":"2025-10-06T09:59:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:18 crc kubenswrapper[4824]: I1006 09:59:18.928398 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:18 crc kubenswrapper[4824]: I1006 09:59:18.928448 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:18 crc kubenswrapper[4824]: I1006 09:59:18.928465 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:18 crc kubenswrapper[4824]: I1006 09:59:18.928493 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:18 crc kubenswrapper[4824]: I1006 09:59:18.928512 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:18Z","lastTransitionTime":"2025-10-06T09:59:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:19 crc kubenswrapper[4824]: I1006 09:59:19.031604 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:19 crc kubenswrapper[4824]: I1006 09:59:19.031646 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:19 crc kubenswrapper[4824]: I1006 09:59:19.031656 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:19 crc kubenswrapper[4824]: I1006 09:59:19.031670 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:19 crc kubenswrapper[4824]: I1006 09:59:19.031682 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:19Z","lastTransitionTime":"2025-10-06T09:59:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:19 crc kubenswrapper[4824]: I1006 09:59:19.134779 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:19 crc kubenswrapper[4824]: I1006 09:59:19.134825 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:19 crc kubenswrapper[4824]: I1006 09:59:19.134834 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:19 crc kubenswrapper[4824]: I1006 09:59:19.134849 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:19 crc kubenswrapper[4824]: I1006 09:59:19.134858 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:19Z","lastTransitionTime":"2025-10-06T09:59:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:19 crc kubenswrapper[4824]: I1006 09:59:19.237809 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:19 crc kubenswrapper[4824]: I1006 09:59:19.237840 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:19 crc kubenswrapper[4824]: I1006 09:59:19.237850 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:19 crc kubenswrapper[4824]: I1006 09:59:19.237861 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:19 crc kubenswrapper[4824]: I1006 09:59:19.237870 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:19Z","lastTransitionTime":"2025-10-06T09:59:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:19 crc kubenswrapper[4824]: I1006 09:59:19.340916 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:19 crc kubenswrapper[4824]: I1006 09:59:19.341009 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:19 crc kubenswrapper[4824]: I1006 09:59:19.341027 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:19 crc kubenswrapper[4824]: I1006 09:59:19.341076 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:19 crc kubenswrapper[4824]: I1006 09:59:19.341094 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:19Z","lastTransitionTime":"2025-10-06T09:59:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:19 crc kubenswrapper[4824]: I1006 09:59:19.443968 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:19 crc kubenswrapper[4824]: I1006 09:59:19.444078 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:19 crc kubenswrapper[4824]: I1006 09:59:19.444105 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:19 crc kubenswrapper[4824]: I1006 09:59:19.444136 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:19 crc kubenswrapper[4824]: I1006 09:59:19.444163 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:19Z","lastTransitionTime":"2025-10-06T09:59:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:19 crc kubenswrapper[4824]: I1006 09:59:19.547952 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:19 crc kubenswrapper[4824]: I1006 09:59:19.548065 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:19 crc kubenswrapper[4824]: I1006 09:59:19.548083 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:19 crc kubenswrapper[4824]: I1006 09:59:19.548149 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:19 crc kubenswrapper[4824]: I1006 09:59:19.548174 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:19Z","lastTransitionTime":"2025-10-06T09:59:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:19 crc kubenswrapper[4824]: I1006 09:59:19.651845 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:19 crc kubenswrapper[4824]: I1006 09:59:19.651944 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:19 crc kubenswrapper[4824]: I1006 09:59:19.652016 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:19 crc kubenswrapper[4824]: I1006 09:59:19.652055 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:19 crc kubenswrapper[4824]: I1006 09:59:19.652079 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:19Z","lastTransitionTime":"2025-10-06T09:59:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:19 crc kubenswrapper[4824]: I1006 09:59:19.755388 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:19 crc kubenswrapper[4824]: I1006 09:59:19.755461 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:19 crc kubenswrapper[4824]: I1006 09:59:19.755480 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:19 crc kubenswrapper[4824]: I1006 09:59:19.755509 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:19 crc kubenswrapper[4824]: I1006 09:59:19.755529 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:19Z","lastTransitionTime":"2025-10-06T09:59:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:19 crc kubenswrapper[4824]: I1006 09:59:19.833353 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-szn8q_b65abf9f-4e88-4571-960d-3ca997d9c344/kube-multus/0.log" Oct 06 09:59:19 crc kubenswrapper[4824]: I1006 09:59:19.833417 4824 generic.go:334] "Generic (PLEG): container finished" podID="b65abf9f-4e88-4571-960d-3ca997d9c344" containerID="664bf42e1324a9c9529f36f08d61e26999b3386bdd1fac77851fb7e7411e8934" exitCode=1 Oct 06 09:59:19 crc kubenswrapper[4824]: I1006 09:59:19.833456 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-szn8q" event={"ID":"b65abf9f-4e88-4571-960d-3ca997d9c344","Type":"ContainerDied","Data":"664bf42e1324a9c9529f36f08d61e26999b3386bdd1fac77851fb7e7411e8934"} Oct 06 09:59:19 crc kubenswrapper[4824]: I1006 09:59:19.834180 4824 scope.go:117] "RemoveContainer" containerID="664bf42e1324a9c9529f36f08d61e26999b3386bdd1fac77851fb7e7411e8934" Oct 06 09:59:19 crc kubenswrapper[4824]: I1006 09:59:19.859134 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:19 crc kubenswrapper[4824]: I1006 09:59:19.859233 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:19 crc kubenswrapper[4824]: I1006 09:59:19.859246 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:19 crc kubenswrapper[4824]: I1006 09:59:19.859291 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:19 crc kubenswrapper[4824]: I1006 09:59:19.859313 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:19Z","lastTransitionTime":"2025-10-06T09:59:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:19 crc kubenswrapper[4824]: I1006 09:59:19.861077 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-7c22q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7eaaa68-d971-4a1c-b6c5-551ded8ff27f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f5150c8c6cfc65ecdcd0e7bd52199dd60e2472c96874f103b49919c81b5e5f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53336c7d3c509b4e022708998aedc5dee9fdaf7136ae3ee69b4181394c3a22fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53336c7d3c509b4e022708998aedc5dee9fdaf7136ae3ee69b4181394c3a22fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://652ab0a13a2bb421872edda3f7188dce8e296247052550c937f8447fea206756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://652ab0a13a2bb421872edda3f7188dce8e296247052550c937f8447fea206756\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3514dec0386bb8c016a7f555d7191ea420d5058b0ee98f97d965dd857c5376e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3514dec0386bb8c016a7f555d7191ea420d5058b0ee98f97d965dd857c5376e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b035e0bf34e3429400ac8942124a3e9834231899487d03d15609fe5624747a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b035e0bf34e3429400ac8942124a3e9834231899487d03d15609fe5624747a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6743e027a11864ad3f433d110e60451fb36d49791e03b22d437b70f89d056ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6743e027a11864ad3f433d110e60451fb36d49791e03b22d437b70f89d056ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8ea48b6886fd39a1c5e5b29ca1fffc2bd4552d61dc621eb2fba66fb464cf120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8ea48b6886fd39a1c5e5b29ca1fffc2bd4552d61dc621eb2fba66fb464cf120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-7c22q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:19Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:19 crc kubenswrapper[4824]: I1006 09:59:19.878086 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b67b743d-7d7e-4a7f-b648-c721f50bb3a9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:59:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:59:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b5bb30b2f9b1a35bf1dfd2274c7e3d280757902687d9ae79b710fbd5f07f398\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0658bdad936b07337e1eb58a4d62c1d7fc98fb6b3b115d9be4c1d5d79ecbc8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://268983a8eccc2af9242e8f5ebea3722e30cd66086272ec6f681ff96ed9a0e7fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6f4a5101060399a25f1d98b0495745ab0a679b88cc6bee5ea1bf0caff7b6767\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6f4a5101060399a25f1d98b0495745ab0a679b88cc6bee5ea1bf0caff7b6767\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:11Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:19Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:19 crc kubenswrapper[4824]: I1006 09:59:19.894054 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xcgwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74317f75-9ebc-4327-b88c-598c2241240f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8def09a126f9bb910affad9b7fb32303b960add9fbfd5bb709d59b4aeba7d245\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v6sg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xcgwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:19Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:19 crc kubenswrapper[4824]: I1006 09:59:19.909432 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1140c8c-93f1-4fce-8c06-a69561ff8a6d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5462cb77b7d61aed7d4b4be2fbc6ba1451d267a66afd6dede2f259c15230e7c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gq49w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5512be36013baeca611ee92645a1d339da237b70d63aac43334beafe24513d44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gq49w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-khgzw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:19Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:19 crc kubenswrapper[4824]: I1006 09:59:19.925396 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2035278cf272827b39a45aa41ecd5b1e41aafc1d9eb8bccd853660fb6e587159\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff3152fceadf36a55578f587623f37d748f90f2afeda5702a05504b372d399be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:19Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:19 crc kubenswrapper[4824]: I1006 09:59:19.943099 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f9011f6673310dffea3acd7bdc2a56f2af91a87641631490fdc63da610f047a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:19Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:19 crc kubenswrapper[4824]: I1006 09:59:19.959729 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:19Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:19 crc kubenswrapper[4824]: I1006 09:59:19.962471 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:19 crc kubenswrapper[4824]: I1006 09:59:19.962502 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:19 crc kubenswrapper[4824]: I1006 09:59:19.962509 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:19 crc kubenswrapper[4824]: I1006 09:59:19.962524 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:19 crc kubenswrapper[4824]: I1006 09:59:19.962539 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:19Z","lastTransitionTime":"2025-10-06T09:59:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:19 crc kubenswrapper[4824]: I1006 09:59:19.977599 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:19Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:19 crc kubenswrapper[4824]: I1006 09:59:19.991706 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab36ed34d6ca02b59043c46e151019620ce31733209d072c21394a79fe36be4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:19Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:20 crc kubenswrapper[4824]: I1006 09:59:20.003666 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5rx6l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0c320fc-94d9-4d82-81ee-ccccfd125efc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f8be7bad4e9644494dfb42380c7f5481e1039a1d27ad379c623515948d87392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njtbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5rx6l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:20Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:20 crc kubenswrapper[4824]: I1006 09:59:20.019179 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d739433-0a38-42fe-b884-b8433011967e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://464e279c5903770c34d80f31ed231b194d5628b39055c4058908039d10690059\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://efa1b62728466ef7aac0e745230348ce920eb9ec9b07fb2cd5a276150dac13fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fedc40f0217626c56d155fc7e78e0592556fef9464db2aaf09c5facc486aad8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78fac300cbe2ac52f5788306b8fc0011f6ba3a22c3546974e70998f2d394987a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:20Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:20 crc kubenswrapper[4824]: I1006 09:59:20.036315 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:20Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:20 crc kubenswrapper[4824]: I1006 09:59:20.054695 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-szn8q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b65abf9f-4e88-4571-960d-3ca997d9c344\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:59:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:59:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://664bf42e1324a9c9529f36f08d61e26999b3386bdd1fac77851fb7e7411e8934\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://664bf42e1324a9c9529f36f08d61e26999b3386bdd1fac77851fb7e7411e8934\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-06T09:59:19Z\\\",\\\"message\\\":\\\"2025-10-06T09:58:34+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_cffdfe89-8b06-4a09-9c9b-7ac6a5561d38\\\\n2025-10-06T09:58:34+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_cffdfe89-8b06-4a09-9c9b-7ac6a5561d38 to /host/opt/cni/bin/\\\\n2025-10-06T09:58:34Z [verbose] multus-daemon started\\\\n2025-10-06T09:58:34Z [verbose] Readiness Indicator file check\\\\n2025-10-06T09:59:19Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49s5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-szn8q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:20Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:20 crc kubenswrapper[4824]: I1006 09:59:20.065545 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:20 crc kubenswrapper[4824]: I1006 09:59:20.065584 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:20 crc kubenswrapper[4824]: I1006 09:59:20.065601 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:20 crc kubenswrapper[4824]: I1006 09:59:20.065623 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:20 crc kubenswrapper[4824]: I1006 09:59:20.065639 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:20Z","lastTransitionTime":"2025-10-06T09:59:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:20 crc kubenswrapper[4824]: I1006 09:59:20.082129 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f511ece-25eb-465f-b85c-9e5f3f886c21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10bfed3e523b7422ce1e981f98166eca78bd12dd4a7cf0634915e1f4ecc3de2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43199f86ad4a1f02bc57a926cd96b2c3c0e2ea46fb36a3f495668706be4c8041\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d6064ad3bcdafebb4c66e7d91f183cc787e9ce13660545c71a722b3ff04ce75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6213164d366243c2139d5971041d6f076cf38fd0cce4c0ca5495111963487ec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa6d59fac987f60a4340ab35d84aa059f2b7085a06362e7207a79702d17eee6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09638fbc9d163ad0b63fa57d24ea5298207ad1f55bab8671be311960be2627a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33caf222d92beb8a42c203fc61d69e804ec3a22d6889f8ead2ce783237d8b464\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33caf222d92beb8a42c203fc61d69e804ec3a22d6889f8ead2ce783237d8b464\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-06T09:59:01Z\\\",\\\"message\\\":\\\"nt-go/informers/factory.go:160\\\\nI1006 09:59:01.318023 6499 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1006 09:59:01.317995 6499 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1006 09:59:01.318223 6499 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1006 09:59:01.318391 6499 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1006 09:59:01.319128 6499 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1006 09:59:01.319521 6499 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1006 09:59:01.319562 6499 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1006 09:59:01.319610 6499 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1006 09:59:01.319617 6499 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1006 09:59:01.319629 6499 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1006 09:59:01.319666 6499 factory.go:656] Stopping watch factory\\\\nI1006 09:59:01.319698 6499 ovnkube.go:599] Stopped ovnkube\\\\nI1006 09:59:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T09:59:00Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-4vwms_openshift-ovn-kubernetes(2f511ece-25eb-465f-b85c-9e5f3f886c21)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0644c7db28c7d26aa8d69cd5b9fbcff4d4c06b96df113137dea22a72d9fecd85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6da4c07be4d507a6f20bd8b42be6e6ea4712a956f57ef164036cfded7142a87e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6da4c07be4d507a6f20bd8b42be6e6ea4712a956f57ef164036cfded7142a87e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4vwms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:20Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:20 crc kubenswrapper[4824]: I1006 09:59:20.101108 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lszvg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0cf70ab4-4cfe-41c5-8db7-035451bafcfa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7b80b5b375b3b32132501ee0d661e06e199654c696ac76ad56e7ba96918cc77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fkbf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://519653c6b241cd89ef3ece003667d420a7a55cdf513f4577a69e14ade5fb4c76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fkbf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lszvg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:20Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:20 crc kubenswrapper[4824]: I1006 09:59:20.116944 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gnw94" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7cf1d355-64cb-48a6-acfd-7d258b7afa10\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-psbq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-psbq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:46Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gnw94\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:20Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:20 crc kubenswrapper[4824]: I1006 09:59:20.133412 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b665cda3-da4e-482e-8248-023abec382f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9235c742d8c45bf667f32979c2891fece50f4ee1bd0c7cf6a753ed2d2b9ab30e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b2a920dbc506df2426232990e31fbb60d255c00be54941826a0f094af4a92f0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d516f278f362d7c45e7149de82fa835997ac198f1165c04352910d51fb674d69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c820d8bea16cff0d463ce68d2f2acb72d7a7c73372757ff104e14fcec1025a05\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b8cdf54407cc95783919eba8e92f093bbc09eac5e908c4ba0685629948c71bd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-06T09:58:27Z\\\",\\\"message\\\":\\\"arting RequestHeaderAuthRequestController\\\\nI1006 09:58:27.156944 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1006 09:58:27.156783 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1006 09:58:27.157213 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-432254888/tls.crt::/tmp/serving-cert-432254888/tls.key\\\\\\\"\\\\nI1006 09:58:27.157407 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1006 09:58:27.157443 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1006 09:58:27.157410 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-432254888/tls.crt::/tmp/serving-cert-432254888/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759744705\\\\\\\\\\\\\\\" (2025-10-06 09:58:25 +0000 UTC to 2025-11-05 09:58:26 +0000 UTC (now=2025-10-06 09:58:27.157390076 +0000 UTC))\\\\\\\"\\\\nI1006 09:58:27.157632 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1006 09:58:27.158158 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759744707\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759744707\\\\\\\\\\\\\\\" (2025-10-06 08:58:26 +0000 UTC to 2026-10-06 08:58:26 +0000 UTC (now=2025-10-06 09:58:27.158090663 +0000 UTC))\\\\\\\"\\\\nI1006 09:58:27.158220 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1006 09:58:27.158270 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1006 09:58:27.158310 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nF1006 09:58:27.158617 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://413822c99ba2866521177cce4f1894aef17c160aebb2662da91824a04827d510\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e37e8c7400f7fd14a5c5390a5985e3d8d277e9c5ca60d5e9e9757d07d1778570\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e37e8c7400f7fd14a5c5390a5985e3d8d277e9c5ca60d5e9e9757d07d1778570\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:20Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:20 crc kubenswrapper[4824]: I1006 09:59:20.168720 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:20 crc kubenswrapper[4824]: I1006 09:59:20.168775 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:20 crc kubenswrapper[4824]: I1006 09:59:20.168786 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:20 crc kubenswrapper[4824]: I1006 09:59:20.168805 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:20 crc kubenswrapper[4824]: I1006 09:59:20.168815 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:20Z","lastTransitionTime":"2025-10-06T09:59:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:20 crc kubenswrapper[4824]: I1006 09:59:20.271747 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:20 crc kubenswrapper[4824]: I1006 09:59:20.272217 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:20 crc kubenswrapper[4824]: I1006 09:59:20.272401 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:20 crc kubenswrapper[4824]: I1006 09:59:20.272561 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:20 crc kubenswrapper[4824]: I1006 09:59:20.272734 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:20Z","lastTransitionTime":"2025-10-06T09:59:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:20 crc kubenswrapper[4824]: I1006 09:59:20.273044 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 09:59:20 crc kubenswrapper[4824]: I1006 09:59:20.273121 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 09:59:20 crc kubenswrapper[4824]: I1006 09:59:20.273210 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gnw94" Oct 06 09:59:20 crc kubenswrapper[4824]: E1006 09:59:20.273215 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 06 09:59:20 crc kubenswrapper[4824]: E1006 09:59:20.273326 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 06 09:59:20 crc kubenswrapper[4824]: E1006 09:59:20.273511 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gnw94" podUID="7cf1d355-64cb-48a6-acfd-7d258b7afa10" Oct 06 09:59:20 crc kubenswrapper[4824]: I1006 09:59:20.273531 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 09:59:20 crc kubenswrapper[4824]: E1006 09:59:20.273690 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 06 09:59:20 crc kubenswrapper[4824]: I1006 09:59:20.376535 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:20 crc kubenswrapper[4824]: I1006 09:59:20.376620 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:20 crc kubenswrapper[4824]: I1006 09:59:20.376636 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:20 crc kubenswrapper[4824]: I1006 09:59:20.376661 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:20 crc kubenswrapper[4824]: I1006 09:59:20.376672 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:20Z","lastTransitionTime":"2025-10-06T09:59:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:20 crc kubenswrapper[4824]: I1006 09:59:20.479353 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:20 crc kubenswrapper[4824]: I1006 09:59:20.479389 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:20 crc kubenswrapper[4824]: I1006 09:59:20.479398 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:20 crc kubenswrapper[4824]: I1006 09:59:20.479414 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:20 crc kubenswrapper[4824]: I1006 09:59:20.479425 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:20Z","lastTransitionTime":"2025-10-06T09:59:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:20 crc kubenswrapper[4824]: I1006 09:59:20.582437 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:20 crc kubenswrapper[4824]: I1006 09:59:20.582482 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:20 crc kubenswrapper[4824]: I1006 09:59:20.582492 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:20 crc kubenswrapper[4824]: I1006 09:59:20.582507 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:20 crc kubenswrapper[4824]: I1006 09:59:20.582518 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:20Z","lastTransitionTime":"2025-10-06T09:59:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:20 crc kubenswrapper[4824]: I1006 09:59:20.685354 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:20 crc kubenswrapper[4824]: I1006 09:59:20.685399 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:20 crc kubenswrapper[4824]: I1006 09:59:20.685408 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:20 crc kubenswrapper[4824]: I1006 09:59:20.685424 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:20 crc kubenswrapper[4824]: I1006 09:59:20.685434 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:20Z","lastTransitionTime":"2025-10-06T09:59:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:20 crc kubenswrapper[4824]: I1006 09:59:20.787647 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:20 crc kubenswrapper[4824]: I1006 09:59:20.787707 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:20 crc kubenswrapper[4824]: I1006 09:59:20.787722 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:20 crc kubenswrapper[4824]: I1006 09:59:20.787747 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:20 crc kubenswrapper[4824]: I1006 09:59:20.787768 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:20Z","lastTransitionTime":"2025-10-06T09:59:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:20 crc kubenswrapper[4824]: I1006 09:59:20.838400 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-szn8q_b65abf9f-4e88-4571-960d-3ca997d9c344/kube-multus/0.log" Oct 06 09:59:20 crc kubenswrapper[4824]: I1006 09:59:20.838458 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-szn8q" event={"ID":"b65abf9f-4e88-4571-960d-3ca997d9c344","Type":"ContainerStarted","Data":"c80f7c0d85c4c76aa3c4ff4e078900ccc3be12fa5bb75f595ead6eeacf39e3f4"} Oct 06 09:59:20 crc kubenswrapper[4824]: I1006 09:59:20.860163 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f9011f6673310dffea3acd7bdc2a56f2af91a87641631490fdc63da610f047a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:20Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:20 crc kubenswrapper[4824]: I1006 09:59:20.875655 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:20Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:20 crc kubenswrapper[4824]: I1006 09:59:20.888694 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:20Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:20 crc kubenswrapper[4824]: I1006 09:59:20.891066 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:20 crc kubenswrapper[4824]: I1006 09:59:20.891100 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:20 crc kubenswrapper[4824]: I1006 09:59:20.891112 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:20 crc kubenswrapper[4824]: I1006 09:59:20.891133 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:20 crc kubenswrapper[4824]: I1006 09:59:20.891145 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:20Z","lastTransitionTime":"2025-10-06T09:59:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:20 crc kubenswrapper[4824]: I1006 09:59:20.906743 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab36ed34d6ca02b59043c46e151019620ce31733209d072c21394a79fe36be4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:20Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:20 crc kubenswrapper[4824]: I1006 09:59:20.918189 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5rx6l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0c320fc-94d9-4d82-81ee-ccccfd125efc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f8be7bad4e9644494dfb42380c7f5481e1039a1d27ad379c623515948d87392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njtbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5rx6l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:20Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:20 crc kubenswrapper[4824]: I1006 09:59:20.939722 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b665cda3-da4e-482e-8248-023abec382f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9235c742d8c45bf667f32979c2891fece50f4ee1bd0c7cf6a753ed2d2b9ab30e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b2a920dbc506df2426232990e31fbb60d255c00be54941826a0f094af4a92f0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d516f278f362d7c45e7149de82fa835997ac198f1165c04352910d51fb674d69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c820d8bea16cff0d463ce68d2f2acb72d7a7c73372757ff104e14fcec1025a05\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b8cdf54407cc95783919eba8e92f093bbc09eac5e908c4ba0685629948c71bd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-06T09:58:27Z\\\",\\\"message\\\":\\\"arting RequestHeaderAuthRequestController\\\\nI1006 09:58:27.156944 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1006 09:58:27.156783 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1006 09:58:27.157213 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-432254888/tls.crt::/tmp/serving-cert-432254888/tls.key\\\\\\\"\\\\nI1006 09:58:27.157407 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1006 09:58:27.157443 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1006 09:58:27.157410 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-432254888/tls.crt::/tmp/serving-cert-432254888/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759744705\\\\\\\\\\\\\\\" (2025-10-06 09:58:25 +0000 UTC to 2025-11-05 09:58:26 +0000 UTC (now=2025-10-06 09:58:27.157390076 +0000 UTC))\\\\\\\"\\\\nI1006 09:58:27.157632 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1006 09:58:27.158158 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759744707\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759744707\\\\\\\\\\\\\\\" (2025-10-06 08:58:26 +0000 UTC to 2026-10-06 08:58:26 +0000 UTC (now=2025-10-06 09:58:27.158090663 +0000 UTC))\\\\\\\"\\\\nI1006 09:58:27.158220 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1006 09:58:27.158270 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1006 09:58:27.158310 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nF1006 09:58:27.158617 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://413822c99ba2866521177cce4f1894aef17c160aebb2662da91824a04827d510\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e37e8c7400f7fd14a5c5390a5985e3d8d277e9c5ca60d5e9e9757d07d1778570\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e37e8c7400f7fd14a5c5390a5985e3d8d277e9c5ca60d5e9e9757d07d1778570\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:20Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:20 crc kubenswrapper[4824]: I1006 09:59:20.970656 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d739433-0a38-42fe-b884-b8433011967e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://464e279c5903770c34d80f31ed231b194d5628b39055c4058908039d10690059\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://efa1b62728466ef7aac0e745230348ce920eb9ec9b07fb2cd5a276150dac13fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fedc40f0217626c56d155fc7e78e0592556fef9464db2aaf09c5facc486aad8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78fac300cbe2ac52f5788306b8fc0011f6ba3a22c3546974e70998f2d394987a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:20Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:20 crc kubenswrapper[4824]: I1006 09:59:20.989280 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:20Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:20 crc kubenswrapper[4824]: I1006 09:59:20.993511 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:20 crc kubenswrapper[4824]: I1006 09:59:20.993540 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:20 crc kubenswrapper[4824]: I1006 09:59:20.993552 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:20 crc kubenswrapper[4824]: I1006 09:59:20.993571 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:20 crc kubenswrapper[4824]: I1006 09:59:20.993585 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:20Z","lastTransitionTime":"2025-10-06T09:59:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:21 crc kubenswrapper[4824]: I1006 09:59:21.004701 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-szn8q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b65abf9f-4e88-4571-960d-3ca997d9c344\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:59:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:59:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c80f7c0d85c4c76aa3c4ff4e078900ccc3be12fa5bb75f595ead6eeacf39e3f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://664bf42e1324a9c9529f36f08d61e26999b3386bdd1fac77851fb7e7411e8934\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-06T09:59:19Z\\\",\\\"message\\\":\\\"2025-10-06T09:58:34+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_cffdfe89-8b06-4a09-9c9b-7ac6a5561d38\\\\n2025-10-06T09:58:34+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_cffdfe89-8b06-4a09-9c9b-7ac6a5561d38 to /host/opt/cni/bin/\\\\n2025-10-06T09:58:34Z [verbose] multus-daemon started\\\\n2025-10-06T09:58:34Z [verbose] Readiness Indicator file check\\\\n2025-10-06T09:59:19Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:59:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49s5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-szn8q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:21Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:21 crc kubenswrapper[4824]: I1006 09:59:21.034678 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f511ece-25eb-465f-b85c-9e5f3f886c21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10bfed3e523b7422ce1e981f98166eca78bd12dd4a7cf0634915e1f4ecc3de2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43199f86ad4a1f02bc57a926cd96b2c3c0e2ea46fb36a3f495668706be4c8041\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d6064ad3bcdafebb4c66e7d91f183cc787e9ce13660545c71a722b3ff04ce75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6213164d366243c2139d5971041d6f076cf38fd0cce4c0ca5495111963487ec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa6d59fac987f60a4340ab35d84aa059f2b7085a06362e7207a79702d17eee6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09638fbc9d163ad0b63fa57d24ea5298207ad1f55bab8671be311960be2627a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33caf222d92beb8a42c203fc61d69e804ec3a22d6889f8ead2ce783237d8b464\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33caf222d92beb8a42c203fc61d69e804ec3a22d6889f8ead2ce783237d8b464\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-06T09:59:01Z\\\",\\\"message\\\":\\\"nt-go/informers/factory.go:160\\\\nI1006 09:59:01.318023 6499 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1006 09:59:01.317995 6499 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1006 09:59:01.318223 6499 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1006 09:59:01.318391 6499 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1006 09:59:01.319128 6499 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1006 09:59:01.319521 6499 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1006 09:59:01.319562 6499 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1006 09:59:01.319610 6499 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1006 09:59:01.319617 6499 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1006 09:59:01.319629 6499 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1006 09:59:01.319666 6499 factory.go:656] Stopping watch factory\\\\nI1006 09:59:01.319698 6499 ovnkube.go:599] Stopped ovnkube\\\\nI1006 09:59:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T09:59:00Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-4vwms_openshift-ovn-kubernetes(2f511ece-25eb-465f-b85c-9e5f3f886c21)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0644c7db28c7d26aa8d69cd5b9fbcff4d4c06b96df113137dea22a72d9fecd85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6da4c07be4d507a6f20bd8b42be6e6ea4712a956f57ef164036cfded7142a87e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6da4c07be4d507a6f20bd8b42be6e6ea4712a956f57ef164036cfded7142a87e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4vwms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:21Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:21 crc kubenswrapper[4824]: I1006 09:59:21.044658 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lszvg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0cf70ab4-4cfe-41c5-8db7-035451bafcfa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7b80b5b375b3b32132501ee0d661e06e199654c696ac76ad56e7ba96918cc77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fkbf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://519653c6b241cd89ef3ece003667d420a7a55cdf513f4577a69e14ade5fb4c76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fkbf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lszvg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:21Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:21 crc kubenswrapper[4824]: I1006 09:59:21.056095 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gnw94" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7cf1d355-64cb-48a6-acfd-7d258b7afa10\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-psbq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-psbq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:46Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gnw94\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:21Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:21 crc kubenswrapper[4824]: I1006 09:59:21.067018 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b67b743d-7d7e-4a7f-b648-c721f50bb3a9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:59:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:59:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b5bb30b2f9b1a35bf1dfd2274c7e3d280757902687d9ae79b710fbd5f07f398\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0658bdad936b07337e1eb58a4d62c1d7fc98fb6b3b115d9be4c1d5d79ecbc8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://268983a8eccc2af9242e8f5ebea3722e30cd66086272ec6f681ff96ed9a0e7fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6f4a5101060399a25f1d98b0495745ab0a679b88cc6bee5ea1bf0caff7b6767\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6f4a5101060399a25f1d98b0495745ab0a679b88cc6bee5ea1bf0caff7b6767\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:11Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:21Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:21 crc kubenswrapper[4824]: I1006 09:59:21.081037 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-7c22q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7eaaa68-d971-4a1c-b6c5-551ded8ff27f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f5150c8c6cfc65ecdcd0e7bd52199dd60e2472c96874f103b49919c81b5e5f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53336c7d3c509b4e022708998aedc5dee9fdaf7136ae3ee69b4181394c3a22fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53336c7d3c509b4e022708998aedc5dee9fdaf7136ae3ee69b4181394c3a22fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://652ab0a13a2bb421872edda3f7188dce8e296247052550c937f8447fea206756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://652ab0a13a2bb421872edda3f7188dce8e296247052550c937f8447fea206756\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3514dec0386bb8c016a7f555d7191ea420d5058b0ee98f97d965dd857c5376e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3514dec0386bb8c016a7f555d7191ea420d5058b0ee98f97d965dd857c5376e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b035e0bf34e3429400ac8942124a3e9834231899487d03d15609fe5624747a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b035e0bf34e3429400ac8942124a3e9834231899487d03d15609fe5624747a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6743e027a11864ad3f433d110e60451fb36d49791e03b22d437b70f89d056ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6743e027a11864ad3f433d110e60451fb36d49791e03b22d437b70f89d056ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8ea48b6886fd39a1c5e5b29ca1fffc2bd4552d61dc621eb2fba66fb464cf120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8ea48b6886fd39a1c5e5b29ca1fffc2bd4552d61dc621eb2fba66fb464cf120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-7c22q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:21Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:21 crc kubenswrapper[4824]: I1006 09:59:21.093135 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2035278cf272827b39a45aa41ecd5b1e41aafc1d9eb8bccd853660fb6e587159\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff3152fceadf36a55578f587623f37d748f90f2afeda5702a05504b372d399be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:21Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:21 crc kubenswrapper[4824]: I1006 09:59:21.095860 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:21 crc kubenswrapper[4824]: I1006 09:59:21.095905 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:21 crc kubenswrapper[4824]: I1006 09:59:21.095924 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:21 crc kubenswrapper[4824]: I1006 09:59:21.095951 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:21 crc kubenswrapper[4824]: I1006 09:59:21.095971 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:21Z","lastTransitionTime":"2025-10-06T09:59:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:21 crc kubenswrapper[4824]: I1006 09:59:21.106954 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xcgwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74317f75-9ebc-4327-b88c-598c2241240f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8def09a126f9bb910affad9b7fb32303b960add9fbfd5bb709d59b4aeba7d245\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v6sg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xcgwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:21Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:21 crc kubenswrapper[4824]: I1006 09:59:21.117467 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1140c8c-93f1-4fce-8c06-a69561ff8a6d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5462cb77b7d61aed7d4b4be2fbc6ba1451d267a66afd6dede2f259c15230e7c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gq49w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5512be36013baeca611ee92645a1d339da237b70d63aac43334beafe24513d44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gq49w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-khgzw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:21Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:21 crc kubenswrapper[4824]: I1006 09:59:21.198567 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:21 crc kubenswrapper[4824]: I1006 09:59:21.198625 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:21 crc kubenswrapper[4824]: I1006 09:59:21.198645 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:21 crc kubenswrapper[4824]: I1006 09:59:21.198673 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:21 crc kubenswrapper[4824]: I1006 09:59:21.198693 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:21Z","lastTransitionTime":"2025-10-06T09:59:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:21 crc kubenswrapper[4824]: I1006 09:59:21.295446 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f9011f6673310dffea3acd7bdc2a56f2af91a87641631490fdc63da610f047a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:21Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:21 crc kubenswrapper[4824]: I1006 09:59:21.301522 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:21 crc kubenswrapper[4824]: I1006 09:59:21.301630 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:21 crc kubenswrapper[4824]: I1006 09:59:21.301709 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:21 crc kubenswrapper[4824]: I1006 09:59:21.301793 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:21 crc kubenswrapper[4824]: I1006 09:59:21.301866 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:21Z","lastTransitionTime":"2025-10-06T09:59:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:21 crc kubenswrapper[4824]: I1006 09:59:21.313139 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:21Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:21 crc kubenswrapper[4824]: I1006 09:59:21.334090 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:21Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:21 crc kubenswrapper[4824]: I1006 09:59:21.353020 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab36ed34d6ca02b59043c46e151019620ce31733209d072c21394a79fe36be4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:21Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:21 crc kubenswrapper[4824]: I1006 09:59:21.369485 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5rx6l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0c320fc-94d9-4d82-81ee-ccccfd125efc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f8be7bad4e9644494dfb42380c7f5481e1039a1d27ad379c623515948d87392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njtbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5rx6l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:21Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:21 crc kubenswrapper[4824]: I1006 09:59:21.391553 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b665cda3-da4e-482e-8248-023abec382f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9235c742d8c45bf667f32979c2891fece50f4ee1bd0c7cf6a753ed2d2b9ab30e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b2a920dbc506df2426232990e31fbb60d255c00be54941826a0f094af4a92f0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d516f278f362d7c45e7149de82fa835997ac198f1165c04352910d51fb674d69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c820d8bea16cff0d463ce68d2f2acb72d7a7c73372757ff104e14fcec1025a05\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b8cdf54407cc95783919eba8e92f093bbc09eac5e908c4ba0685629948c71bd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-06T09:58:27Z\\\",\\\"message\\\":\\\"arting RequestHeaderAuthRequestController\\\\nI1006 09:58:27.156944 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1006 09:58:27.156783 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1006 09:58:27.157213 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-432254888/tls.crt::/tmp/serving-cert-432254888/tls.key\\\\\\\"\\\\nI1006 09:58:27.157407 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1006 09:58:27.157443 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1006 09:58:27.157410 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-432254888/tls.crt::/tmp/serving-cert-432254888/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759744705\\\\\\\\\\\\\\\" (2025-10-06 09:58:25 +0000 UTC to 2025-11-05 09:58:26 +0000 UTC (now=2025-10-06 09:58:27.157390076 +0000 UTC))\\\\\\\"\\\\nI1006 09:58:27.157632 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1006 09:58:27.158158 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759744707\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759744707\\\\\\\\\\\\\\\" (2025-10-06 08:58:26 +0000 UTC to 2026-10-06 08:58:26 +0000 UTC (now=2025-10-06 09:58:27.158090663 +0000 UTC))\\\\\\\"\\\\nI1006 09:58:27.158220 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1006 09:58:27.158270 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1006 09:58:27.158310 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nF1006 09:58:27.158617 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://413822c99ba2866521177cce4f1894aef17c160aebb2662da91824a04827d510\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e37e8c7400f7fd14a5c5390a5985e3d8d277e9c5ca60d5e9e9757d07d1778570\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e37e8c7400f7fd14a5c5390a5985e3d8d277e9c5ca60d5e9e9757d07d1778570\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:21Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:21 crc kubenswrapper[4824]: I1006 09:59:21.404234 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:21 crc kubenswrapper[4824]: I1006 09:59:21.404260 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:21 crc kubenswrapper[4824]: I1006 09:59:21.404271 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:21 crc kubenswrapper[4824]: I1006 09:59:21.404286 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:21 crc kubenswrapper[4824]: I1006 09:59:21.404297 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:21Z","lastTransitionTime":"2025-10-06T09:59:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:21 crc kubenswrapper[4824]: I1006 09:59:21.412366 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d739433-0a38-42fe-b884-b8433011967e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://464e279c5903770c34d80f31ed231b194d5628b39055c4058908039d10690059\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://efa1b62728466ef7aac0e745230348ce920eb9ec9b07fb2cd5a276150dac13fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fedc40f0217626c56d155fc7e78e0592556fef9464db2aaf09c5facc486aad8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78fac300cbe2ac52f5788306b8fc0011f6ba3a22c3546974e70998f2d394987a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:21Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:21 crc kubenswrapper[4824]: I1006 09:59:21.426333 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:21Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:21 crc kubenswrapper[4824]: I1006 09:59:21.481903 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-szn8q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b65abf9f-4e88-4571-960d-3ca997d9c344\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:59:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:59:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c80f7c0d85c4c76aa3c4ff4e078900ccc3be12fa5bb75f595ead6eeacf39e3f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://664bf42e1324a9c9529f36f08d61e26999b3386bdd1fac77851fb7e7411e8934\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-06T09:59:19Z\\\",\\\"message\\\":\\\"2025-10-06T09:58:34+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_cffdfe89-8b06-4a09-9c9b-7ac6a5561d38\\\\n2025-10-06T09:58:34+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_cffdfe89-8b06-4a09-9c9b-7ac6a5561d38 to /host/opt/cni/bin/\\\\n2025-10-06T09:58:34Z [verbose] multus-daemon started\\\\n2025-10-06T09:58:34Z [verbose] Readiness Indicator file check\\\\n2025-10-06T09:59:19Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:59:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49s5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-szn8q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:21Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:21 crc kubenswrapper[4824]: I1006 09:59:21.507248 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:21 crc kubenswrapper[4824]: I1006 09:59:21.507286 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:21 crc kubenswrapper[4824]: I1006 09:59:21.507296 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:21 crc kubenswrapper[4824]: I1006 09:59:21.507313 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:21 crc kubenswrapper[4824]: I1006 09:59:21.507325 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:21Z","lastTransitionTime":"2025-10-06T09:59:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:21 crc kubenswrapper[4824]: I1006 09:59:21.517577 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f511ece-25eb-465f-b85c-9e5f3f886c21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10bfed3e523b7422ce1e981f98166eca78bd12dd4a7cf0634915e1f4ecc3de2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43199f86ad4a1f02bc57a926cd96b2c3c0e2ea46fb36a3f495668706be4c8041\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d6064ad3bcdafebb4c66e7d91f183cc787e9ce13660545c71a722b3ff04ce75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6213164d366243c2139d5971041d6f076cf38fd0cce4c0ca5495111963487ec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa6d59fac987f60a4340ab35d84aa059f2b7085a06362e7207a79702d17eee6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09638fbc9d163ad0b63fa57d24ea5298207ad1f55bab8671be311960be2627a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33caf222d92beb8a42c203fc61d69e804ec3a22d6889f8ead2ce783237d8b464\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33caf222d92beb8a42c203fc61d69e804ec3a22d6889f8ead2ce783237d8b464\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-06T09:59:01Z\\\",\\\"message\\\":\\\"nt-go/informers/factory.go:160\\\\nI1006 09:59:01.318023 6499 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1006 09:59:01.317995 6499 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1006 09:59:01.318223 6499 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1006 09:59:01.318391 6499 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1006 09:59:01.319128 6499 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1006 09:59:01.319521 6499 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1006 09:59:01.319562 6499 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1006 09:59:01.319610 6499 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1006 09:59:01.319617 6499 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1006 09:59:01.319629 6499 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1006 09:59:01.319666 6499 factory.go:656] Stopping watch factory\\\\nI1006 09:59:01.319698 6499 ovnkube.go:599] Stopped ovnkube\\\\nI1006 09:59:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T09:59:00Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-4vwms_openshift-ovn-kubernetes(2f511ece-25eb-465f-b85c-9e5f3f886c21)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0644c7db28c7d26aa8d69cd5b9fbcff4d4c06b96df113137dea22a72d9fecd85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6da4c07be4d507a6f20bd8b42be6e6ea4712a956f57ef164036cfded7142a87e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6da4c07be4d507a6f20bd8b42be6e6ea4712a956f57ef164036cfded7142a87e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4vwms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:21Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:21 crc kubenswrapper[4824]: I1006 09:59:21.531390 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lszvg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0cf70ab4-4cfe-41c5-8db7-035451bafcfa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7b80b5b375b3b32132501ee0d661e06e199654c696ac76ad56e7ba96918cc77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fkbf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://519653c6b241cd89ef3ece003667d420a7a55cdf513f4577a69e14ade5fb4c76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fkbf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lszvg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:21Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:21 crc kubenswrapper[4824]: I1006 09:59:21.543453 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gnw94" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7cf1d355-64cb-48a6-acfd-7d258b7afa10\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-psbq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-psbq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:46Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gnw94\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:21Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:21 crc kubenswrapper[4824]: I1006 09:59:21.555918 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b67b743d-7d7e-4a7f-b648-c721f50bb3a9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:59:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:59:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b5bb30b2f9b1a35bf1dfd2274c7e3d280757902687d9ae79b710fbd5f07f398\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0658bdad936b07337e1eb58a4d62c1d7fc98fb6b3b115d9be4c1d5d79ecbc8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://268983a8eccc2af9242e8f5ebea3722e30cd66086272ec6f681ff96ed9a0e7fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6f4a5101060399a25f1d98b0495745ab0a679b88cc6bee5ea1bf0caff7b6767\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6f4a5101060399a25f1d98b0495745ab0a679b88cc6bee5ea1bf0caff7b6767\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:11Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:21Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:21 crc kubenswrapper[4824]: I1006 09:59:21.572938 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-7c22q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7eaaa68-d971-4a1c-b6c5-551ded8ff27f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f5150c8c6cfc65ecdcd0e7bd52199dd60e2472c96874f103b49919c81b5e5f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53336c7d3c509b4e022708998aedc5dee9fdaf7136ae3ee69b4181394c3a22fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53336c7d3c509b4e022708998aedc5dee9fdaf7136ae3ee69b4181394c3a22fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://652ab0a13a2bb421872edda3f7188dce8e296247052550c937f8447fea206756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://652ab0a13a2bb421872edda3f7188dce8e296247052550c937f8447fea206756\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3514dec0386bb8c016a7f555d7191ea420d5058b0ee98f97d965dd857c5376e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3514dec0386bb8c016a7f555d7191ea420d5058b0ee98f97d965dd857c5376e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b035e0bf34e3429400ac8942124a3e9834231899487d03d15609fe5624747a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b035e0bf34e3429400ac8942124a3e9834231899487d03d15609fe5624747a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6743e027a11864ad3f433d110e60451fb36d49791e03b22d437b70f89d056ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6743e027a11864ad3f433d110e60451fb36d49791e03b22d437b70f89d056ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8ea48b6886fd39a1c5e5b29ca1fffc2bd4552d61dc621eb2fba66fb464cf120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8ea48b6886fd39a1c5e5b29ca1fffc2bd4552d61dc621eb2fba66fb464cf120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-7c22q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:21Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:21 crc kubenswrapper[4824]: I1006 09:59:21.583849 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2035278cf272827b39a45aa41ecd5b1e41aafc1d9eb8bccd853660fb6e587159\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff3152fceadf36a55578f587623f37d748f90f2afeda5702a05504b372d399be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:21Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:21 crc kubenswrapper[4824]: I1006 09:59:21.595184 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xcgwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74317f75-9ebc-4327-b88c-598c2241240f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8def09a126f9bb910affad9b7fb32303b960add9fbfd5bb709d59b4aeba7d245\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v6sg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xcgwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:21Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:21 crc kubenswrapper[4824]: I1006 09:59:21.606902 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1140c8c-93f1-4fce-8c06-a69561ff8a6d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5462cb77b7d61aed7d4b4be2fbc6ba1451d267a66afd6dede2f259c15230e7c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gq49w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5512be36013baeca611ee92645a1d339da237b70d63aac43334beafe24513d44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gq49w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-khgzw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:21Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:21 crc kubenswrapper[4824]: I1006 09:59:21.610009 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:21 crc kubenswrapper[4824]: I1006 09:59:21.610045 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:21 crc kubenswrapper[4824]: I1006 09:59:21.610055 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:21 crc kubenswrapper[4824]: I1006 09:59:21.610076 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:21 crc kubenswrapper[4824]: I1006 09:59:21.610087 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:21Z","lastTransitionTime":"2025-10-06T09:59:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:21 crc kubenswrapper[4824]: I1006 09:59:21.711654 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:21 crc kubenswrapper[4824]: I1006 09:59:21.711697 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:21 crc kubenswrapper[4824]: I1006 09:59:21.711709 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:21 crc kubenswrapper[4824]: I1006 09:59:21.711726 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:21 crc kubenswrapper[4824]: I1006 09:59:21.711736 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:21Z","lastTransitionTime":"2025-10-06T09:59:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:21 crc kubenswrapper[4824]: I1006 09:59:21.814213 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:21 crc kubenswrapper[4824]: I1006 09:59:21.814240 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:21 crc kubenswrapper[4824]: I1006 09:59:21.814249 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:21 crc kubenswrapper[4824]: I1006 09:59:21.814262 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:21 crc kubenswrapper[4824]: I1006 09:59:21.814270 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:21Z","lastTransitionTime":"2025-10-06T09:59:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:21 crc kubenswrapper[4824]: I1006 09:59:21.917831 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:21 crc kubenswrapper[4824]: I1006 09:59:21.917894 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:21 crc kubenswrapper[4824]: I1006 09:59:21.917913 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:21 crc kubenswrapper[4824]: I1006 09:59:21.917936 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:21 crc kubenswrapper[4824]: I1006 09:59:21.917958 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:21Z","lastTransitionTime":"2025-10-06T09:59:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:22 crc kubenswrapper[4824]: I1006 09:59:22.021624 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:22 crc kubenswrapper[4824]: I1006 09:59:22.021710 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:22 crc kubenswrapper[4824]: I1006 09:59:22.021737 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:22 crc kubenswrapper[4824]: I1006 09:59:22.021775 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:22 crc kubenswrapper[4824]: I1006 09:59:22.021801 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:22Z","lastTransitionTime":"2025-10-06T09:59:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:22 crc kubenswrapper[4824]: I1006 09:59:22.124671 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:22 crc kubenswrapper[4824]: I1006 09:59:22.124742 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:22 crc kubenswrapper[4824]: I1006 09:59:22.124760 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:22 crc kubenswrapper[4824]: I1006 09:59:22.124787 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:22 crc kubenswrapper[4824]: I1006 09:59:22.124807 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:22Z","lastTransitionTime":"2025-10-06T09:59:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:22 crc kubenswrapper[4824]: I1006 09:59:22.227747 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:22 crc kubenswrapper[4824]: I1006 09:59:22.227788 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:22 crc kubenswrapper[4824]: I1006 09:59:22.227803 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:22 crc kubenswrapper[4824]: I1006 09:59:22.227821 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:22 crc kubenswrapper[4824]: I1006 09:59:22.227832 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:22Z","lastTransitionTime":"2025-10-06T09:59:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:22 crc kubenswrapper[4824]: I1006 09:59:22.274182 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 09:59:22 crc kubenswrapper[4824]: I1006 09:59:22.274271 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 09:59:22 crc kubenswrapper[4824]: I1006 09:59:22.274313 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gnw94" Oct 06 09:59:22 crc kubenswrapper[4824]: E1006 09:59:22.274363 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 06 09:59:22 crc kubenswrapper[4824]: I1006 09:59:22.274194 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 09:59:22 crc kubenswrapper[4824]: E1006 09:59:22.274611 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gnw94" podUID="7cf1d355-64cb-48a6-acfd-7d258b7afa10" Oct 06 09:59:22 crc kubenswrapper[4824]: E1006 09:59:22.274684 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 06 09:59:22 crc kubenswrapper[4824]: E1006 09:59:22.274827 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 06 09:59:22 crc kubenswrapper[4824]: I1006 09:59:22.331333 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:22 crc kubenswrapper[4824]: I1006 09:59:22.331409 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:22 crc kubenswrapper[4824]: I1006 09:59:22.331430 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:22 crc kubenswrapper[4824]: I1006 09:59:22.331461 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:22 crc kubenswrapper[4824]: I1006 09:59:22.331482 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:22Z","lastTransitionTime":"2025-10-06T09:59:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:22 crc kubenswrapper[4824]: I1006 09:59:22.433938 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:22 crc kubenswrapper[4824]: I1006 09:59:22.434010 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:22 crc kubenswrapper[4824]: I1006 09:59:22.434022 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:22 crc kubenswrapper[4824]: I1006 09:59:22.434043 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:22 crc kubenswrapper[4824]: I1006 09:59:22.434056 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:22Z","lastTransitionTime":"2025-10-06T09:59:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:22 crc kubenswrapper[4824]: I1006 09:59:22.536929 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:22 crc kubenswrapper[4824]: I1006 09:59:22.537029 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:22 crc kubenswrapper[4824]: I1006 09:59:22.537048 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:22 crc kubenswrapper[4824]: I1006 09:59:22.537077 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:22 crc kubenswrapper[4824]: I1006 09:59:22.537095 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:22Z","lastTransitionTime":"2025-10-06T09:59:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:22 crc kubenswrapper[4824]: I1006 09:59:22.640151 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:22 crc kubenswrapper[4824]: I1006 09:59:22.640226 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:22 crc kubenswrapper[4824]: I1006 09:59:22.640246 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:22 crc kubenswrapper[4824]: I1006 09:59:22.640283 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:22 crc kubenswrapper[4824]: I1006 09:59:22.640305 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:22Z","lastTransitionTime":"2025-10-06T09:59:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:22 crc kubenswrapper[4824]: I1006 09:59:22.743524 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:22 crc kubenswrapper[4824]: I1006 09:59:22.743583 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:22 crc kubenswrapper[4824]: I1006 09:59:22.743595 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:22 crc kubenswrapper[4824]: I1006 09:59:22.743614 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:22 crc kubenswrapper[4824]: I1006 09:59:22.743632 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:22Z","lastTransitionTime":"2025-10-06T09:59:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:22 crc kubenswrapper[4824]: I1006 09:59:22.846358 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:22 crc kubenswrapper[4824]: I1006 09:59:22.846422 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:22 crc kubenswrapper[4824]: I1006 09:59:22.846439 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:22 crc kubenswrapper[4824]: I1006 09:59:22.846463 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:22 crc kubenswrapper[4824]: I1006 09:59:22.846483 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:22Z","lastTransitionTime":"2025-10-06T09:59:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:22 crc kubenswrapper[4824]: I1006 09:59:22.949748 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:22 crc kubenswrapper[4824]: I1006 09:59:22.949806 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:22 crc kubenswrapper[4824]: I1006 09:59:22.949823 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:22 crc kubenswrapper[4824]: I1006 09:59:22.949848 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:22 crc kubenswrapper[4824]: I1006 09:59:22.949866 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:22Z","lastTransitionTime":"2025-10-06T09:59:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:23 crc kubenswrapper[4824]: I1006 09:59:23.053959 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:23 crc kubenswrapper[4824]: I1006 09:59:23.054065 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:23 crc kubenswrapper[4824]: I1006 09:59:23.054085 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:23 crc kubenswrapper[4824]: I1006 09:59:23.054117 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:23 crc kubenswrapper[4824]: I1006 09:59:23.054135 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:23Z","lastTransitionTime":"2025-10-06T09:59:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:23 crc kubenswrapper[4824]: I1006 09:59:23.157594 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:23 crc kubenswrapper[4824]: I1006 09:59:23.157669 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:23 crc kubenswrapper[4824]: I1006 09:59:23.157694 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:23 crc kubenswrapper[4824]: I1006 09:59:23.157721 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:23 crc kubenswrapper[4824]: I1006 09:59:23.157744 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:23Z","lastTransitionTime":"2025-10-06T09:59:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:23 crc kubenswrapper[4824]: I1006 09:59:23.261481 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:23 crc kubenswrapper[4824]: I1006 09:59:23.261539 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:23 crc kubenswrapper[4824]: I1006 09:59:23.261556 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:23 crc kubenswrapper[4824]: I1006 09:59:23.261580 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:23 crc kubenswrapper[4824]: I1006 09:59:23.261596 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:23Z","lastTransitionTime":"2025-10-06T09:59:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:23 crc kubenswrapper[4824]: I1006 09:59:23.295918 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Oct 06 09:59:23 crc kubenswrapper[4824]: I1006 09:59:23.364327 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:23 crc kubenswrapper[4824]: I1006 09:59:23.364373 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:23 crc kubenswrapper[4824]: I1006 09:59:23.364385 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:23 crc kubenswrapper[4824]: I1006 09:59:23.364403 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:23 crc kubenswrapper[4824]: I1006 09:59:23.364415 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:23Z","lastTransitionTime":"2025-10-06T09:59:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:23 crc kubenswrapper[4824]: I1006 09:59:23.467635 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:23 crc kubenswrapper[4824]: I1006 09:59:23.467674 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:23 crc kubenswrapper[4824]: I1006 09:59:23.467685 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:23 crc kubenswrapper[4824]: I1006 09:59:23.467701 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:23 crc kubenswrapper[4824]: I1006 09:59:23.467711 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:23Z","lastTransitionTime":"2025-10-06T09:59:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:23 crc kubenswrapper[4824]: I1006 09:59:23.570414 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:23 crc kubenswrapper[4824]: I1006 09:59:23.570831 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:23 crc kubenswrapper[4824]: I1006 09:59:23.571169 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:23 crc kubenswrapper[4824]: I1006 09:59:23.571318 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:23 crc kubenswrapper[4824]: I1006 09:59:23.571462 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:23Z","lastTransitionTime":"2025-10-06T09:59:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:23 crc kubenswrapper[4824]: I1006 09:59:23.675028 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:23 crc kubenswrapper[4824]: I1006 09:59:23.675130 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:23 crc kubenswrapper[4824]: I1006 09:59:23.675149 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:23 crc kubenswrapper[4824]: I1006 09:59:23.675220 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:23 crc kubenswrapper[4824]: I1006 09:59:23.675239 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:23Z","lastTransitionTime":"2025-10-06T09:59:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:23 crc kubenswrapper[4824]: I1006 09:59:23.777713 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:23 crc kubenswrapper[4824]: I1006 09:59:23.777766 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:23 crc kubenswrapper[4824]: I1006 09:59:23.777777 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:23 crc kubenswrapper[4824]: I1006 09:59:23.777796 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:23 crc kubenswrapper[4824]: I1006 09:59:23.777810 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:23Z","lastTransitionTime":"2025-10-06T09:59:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:23 crc kubenswrapper[4824]: I1006 09:59:23.881558 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:23 crc kubenswrapper[4824]: I1006 09:59:23.881630 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:23 crc kubenswrapper[4824]: I1006 09:59:23.881645 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:23 crc kubenswrapper[4824]: I1006 09:59:23.881667 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:23 crc kubenswrapper[4824]: I1006 09:59:23.881683 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:23Z","lastTransitionTime":"2025-10-06T09:59:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:23 crc kubenswrapper[4824]: I1006 09:59:23.985140 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:23 crc kubenswrapper[4824]: I1006 09:59:23.985214 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:23 crc kubenswrapper[4824]: I1006 09:59:23.985230 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:23 crc kubenswrapper[4824]: I1006 09:59:23.985257 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:23 crc kubenswrapper[4824]: I1006 09:59:23.985274 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:23Z","lastTransitionTime":"2025-10-06T09:59:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:24 crc kubenswrapper[4824]: I1006 09:59:24.089756 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:24 crc kubenswrapper[4824]: I1006 09:59:24.089829 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:24 crc kubenswrapper[4824]: I1006 09:59:24.089853 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:24 crc kubenswrapper[4824]: I1006 09:59:24.089883 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:24 crc kubenswrapper[4824]: I1006 09:59:24.089904 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:24Z","lastTransitionTime":"2025-10-06T09:59:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:24 crc kubenswrapper[4824]: I1006 09:59:24.192127 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:24 crc kubenswrapper[4824]: I1006 09:59:24.192170 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:24 crc kubenswrapper[4824]: I1006 09:59:24.192181 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:24 crc kubenswrapper[4824]: I1006 09:59:24.192200 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:24 crc kubenswrapper[4824]: I1006 09:59:24.192213 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:24Z","lastTransitionTime":"2025-10-06T09:59:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:24 crc kubenswrapper[4824]: I1006 09:59:24.273303 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 09:59:24 crc kubenswrapper[4824]: I1006 09:59:24.273365 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 09:59:24 crc kubenswrapper[4824]: I1006 09:59:24.273379 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gnw94" Oct 06 09:59:24 crc kubenswrapper[4824]: I1006 09:59:24.273303 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 09:59:24 crc kubenswrapper[4824]: E1006 09:59:24.273455 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 06 09:59:24 crc kubenswrapper[4824]: E1006 09:59:24.273600 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gnw94" podUID="7cf1d355-64cb-48a6-acfd-7d258b7afa10" Oct 06 09:59:24 crc kubenswrapper[4824]: E1006 09:59:24.273720 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 06 09:59:24 crc kubenswrapper[4824]: E1006 09:59:24.273817 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 06 09:59:24 crc kubenswrapper[4824]: I1006 09:59:24.294378 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:24 crc kubenswrapper[4824]: I1006 09:59:24.294413 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:24 crc kubenswrapper[4824]: I1006 09:59:24.294423 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:24 crc kubenswrapper[4824]: I1006 09:59:24.294437 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:24 crc kubenswrapper[4824]: I1006 09:59:24.294452 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:24Z","lastTransitionTime":"2025-10-06T09:59:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:24 crc kubenswrapper[4824]: I1006 09:59:24.397173 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:24 crc kubenswrapper[4824]: I1006 09:59:24.397204 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:24 crc kubenswrapper[4824]: I1006 09:59:24.397214 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:24 crc kubenswrapper[4824]: I1006 09:59:24.397226 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:24 crc kubenswrapper[4824]: I1006 09:59:24.397235 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:24Z","lastTransitionTime":"2025-10-06T09:59:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:24 crc kubenswrapper[4824]: I1006 09:59:24.499595 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:24 crc kubenswrapper[4824]: I1006 09:59:24.499642 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:24 crc kubenswrapper[4824]: I1006 09:59:24.499652 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:24 crc kubenswrapper[4824]: I1006 09:59:24.499671 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:24 crc kubenswrapper[4824]: I1006 09:59:24.499686 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:24Z","lastTransitionTime":"2025-10-06T09:59:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:24 crc kubenswrapper[4824]: I1006 09:59:24.602633 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:24 crc kubenswrapper[4824]: I1006 09:59:24.602695 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:24 crc kubenswrapper[4824]: I1006 09:59:24.602716 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:24 crc kubenswrapper[4824]: I1006 09:59:24.602740 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:24 crc kubenswrapper[4824]: I1006 09:59:24.602760 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:24Z","lastTransitionTime":"2025-10-06T09:59:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:24 crc kubenswrapper[4824]: I1006 09:59:24.705805 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:24 crc kubenswrapper[4824]: I1006 09:59:24.705859 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:24 crc kubenswrapper[4824]: I1006 09:59:24.705878 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:24 crc kubenswrapper[4824]: I1006 09:59:24.705902 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:24 crc kubenswrapper[4824]: I1006 09:59:24.705922 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:24Z","lastTransitionTime":"2025-10-06T09:59:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:24 crc kubenswrapper[4824]: I1006 09:59:24.808378 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:24 crc kubenswrapper[4824]: I1006 09:59:24.808471 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:24 crc kubenswrapper[4824]: I1006 09:59:24.808492 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:24 crc kubenswrapper[4824]: I1006 09:59:24.808516 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:24 crc kubenswrapper[4824]: I1006 09:59:24.808535 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:24Z","lastTransitionTime":"2025-10-06T09:59:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:24 crc kubenswrapper[4824]: I1006 09:59:24.911258 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:24 crc kubenswrapper[4824]: I1006 09:59:24.911331 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:24 crc kubenswrapper[4824]: I1006 09:59:24.911356 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:24 crc kubenswrapper[4824]: I1006 09:59:24.911392 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:24 crc kubenswrapper[4824]: I1006 09:59:24.911416 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:24Z","lastTransitionTime":"2025-10-06T09:59:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:25 crc kubenswrapper[4824]: I1006 09:59:25.015587 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:25 crc kubenswrapper[4824]: I1006 09:59:25.015651 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:25 crc kubenswrapper[4824]: I1006 09:59:25.015669 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:25 crc kubenswrapper[4824]: I1006 09:59:25.015699 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:25 crc kubenswrapper[4824]: I1006 09:59:25.015720 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:25Z","lastTransitionTime":"2025-10-06T09:59:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:25 crc kubenswrapper[4824]: I1006 09:59:25.118453 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:25 crc kubenswrapper[4824]: I1006 09:59:25.118510 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:25 crc kubenswrapper[4824]: I1006 09:59:25.118523 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:25 crc kubenswrapper[4824]: I1006 09:59:25.118546 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:25 crc kubenswrapper[4824]: I1006 09:59:25.118560 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:25Z","lastTransitionTime":"2025-10-06T09:59:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:25 crc kubenswrapper[4824]: I1006 09:59:25.221665 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:25 crc kubenswrapper[4824]: I1006 09:59:25.221735 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:25 crc kubenswrapper[4824]: I1006 09:59:25.221759 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:25 crc kubenswrapper[4824]: I1006 09:59:25.221788 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:25 crc kubenswrapper[4824]: I1006 09:59:25.221813 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:25Z","lastTransitionTime":"2025-10-06T09:59:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:25 crc kubenswrapper[4824]: I1006 09:59:25.324581 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:25 crc kubenswrapper[4824]: I1006 09:59:25.324644 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:25 crc kubenswrapper[4824]: I1006 09:59:25.324662 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:25 crc kubenswrapper[4824]: I1006 09:59:25.324688 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:25 crc kubenswrapper[4824]: I1006 09:59:25.324714 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:25Z","lastTransitionTime":"2025-10-06T09:59:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:25 crc kubenswrapper[4824]: I1006 09:59:25.428144 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:25 crc kubenswrapper[4824]: I1006 09:59:25.428337 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:25 crc kubenswrapper[4824]: I1006 09:59:25.428365 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:25 crc kubenswrapper[4824]: I1006 09:59:25.428470 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:25 crc kubenswrapper[4824]: I1006 09:59:25.428555 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:25Z","lastTransitionTime":"2025-10-06T09:59:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:25 crc kubenswrapper[4824]: I1006 09:59:25.531876 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:25 crc kubenswrapper[4824]: I1006 09:59:25.531959 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:25 crc kubenswrapper[4824]: I1006 09:59:25.532003 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:25 crc kubenswrapper[4824]: I1006 09:59:25.532030 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:25 crc kubenswrapper[4824]: I1006 09:59:25.532048 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:25Z","lastTransitionTime":"2025-10-06T09:59:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:25 crc kubenswrapper[4824]: I1006 09:59:25.635732 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:25 crc kubenswrapper[4824]: I1006 09:59:25.635812 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:25 crc kubenswrapper[4824]: I1006 09:59:25.635834 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:25 crc kubenswrapper[4824]: I1006 09:59:25.635865 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:25 crc kubenswrapper[4824]: I1006 09:59:25.635891 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:25Z","lastTransitionTime":"2025-10-06T09:59:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:25 crc kubenswrapper[4824]: I1006 09:59:25.739478 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:25 crc kubenswrapper[4824]: I1006 09:59:25.739546 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:25 crc kubenswrapper[4824]: I1006 09:59:25.739571 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:25 crc kubenswrapper[4824]: I1006 09:59:25.739600 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:25 crc kubenswrapper[4824]: I1006 09:59:25.739621 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:25Z","lastTransitionTime":"2025-10-06T09:59:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:25 crc kubenswrapper[4824]: I1006 09:59:25.841919 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:25 crc kubenswrapper[4824]: I1006 09:59:25.842042 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:25 crc kubenswrapper[4824]: I1006 09:59:25.842063 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:25 crc kubenswrapper[4824]: I1006 09:59:25.842090 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:25 crc kubenswrapper[4824]: I1006 09:59:25.842114 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:25Z","lastTransitionTime":"2025-10-06T09:59:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:25 crc kubenswrapper[4824]: I1006 09:59:25.945099 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:25 crc kubenswrapper[4824]: I1006 09:59:25.945173 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:25 crc kubenswrapper[4824]: I1006 09:59:25.945197 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:25 crc kubenswrapper[4824]: I1006 09:59:25.945231 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:25 crc kubenswrapper[4824]: I1006 09:59:25.945257 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:25Z","lastTransitionTime":"2025-10-06T09:59:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:26 crc kubenswrapper[4824]: I1006 09:59:26.048395 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:26 crc kubenswrapper[4824]: I1006 09:59:26.048455 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:26 crc kubenswrapper[4824]: I1006 09:59:26.048471 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:26 crc kubenswrapper[4824]: I1006 09:59:26.048495 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:26 crc kubenswrapper[4824]: I1006 09:59:26.048513 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:26Z","lastTransitionTime":"2025-10-06T09:59:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:26 crc kubenswrapper[4824]: I1006 09:59:26.151441 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:26 crc kubenswrapper[4824]: I1006 09:59:26.151502 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:26 crc kubenswrapper[4824]: I1006 09:59:26.151517 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:26 crc kubenswrapper[4824]: I1006 09:59:26.151539 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:26 crc kubenswrapper[4824]: I1006 09:59:26.151555 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:26Z","lastTransitionTime":"2025-10-06T09:59:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:26 crc kubenswrapper[4824]: I1006 09:59:26.254302 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:26 crc kubenswrapper[4824]: I1006 09:59:26.254378 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:26 crc kubenswrapper[4824]: I1006 09:59:26.254397 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:26 crc kubenswrapper[4824]: I1006 09:59:26.254429 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:26 crc kubenswrapper[4824]: I1006 09:59:26.254450 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:26Z","lastTransitionTime":"2025-10-06T09:59:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:26 crc kubenswrapper[4824]: I1006 09:59:26.274078 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gnw94" Oct 06 09:59:26 crc kubenswrapper[4824]: I1006 09:59:26.274155 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 09:59:26 crc kubenswrapper[4824]: I1006 09:59:26.274191 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 09:59:26 crc kubenswrapper[4824]: I1006 09:59:26.274192 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 09:59:26 crc kubenswrapper[4824]: E1006 09:59:26.274256 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gnw94" podUID="7cf1d355-64cb-48a6-acfd-7d258b7afa10" Oct 06 09:59:26 crc kubenswrapper[4824]: E1006 09:59:26.274410 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 06 09:59:26 crc kubenswrapper[4824]: E1006 09:59:26.274559 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 06 09:59:26 crc kubenswrapper[4824]: E1006 09:59:26.274654 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 06 09:59:26 crc kubenswrapper[4824]: I1006 09:59:26.357270 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:26 crc kubenswrapper[4824]: I1006 09:59:26.357297 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:26 crc kubenswrapper[4824]: I1006 09:59:26.357311 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:26 crc kubenswrapper[4824]: I1006 09:59:26.357326 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:26 crc kubenswrapper[4824]: I1006 09:59:26.357336 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:26Z","lastTransitionTime":"2025-10-06T09:59:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:26 crc kubenswrapper[4824]: I1006 09:59:26.459359 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:26 crc kubenswrapper[4824]: I1006 09:59:26.459400 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:26 crc kubenswrapper[4824]: I1006 09:59:26.459410 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:26 crc kubenswrapper[4824]: I1006 09:59:26.459428 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:26 crc kubenswrapper[4824]: I1006 09:59:26.459441 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:26Z","lastTransitionTime":"2025-10-06T09:59:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:26 crc kubenswrapper[4824]: I1006 09:59:26.562646 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:26 crc kubenswrapper[4824]: I1006 09:59:26.562708 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:26 crc kubenswrapper[4824]: I1006 09:59:26.562726 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:26 crc kubenswrapper[4824]: I1006 09:59:26.562754 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:26 crc kubenswrapper[4824]: I1006 09:59:26.562773 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:26Z","lastTransitionTime":"2025-10-06T09:59:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:26 crc kubenswrapper[4824]: I1006 09:59:26.665555 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:26 crc kubenswrapper[4824]: I1006 09:59:26.665641 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:26 crc kubenswrapper[4824]: I1006 09:59:26.665658 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:26 crc kubenswrapper[4824]: I1006 09:59:26.665682 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:26 crc kubenswrapper[4824]: I1006 09:59:26.665700 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:26Z","lastTransitionTime":"2025-10-06T09:59:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:26 crc kubenswrapper[4824]: I1006 09:59:26.769733 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:26 crc kubenswrapper[4824]: I1006 09:59:26.769824 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:26 crc kubenswrapper[4824]: I1006 09:59:26.769848 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:26 crc kubenswrapper[4824]: I1006 09:59:26.769888 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:26 crc kubenswrapper[4824]: I1006 09:59:26.769913 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:26Z","lastTransitionTime":"2025-10-06T09:59:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:26 crc kubenswrapper[4824]: I1006 09:59:26.872823 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:26 crc kubenswrapper[4824]: I1006 09:59:26.872886 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:26 crc kubenswrapper[4824]: I1006 09:59:26.872901 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:26 crc kubenswrapper[4824]: I1006 09:59:26.872920 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:26 crc kubenswrapper[4824]: I1006 09:59:26.872935 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:26Z","lastTransitionTime":"2025-10-06T09:59:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:26 crc kubenswrapper[4824]: I1006 09:59:26.975967 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:26 crc kubenswrapper[4824]: I1006 09:59:26.976065 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:26 crc kubenswrapper[4824]: I1006 09:59:26.976085 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:26 crc kubenswrapper[4824]: I1006 09:59:26.976116 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:26 crc kubenswrapper[4824]: I1006 09:59:26.976138 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:26Z","lastTransitionTime":"2025-10-06T09:59:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:27 crc kubenswrapper[4824]: I1006 09:59:27.080197 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:27 crc kubenswrapper[4824]: I1006 09:59:27.080277 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:27 crc kubenswrapper[4824]: I1006 09:59:27.080297 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:27 crc kubenswrapper[4824]: I1006 09:59:27.080326 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:27 crc kubenswrapper[4824]: I1006 09:59:27.080344 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:27Z","lastTransitionTime":"2025-10-06T09:59:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:27 crc kubenswrapper[4824]: I1006 09:59:27.183433 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:27 crc kubenswrapper[4824]: I1006 09:59:27.183509 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:27 crc kubenswrapper[4824]: I1006 09:59:27.183528 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:27 crc kubenswrapper[4824]: I1006 09:59:27.183555 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:27 crc kubenswrapper[4824]: I1006 09:59:27.183573 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:27Z","lastTransitionTime":"2025-10-06T09:59:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:27 crc kubenswrapper[4824]: I1006 09:59:27.287058 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:27 crc kubenswrapper[4824]: I1006 09:59:27.287133 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:27 crc kubenswrapper[4824]: I1006 09:59:27.287156 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:27 crc kubenswrapper[4824]: I1006 09:59:27.287186 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:27 crc kubenswrapper[4824]: I1006 09:59:27.287212 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:27Z","lastTransitionTime":"2025-10-06T09:59:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:27 crc kubenswrapper[4824]: I1006 09:59:27.391112 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:27 crc kubenswrapper[4824]: I1006 09:59:27.391161 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:27 crc kubenswrapper[4824]: I1006 09:59:27.391176 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:27 crc kubenswrapper[4824]: I1006 09:59:27.391198 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:27 crc kubenswrapper[4824]: I1006 09:59:27.391215 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:27Z","lastTransitionTime":"2025-10-06T09:59:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:27 crc kubenswrapper[4824]: I1006 09:59:27.495027 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:27 crc kubenswrapper[4824]: I1006 09:59:27.495163 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:27 crc kubenswrapper[4824]: I1006 09:59:27.495187 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:27 crc kubenswrapper[4824]: I1006 09:59:27.495213 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:27 crc kubenswrapper[4824]: I1006 09:59:27.495231 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:27Z","lastTransitionTime":"2025-10-06T09:59:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:27 crc kubenswrapper[4824]: I1006 09:59:27.599315 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:27 crc kubenswrapper[4824]: I1006 09:59:27.599373 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:27 crc kubenswrapper[4824]: I1006 09:59:27.599391 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:27 crc kubenswrapper[4824]: I1006 09:59:27.599418 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:27 crc kubenswrapper[4824]: I1006 09:59:27.599437 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:27Z","lastTransitionTime":"2025-10-06T09:59:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:27 crc kubenswrapper[4824]: I1006 09:59:27.702803 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:27 crc kubenswrapper[4824]: I1006 09:59:27.702870 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:27 crc kubenswrapper[4824]: I1006 09:59:27.702888 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:27 crc kubenswrapper[4824]: I1006 09:59:27.702914 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:27 crc kubenswrapper[4824]: I1006 09:59:27.702933 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:27Z","lastTransitionTime":"2025-10-06T09:59:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:27 crc kubenswrapper[4824]: I1006 09:59:27.806367 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:27 crc kubenswrapper[4824]: I1006 09:59:27.806456 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:27 crc kubenswrapper[4824]: I1006 09:59:27.806478 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:27 crc kubenswrapper[4824]: I1006 09:59:27.806509 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:27 crc kubenswrapper[4824]: I1006 09:59:27.806540 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:27Z","lastTransitionTime":"2025-10-06T09:59:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:27 crc kubenswrapper[4824]: I1006 09:59:27.909999 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:27 crc kubenswrapper[4824]: I1006 09:59:27.910070 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:27 crc kubenswrapper[4824]: I1006 09:59:27.910090 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:27 crc kubenswrapper[4824]: I1006 09:59:27.910117 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:27 crc kubenswrapper[4824]: I1006 09:59:27.910136 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:27Z","lastTransitionTime":"2025-10-06T09:59:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:28 crc kubenswrapper[4824]: I1006 09:59:28.013810 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:28 crc kubenswrapper[4824]: I1006 09:59:28.013871 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:28 crc kubenswrapper[4824]: I1006 09:59:28.013889 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:28 crc kubenswrapper[4824]: I1006 09:59:28.013915 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:28 crc kubenswrapper[4824]: I1006 09:59:28.013936 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:28Z","lastTransitionTime":"2025-10-06T09:59:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:28 crc kubenswrapper[4824]: I1006 09:59:28.062693 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:28 crc kubenswrapper[4824]: I1006 09:59:28.062756 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:28 crc kubenswrapper[4824]: I1006 09:59:28.062775 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:28 crc kubenswrapper[4824]: I1006 09:59:28.062805 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:28 crc kubenswrapper[4824]: I1006 09:59:28.062833 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:28Z","lastTransitionTime":"2025-10-06T09:59:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:28 crc kubenswrapper[4824]: E1006 09:59:28.083715 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T09:59:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T09:59:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T09:59:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T09:59:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T09:59:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T09:59:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T09:59:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T09:59:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"740c2e7e-c94c-411f-b080-eb42c9aecdca\\\",\\\"systemUUID\\\":\\\"77448683-227c-4c6b-a334-eb212350758a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:28Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:28 crc kubenswrapper[4824]: I1006 09:59:28.089774 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:28 crc kubenswrapper[4824]: I1006 09:59:28.089834 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:28 crc kubenswrapper[4824]: I1006 09:59:28.089914 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:28 crc kubenswrapper[4824]: I1006 09:59:28.089947 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:28 crc kubenswrapper[4824]: I1006 09:59:28.090054 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:28Z","lastTransitionTime":"2025-10-06T09:59:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:28 crc kubenswrapper[4824]: E1006 09:59:28.108357 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T09:59:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T09:59:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T09:59:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T09:59:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T09:59:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T09:59:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T09:59:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T09:59:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"740c2e7e-c94c-411f-b080-eb42c9aecdca\\\",\\\"systemUUID\\\":\\\"77448683-227c-4c6b-a334-eb212350758a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:28Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:28 crc kubenswrapper[4824]: I1006 09:59:28.113770 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:28 crc kubenswrapper[4824]: I1006 09:59:28.113859 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:28 crc kubenswrapper[4824]: I1006 09:59:28.113886 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:28 crc kubenswrapper[4824]: I1006 09:59:28.113922 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:28 crc kubenswrapper[4824]: I1006 09:59:28.113949 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:28Z","lastTransitionTime":"2025-10-06T09:59:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:28 crc kubenswrapper[4824]: E1006 09:59:28.136767 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T09:59:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T09:59:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T09:59:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T09:59:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T09:59:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T09:59:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T09:59:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T09:59:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"740c2e7e-c94c-411f-b080-eb42c9aecdca\\\",\\\"systemUUID\\\":\\\"77448683-227c-4c6b-a334-eb212350758a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:28Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:28 crc kubenswrapper[4824]: I1006 09:59:28.141828 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:28 crc kubenswrapper[4824]: I1006 09:59:28.141890 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:28 crc kubenswrapper[4824]: I1006 09:59:28.141907 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:28 crc kubenswrapper[4824]: I1006 09:59:28.141932 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:28 crc kubenswrapper[4824]: I1006 09:59:28.141951 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:28Z","lastTransitionTime":"2025-10-06T09:59:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:28 crc kubenswrapper[4824]: E1006 09:59:28.160846 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T09:59:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T09:59:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T09:59:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T09:59:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T09:59:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T09:59:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T09:59:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T09:59:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"740c2e7e-c94c-411f-b080-eb42c9aecdca\\\",\\\"systemUUID\\\":\\\"77448683-227c-4c6b-a334-eb212350758a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:28Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:28 crc kubenswrapper[4824]: I1006 09:59:28.166307 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:28 crc kubenswrapper[4824]: I1006 09:59:28.166358 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:28 crc kubenswrapper[4824]: I1006 09:59:28.166375 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:28 crc kubenswrapper[4824]: I1006 09:59:28.166399 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:28 crc kubenswrapper[4824]: I1006 09:59:28.166420 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:28Z","lastTransitionTime":"2025-10-06T09:59:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:28 crc kubenswrapper[4824]: E1006 09:59:28.189433 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T09:59:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T09:59:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T09:59:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T09:59:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T09:59:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T09:59:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T09:59:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T09:59:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"740c2e7e-c94c-411f-b080-eb42c9aecdca\\\",\\\"systemUUID\\\":\\\"77448683-227c-4c6b-a334-eb212350758a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:28Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:28 crc kubenswrapper[4824]: E1006 09:59:28.189682 4824 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 06 09:59:28 crc kubenswrapper[4824]: I1006 09:59:28.192203 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:28 crc kubenswrapper[4824]: I1006 09:59:28.192253 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:28 crc kubenswrapper[4824]: I1006 09:59:28.192273 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:28 crc kubenswrapper[4824]: I1006 09:59:28.192304 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:28 crc kubenswrapper[4824]: I1006 09:59:28.192327 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:28Z","lastTransitionTime":"2025-10-06T09:59:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:28 crc kubenswrapper[4824]: I1006 09:59:28.273531 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 09:59:28 crc kubenswrapper[4824]: E1006 09:59:28.273704 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 06 09:59:28 crc kubenswrapper[4824]: I1006 09:59:28.273923 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 09:59:28 crc kubenswrapper[4824]: E1006 09:59:28.274011 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 06 09:59:28 crc kubenswrapper[4824]: I1006 09:59:28.274293 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 09:59:28 crc kubenswrapper[4824]: I1006 09:59:28.274329 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gnw94" Oct 06 09:59:28 crc kubenswrapper[4824]: E1006 09:59:28.274495 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 06 09:59:28 crc kubenswrapper[4824]: E1006 09:59:28.274798 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gnw94" podUID="7cf1d355-64cb-48a6-acfd-7d258b7afa10" Oct 06 09:59:28 crc kubenswrapper[4824]: I1006 09:59:28.295291 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:28 crc kubenswrapper[4824]: I1006 09:59:28.295341 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:28 crc kubenswrapper[4824]: I1006 09:59:28.295358 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:28 crc kubenswrapper[4824]: I1006 09:59:28.295385 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:28 crc kubenswrapper[4824]: I1006 09:59:28.295405 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:28Z","lastTransitionTime":"2025-10-06T09:59:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:28 crc kubenswrapper[4824]: I1006 09:59:28.399682 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:28 crc kubenswrapper[4824]: I1006 09:59:28.400134 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:28 crc kubenswrapper[4824]: I1006 09:59:28.400335 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:28 crc kubenswrapper[4824]: I1006 09:59:28.400519 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:28 crc kubenswrapper[4824]: I1006 09:59:28.400702 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:28Z","lastTransitionTime":"2025-10-06T09:59:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:28 crc kubenswrapper[4824]: I1006 09:59:28.504812 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:28 crc kubenswrapper[4824]: I1006 09:59:28.505218 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:28 crc kubenswrapper[4824]: I1006 09:59:28.505385 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:28 crc kubenswrapper[4824]: I1006 09:59:28.505582 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:28 crc kubenswrapper[4824]: I1006 09:59:28.505786 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:28Z","lastTransitionTime":"2025-10-06T09:59:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:28 crc kubenswrapper[4824]: I1006 09:59:28.609206 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:28 crc kubenswrapper[4824]: I1006 09:59:28.609281 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:28 crc kubenswrapper[4824]: I1006 09:59:28.609303 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:28 crc kubenswrapper[4824]: I1006 09:59:28.609333 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:28 crc kubenswrapper[4824]: I1006 09:59:28.609357 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:28Z","lastTransitionTime":"2025-10-06T09:59:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:28 crc kubenswrapper[4824]: I1006 09:59:28.713276 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:28 crc kubenswrapper[4824]: I1006 09:59:28.713334 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:28 crc kubenswrapper[4824]: I1006 09:59:28.713345 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:28 crc kubenswrapper[4824]: I1006 09:59:28.713367 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:28 crc kubenswrapper[4824]: I1006 09:59:28.713382 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:28Z","lastTransitionTime":"2025-10-06T09:59:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:28 crc kubenswrapper[4824]: I1006 09:59:28.817003 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:28 crc kubenswrapper[4824]: I1006 09:59:28.817071 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:28 crc kubenswrapper[4824]: I1006 09:59:28.817087 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:28 crc kubenswrapper[4824]: I1006 09:59:28.817112 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:28 crc kubenswrapper[4824]: I1006 09:59:28.817128 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:28Z","lastTransitionTime":"2025-10-06T09:59:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:28 crc kubenswrapper[4824]: I1006 09:59:28.920310 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:28 crc kubenswrapper[4824]: I1006 09:59:28.920357 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:28 crc kubenswrapper[4824]: I1006 09:59:28.920371 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:28 crc kubenswrapper[4824]: I1006 09:59:28.920395 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:28 crc kubenswrapper[4824]: I1006 09:59:28.920409 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:28Z","lastTransitionTime":"2025-10-06T09:59:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:29 crc kubenswrapper[4824]: I1006 09:59:29.024044 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:29 crc kubenswrapper[4824]: I1006 09:59:29.024105 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:29 crc kubenswrapper[4824]: I1006 09:59:29.024126 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:29 crc kubenswrapper[4824]: I1006 09:59:29.024153 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:29 crc kubenswrapper[4824]: I1006 09:59:29.024175 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:29Z","lastTransitionTime":"2025-10-06T09:59:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:29 crc kubenswrapper[4824]: I1006 09:59:29.127813 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:29 crc kubenswrapper[4824]: I1006 09:59:29.127885 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:29 crc kubenswrapper[4824]: I1006 09:59:29.127907 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:29 crc kubenswrapper[4824]: I1006 09:59:29.127943 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:29 crc kubenswrapper[4824]: I1006 09:59:29.127973 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:29Z","lastTransitionTime":"2025-10-06T09:59:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:29 crc kubenswrapper[4824]: I1006 09:59:29.231560 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:29 crc kubenswrapper[4824]: I1006 09:59:29.231641 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:29 crc kubenswrapper[4824]: I1006 09:59:29.231668 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:29 crc kubenswrapper[4824]: I1006 09:59:29.231699 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:29 crc kubenswrapper[4824]: I1006 09:59:29.231722 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:29Z","lastTransitionTime":"2025-10-06T09:59:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:29 crc kubenswrapper[4824]: I1006 09:59:29.274742 4824 scope.go:117] "RemoveContainer" containerID="33caf222d92beb8a42c203fc61d69e804ec3a22d6889f8ead2ce783237d8b464" Oct 06 09:59:29 crc kubenswrapper[4824]: I1006 09:59:29.335931 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:29 crc kubenswrapper[4824]: I1006 09:59:29.336037 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:29 crc kubenswrapper[4824]: I1006 09:59:29.336061 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:29 crc kubenswrapper[4824]: I1006 09:59:29.336090 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:29 crc kubenswrapper[4824]: I1006 09:59:29.336111 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:29Z","lastTransitionTime":"2025-10-06T09:59:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:29 crc kubenswrapper[4824]: I1006 09:59:29.440927 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:29 crc kubenswrapper[4824]: I1006 09:59:29.441550 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:29 crc kubenswrapper[4824]: I1006 09:59:29.441578 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:29 crc kubenswrapper[4824]: I1006 09:59:29.441609 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:29 crc kubenswrapper[4824]: I1006 09:59:29.441632 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:29Z","lastTransitionTime":"2025-10-06T09:59:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:29 crc kubenswrapper[4824]: I1006 09:59:29.545022 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:29 crc kubenswrapper[4824]: I1006 09:59:29.545110 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:29 crc kubenswrapper[4824]: I1006 09:59:29.545134 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:29 crc kubenswrapper[4824]: I1006 09:59:29.545168 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:29 crc kubenswrapper[4824]: I1006 09:59:29.545188 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:29Z","lastTransitionTime":"2025-10-06T09:59:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:29 crc kubenswrapper[4824]: I1006 09:59:29.648243 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:29 crc kubenswrapper[4824]: I1006 09:59:29.648339 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:29 crc kubenswrapper[4824]: I1006 09:59:29.648362 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:29 crc kubenswrapper[4824]: I1006 09:59:29.648403 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:29 crc kubenswrapper[4824]: I1006 09:59:29.648426 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:29Z","lastTransitionTime":"2025-10-06T09:59:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:29 crc kubenswrapper[4824]: I1006 09:59:29.751940 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:29 crc kubenswrapper[4824]: I1006 09:59:29.752019 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:29 crc kubenswrapper[4824]: I1006 09:59:29.752034 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:29 crc kubenswrapper[4824]: I1006 09:59:29.752057 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:29 crc kubenswrapper[4824]: I1006 09:59:29.752072 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:29Z","lastTransitionTime":"2025-10-06T09:59:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:29 crc kubenswrapper[4824]: I1006 09:59:29.854682 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:29 crc kubenswrapper[4824]: I1006 09:59:29.854748 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:29 crc kubenswrapper[4824]: I1006 09:59:29.854760 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:29 crc kubenswrapper[4824]: I1006 09:59:29.854784 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:29 crc kubenswrapper[4824]: I1006 09:59:29.854805 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:29Z","lastTransitionTime":"2025-10-06T09:59:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:29 crc kubenswrapper[4824]: I1006 09:59:29.876215 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-4vwms_2f511ece-25eb-465f-b85c-9e5f3f886c21/ovnkube-controller/2.log" Oct 06 09:59:29 crc kubenswrapper[4824]: I1006 09:59:29.880109 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" event={"ID":"2f511ece-25eb-465f-b85c-9e5f3f886c21","Type":"ContainerStarted","Data":"9c88940bd9c33c0ee0feef797ad549f874cad42942352072dba2b4eaf1e8b803"} Oct 06 09:59:29 crc kubenswrapper[4824]: I1006 09:59:29.880838 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" Oct 06 09:59:29 crc kubenswrapper[4824]: I1006 09:59:29.898586 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2035278cf272827b39a45aa41ecd5b1e41aafc1d9eb8bccd853660fb6e587159\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff3152fceadf36a55578f587623f37d748f90f2afeda5702a05504b372d399be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:29Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:29 crc kubenswrapper[4824]: I1006 09:59:29.917577 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xcgwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74317f75-9ebc-4327-b88c-598c2241240f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8def09a126f9bb910affad9b7fb32303b960add9fbfd5bb709d59b4aeba7d245\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v6sg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xcgwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:29Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:29 crc kubenswrapper[4824]: I1006 09:59:29.938011 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1140c8c-93f1-4fce-8c06-a69561ff8a6d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5462cb77b7d61aed7d4b4be2fbc6ba1451d267a66afd6dede2f259c15230e7c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gq49w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5512be36013baeca611ee92645a1d339da237b70d63aac43334beafe24513d44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gq49w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-khgzw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:29Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:29 crc kubenswrapper[4824]: I1006 09:59:29.958318 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:29 crc kubenswrapper[4824]: I1006 09:59:29.958369 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:29 crc kubenswrapper[4824]: I1006 09:59:29.958390 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:29 crc kubenswrapper[4824]: I1006 09:59:29.958421 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:29 crc kubenswrapper[4824]: I1006 09:59:29.958434 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:29Z","lastTransitionTime":"2025-10-06T09:59:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:29 crc kubenswrapper[4824]: I1006 09:59:29.962216 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:29Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:29 crc kubenswrapper[4824]: I1006 09:59:29.984178 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab36ed34d6ca02b59043c46e151019620ce31733209d072c21394a79fe36be4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:29Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:30 crc kubenswrapper[4824]: I1006 09:59:30.005748 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5rx6l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0c320fc-94d9-4d82-81ee-ccccfd125efc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f8be7bad4e9644494dfb42380c7f5481e1039a1d27ad379c623515948d87392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njtbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5rx6l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:30Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:30 crc kubenswrapper[4824]: I1006 09:59:30.019720 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da385c16-418f-4c00-a25d-04fd6506a1c6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79b92b42226cd9bba974d119a8e7f3d1d822e8a21cf1e6cdfff48dcbe7a5a898\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://507540c1b777046fb6b90bb1de444f904c92fbd7b6e598217bb844b25495c683\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://507540c1b777046fb6b90bb1de444f904c92fbd7b6e598217bb844b25495c683\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:30Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:30 crc kubenswrapper[4824]: I1006 09:59:30.033079 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f9011f6673310dffea3acd7bdc2a56f2af91a87641631490fdc63da610f047a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:30Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:30 crc kubenswrapper[4824]: I1006 09:59:30.053308 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:30Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:30 crc kubenswrapper[4824]: I1006 09:59:30.061452 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:30 crc kubenswrapper[4824]: I1006 09:59:30.061524 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:30 crc kubenswrapper[4824]: I1006 09:59:30.061537 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:30 crc kubenswrapper[4824]: I1006 09:59:30.061569 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:30 crc kubenswrapper[4824]: I1006 09:59:30.061589 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:30Z","lastTransitionTime":"2025-10-06T09:59:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:30 crc kubenswrapper[4824]: I1006 09:59:30.078647 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f511ece-25eb-465f-b85c-9e5f3f886c21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10bfed3e523b7422ce1e981f98166eca78bd12dd4a7cf0634915e1f4ecc3de2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43199f86ad4a1f02bc57a926cd96b2c3c0e2ea46fb36a3f495668706be4c8041\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d6064ad3bcdafebb4c66e7d91f183cc787e9ce13660545c71a722b3ff04ce75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6213164d366243c2139d5971041d6f076cf38fd0cce4c0ca5495111963487ec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa6d59fac987f60a4340ab35d84aa059f2b7085a06362e7207a79702d17eee6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09638fbc9d163ad0b63fa57d24ea5298207ad1f55bab8671be311960be2627a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c88940bd9c33c0ee0feef797ad549f874cad42942352072dba2b4eaf1e8b803\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33caf222d92beb8a42c203fc61d69e804ec3a22d6889f8ead2ce783237d8b464\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-06T09:59:01Z\\\",\\\"message\\\":\\\"nt-go/informers/factory.go:160\\\\nI1006 09:59:01.318023 6499 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1006 09:59:01.317995 6499 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1006 09:59:01.318223 6499 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1006 09:59:01.318391 6499 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1006 09:59:01.319128 6499 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1006 09:59:01.319521 6499 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1006 09:59:01.319562 6499 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1006 09:59:01.319610 6499 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1006 09:59:01.319617 6499 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1006 09:59:01.319629 6499 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1006 09:59:01.319666 6499 factory.go:656] Stopping watch factory\\\\nI1006 09:59:01.319698 6499 ovnkube.go:599] Stopped ovnkube\\\\nI1006 09:59:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T09:59:00Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:59:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0644c7db28c7d26aa8d69cd5b9fbcff4d4c06b96df113137dea22a72d9fecd85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6da4c07be4d507a6f20bd8b42be6e6ea4712a956f57ef164036cfded7142a87e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6da4c07be4d507a6f20bd8b42be6e6ea4712a956f57ef164036cfded7142a87e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4vwms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:30Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:30 crc kubenswrapper[4824]: I1006 09:59:30.099051 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lszvg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0cf70ab4-4cfe-41c5-8db7-035451bafcfa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7b80b5b375b3b32132501ee0d661e06e199654c696ac76ad56e7ba96918cc77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fkbf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://519653c6b241cd89ef3ece003667d420a7a55cdf513f4577a69e14ade5fb4c76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fkbf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lszvg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:30Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:30 crc kubenswrapper[4824]: I1006 09:59:30.112450 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gnw94" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7cf1d355-64cb-48a6-acfd-7d258b7afa10\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-psbq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-psbq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:46Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gnw94\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:30Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:30 crc kubenswrapper[4824]: I1006 09:59:30.131755 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b665cda3-da4e-482e-8248-023abec382f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9235c742d8c45bf667f32979c2891fece50f4ee1bd0c7cf6a753ed2d2b9ab30e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b2a920dbc506df2426232990e31fbb60d255c00be54941826a0f094af4a92f0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d516f278f362d7c45e7149de82fa835997ac198f1165c04352910d51fb674d69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c820d8bea16cff0d463ce68d2f2acb72d7a7c73372757ff104e14fcec1025a05\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b8cdf54407cc95783919eba8e92f093bbc09eac5e908c4ba0685629948c71bd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-06T09:58:27Z\\\",\\\"message\\\":\\\"arting RequestHeaderAuthRequestController\\\\nI1006 09:58:27.156944 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1006 09:58:27.156783 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1006 09:58:27.157213 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-432254888/tls.crt::/tmp/serving-cert-432254888/tls.key\\\\\\\"\\\\nI1006 09:58:27.157407 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1006 09:58:27.157443 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1006 09:58:27.157410 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-432254888/tls.crt::/tmp/serving-cert-432254888/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759744705\\\\\\\\\\\\\\\" (2025-10-06 09:58:25 +0000 UTC to 2025-11-05 09:58:26 +0000 UTC (now=2025-10-06 09:58:27.157390076 +0000 UTC))\\\\\\\"\\\\nI1006 09:58:27.157632 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1006 09:58:27.158158 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759744707\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759744707\\\\\\\\\\\\\\\" (2025-10-06 08:58:26 +0000 UTC to 2026-10-06 08:58:26 +0000 UTC (now=2025-10-06 09:58:27.158090663 +0000 UTC))\\\\\\\"\\\\nI1006 09:58:27.158220 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1006 09:58:27.158270 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1006 09:58:27.158310 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nF1006 09:58:27.158617 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://413822c99ba2866521177cce4f1894aef17c160aebb2662da91824a04827d510\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e37e8c7400f7fd14a5c5390a5985e3d8d277e9c5ca60d5e9e9757d07d1778570\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e37e8c7400f7fd14a5c5390a5985e3d8d277e9c5ca60d5e9e9757d07d1778570\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:30Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:30 crc kubenswrapper[4824]: I1006 09:59:30.147244 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d739433-0a38-42fe-b884-b8433011967e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://464e279c5903770c34d80f31ed231b194d5628b39055c4058908039d10690059\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://efa1b62728466ef7aac0e745230348ce920eb9ec9b07fb2cd5a276150dac13fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fedc40f0217626c56d155fc7e78e0592556fef9464db2aaf09c5facc486aad8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78fac300cbe2ac52f5788306b8fc0011f6ba3a22c3546974e70998f2d394987a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:30Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:30 crc kubenswrapper[4824]: I1006 09:59:30.163027 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:30Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:30 crc kubenswrapper[4824]: I1006 09:59:30.164370 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:30 crc kubenswrapper[4824]: I1006 09:59:30.164405 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:30 crc kubenswrapper[4824]: I1006 09:59:30.164417 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:30 crc kubenswrapper[4824]: I1006 09:59:30.164437 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:30 crc kubenswrapper[4824]: I1006 09:59:30.164448 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:30Z","lastTransitionTime":"2025-10-06T09:59:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:30 crc kubenswrapper[4824]: I1006 09:59:30.180398 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-szn8q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b65abf9f-4e88-4571-960d-3ca997d9c344\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:59:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:59:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c80f7c0d85c4c76aa3c4ff4e078900ccc3be12fa5bb75f595ead6eeacf39e3f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://664bf42e1324a9c9529f36f08d61e26999b3386bdd1fac77851fb7e7411e8934\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-06T09:59:19Z\\\",\\\"message\\\":\\\"2025-10-06T09:58:34+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_cffdfe89-8b06-4a09-9c9b-7ac6a5561d38\\\\n2025-10-06T09:58:34+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_cffdfe89-8b06-4a09-9c9b-7ac6a5561d38 to /host/opt/cni/bin/\\\\n2025-10-06T09:58:34Z [verbose] multus-daemon started\\\\n2025-10-06T09:58:34Z [verbose] Readiness Indicator file check\\\\n2025-10-06T09:59:19Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:59:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49s5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-szn8q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:30Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:30 crc kubenswrapper[4824]: I1006 09:59:30.200056 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b67b743d-7d7e-4a7f-b648-c721f50bb3a9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:59:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:59:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b5bb30b2f9b1a35bf1dfd2274c7e3d280757902687d9ae79b710fbd5f07f398\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0658bdad936b07337e1eb58a4d62c1d7fc98fb6b3b115d9be4c1d5d79ecbc8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://268983a8eccc2af9242e8f5ebea3722e30cd66086272ec6f681ff96ed9a0e7fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6f4a5101060399a25f1d98b0495745ab0a679b88cc6bee5ea1bf0caff7b6767\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6f4a5101060399a25f1d98b0495745ab0a679b88cc6bee5ea1bf0caff7b6767\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:11Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:30Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:30 crc kubenswrapper[4824]: I1006 09:59:30.221002 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-7c22q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7eaaa68-d971-4a1c-b6c5-551ded8ff27f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f5150c8c6cfc65ecdcd0e7bd52199dd60e2472c96874f103b49919c81b5e5f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53336c7d3c509b4e022708998aedc5dee9fdaf7136ae3ee69b4181394c3a22fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53336c7d3c509b4e022708998aedc5dee9fdaf7136ae3ee69b4181394c3a22fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://652ab0a13a2bb421872edda3f7188dce8e296247052550c937f8447fea206756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://652ab0a13a2bb421872edda3f7188dce8e296247052550c937f8447fea206756\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3514dec0386bb8c016a7f555d7191ea420d5058b0ee98f97d965dd857c5376e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3514dec0386bb8c016a7f555d7191ea420d5058b0ee98f97d965dd857c5376e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b035e0bf34e3429400ac8942124a3e9834231899487d03d15609fe5624747a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b035e0bf34e3429400ac8942124a3e9834231899487d03d15609fe5624747a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6743e027a11864ad3f433d110e60451fb36d49791e03b22d437b70f89d056ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6743e027a11864ad3f433d110e60451fb36d49791e03b22d437b70f89d056ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8ea48b6886fd39a1c5e5b29ca1fffc2bd4552d61dc621eb2fba66fb464cf120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8ea48b6886fd39a1c5e5b29ca1fffc2bd4552d61dc621eb2fba66fb464cf120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-7c22q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:30Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:30 crc kubenswrapper[4824]: I1006 09:59:30.267398 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:30 crc kubenswrapper[4824]: I1006 09:59:30.267467 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:30 crc kubenswrapper[4824]: I1006 09:59:30.267499 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:30 crc kubenswrapper[4824]: I1006 09:59:30.267526 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:30 crc kubenswrapper[4824]: I1006 09:59:30.267542 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:30Z","lastTransitionTime":"2025-10-06T09:59:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:30 crc kubenswrapper[4824]: I1006 09:59:30.273850 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 09:59:30 crc kubenswrapper[4824]: E1006 09:59:30.274087 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 06 09:59:30 crc kubenswrapper[4824]: I1006 09:59:30.274394 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 09:59:30 crc kubenswrapper[4824]: E1006 09:59:30.274459 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 06 09:59:30 crc kubenswrapper[4824]: I1006 09:59:30.274610 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gnw94" Oct 06 09:59:30 crc kubenswrapper[4824]: E1006 09:59:30.274678 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gnw94" podUID="7cf1d355-64cb-48a6-acfd-7d258b7afa10" Oct 06 09:59:30 crc kubenswrapper[4824]: I1006 09:59:30.275043 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 09:59:30 crc kubenswrapper[4824]: E1006 09:59:30.275115 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 06 09:59:30 crc kubenswrapper[4824]: I1006 09:59:30.370837 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:30 crc kubenswrapper[4824]: I1006 09:59:30.370889 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:30 crc kubenswrapper[4824]: I1006 09:59:30.370899 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:30 crc kubenswrapper[4824]: I1006 09:59:30.370920 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:30 crc kubenswrapper[4824]: I1006 09:59:30.370933 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:30Z","lastTransitionTime":"2025-10-06T09:59:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:30 crc kubenswrapper[4824]: I1006 09:59:30.474503 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:30 crc kubenswrapper[4824]: I1006 09:59:30.474570 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:30 crc kubenswrapper[4824]: I1006 09:59:30.474591 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:30 crc kubenswrapper[4824]: I1006 09:59:30.474621 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:30 crc kubenswrapper[4824]: I1006 09:59:30.474643 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:30Z","lastTransitionTime":"2025-10-06T09:59:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:30 crc kubenswrapper[4824]: I1006 09:59:30.577819 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:30 crc kubenswrapper[4824]: I1006 09:59:30.577895 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:30 crc kubenswrapper[4824]: I1006 09:59:30.577908 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:30 crc kubenswrapper[4824]: I1006 09:59:30.577931 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:30 crc kubenswrapper[4824]: I1006 09:59:30.577953 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:30Z","lastTransitionTime":"2025-10-06T09:59:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:30 crc kubenswrapper[4824]: I1006 09:59:30.681824 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:30 crc kubenswrapper[4824]: I1006 09:59:30.681922 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:30 crc kubenswrapper[4824]: I1006 09:59:30.681967 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:30 crc kubenswrapper[4824]: I1006 09:59:30.682040 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:30 crc kubenswrapper[4824]: I1006 09:59:30.682066 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:30Z","lastTransitionTime":"2025-10-06T09:59:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:30 crc kubenswrapper[4824]: I1006 09:59:30.786082 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:30 crc kubenswrapper[4824]: I1006 09:59:30.786213 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:30 crc kubenswrapper[4824]: I1006 09:59:30.786234 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:30 crc kubenswrapper[4824]: I1006 09:59:30.786261 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:30 crc kubenswrapper[4824]: I1006 09:59:30.786314 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:30Z","lastTransitionTime":"2025-10-06T09:59:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:30 crc kubenswrapper[4824]: I1006 09:59:30.888843 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-4vwms_2f511ece-25eb-465f-b85c-9e5f3f886c21/ovnkube-controller/3.log" Oct 06 09:59:30 crc kubenswrapper[4824]: I1006 09:59:30.888864 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:30 crc kubenswrapper[4824]: I1006 09:59:30.889009 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:30 crc kubenswrapper[4824]: I1006 09:59:30.889021 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:30 crc kubenswrapper[4824]: I1006 09:59:30.889039 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:30 crc kubenswrapper[4824]: I1006 09:59:30.889062 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:30Z","lastTransitionTime":"2025-10-06T09:59:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:30 crc kubenswrapper[4824]: I1006 09:59:30.889700 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-4vwms_2f511ece-25eb-465f-b85c-9e5f3f886c21/ovnkube-controller/2.log" Oct 06 09:59:30 crc kubenswrapper[4824]: I1006 09:59:30.893733 4824 generic.go:334] "Generic (PLEG): container finished" podID="2f511ece-25eb-465f-b85c-9e5f3f886c21" containerID="9c88940bd9c33c0ee0feef797ad549f874cad42942352072dba2b4eaf1e8b803" exitCode=1 Oct 06 09:59:30 crc kubenswrapper[4824]: I1006 09:59:30.893769 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" event={"ID":"2f511ece-25eb-465f-b85c-9e5f3f886c21","Type":"ContainerDied","Data":"9c88940bd9c33c0ee0feef797ad549f874cad42942352072dba2b4eaf1e8b803"} Oct 06 09:59:30 crc kubenswrapper[4824]: I1006 09:59:30.893854 4824 scope.go:117] "RemoveContainer" containerID="33caf222d92beb8a42c203fc61d69e804ec3a22d6889f8ead2ce783237d8b464" Oct 06 09:59:30 crc kubenswrapper[4824]: I1006 09:59:30.895335 4824 scope.go:117] "RemoveContainer" containerID="9c88940bd9c33c0ee0feef797ad549f874cad42942352072dba2b4eaf1e8b803" Oct 06 09:59:30 crc kubenswrapper[4824]: E1006 09:59:30.895669 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-4vwms_openshift-ovn-kubernetes(2f511ece-25eb-465f-b85c-9e5f3f886c21)\"" pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" podUID="2f511ece-25eb-465f-b85c-9e5f3f886c21" Oct 06 09:59:30 crc kubenswrapper[4824]: I1006 09:59:30.910120 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da385c16-418f-4c00-a25d-04fd6506a1c6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79b92b42226cd9bba974d119a8e7f3d1d822e8a21cf1e6cdfff48dcbe7a5a898\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://507540c1b777046fb6b90bb1de444f904c92fbd7b6e598217bb844b25495c683\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://507540c1b777046fb6b90bb1de444f904c92fbd7b6e598217bb844b25495c683\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:30Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:30 crc kubenswrapper[4824]: I1006 09:59:30.934520 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f9011f6673310dffea3acd7bdc2a56f2af91a87641631490fdc63da610f047a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:30Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:30 crc kubenswrapper[4824]: I1006 09:59:30.956651 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:30Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:30 crc kubenswrapper[4824]: I1006 09:59:30.979919 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:30Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:30 crc kubenswrapper[4824]: I1006 09:59:30.992698 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:30 crc kubenswrapper[4824]: I1006 09:59:30.992916 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:30 crc kubenswrapper[4824]: I1006 09:59:30.993107 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:30 crc kubenswrapper[4824]: I1006 09:59:30.993255 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:30 crc kubenswrapper[4824]: I1006 09:59:30.993378 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:30Z","lastTransitionTime":"2025-10-06T09:59:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:30 crc kubenswrapper[4824]: I1006 09:59:30.996526 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab36ed34d6ca02b59043c46e151019620ce31733209d072c21394a79fe36be4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:30Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:31 crc kubenswrapper[4824]: I1006 09:59:31.014633 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5rx6l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0c320fc-94d9-4d82-81ee-ccccfd125efc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f8be7bad4e9644494dfb42380c7f5481e1039a1d27ad379c623515948d87392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njtbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5rx6l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:31Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:31 crc kubenswrapper[4824]: I1006 09:59:31.033969 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b665cda3-da4e-482e-8248-023abec382f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9235c742d8c45bf667f32979c2891fece50f4ee1bd0c7cf6a753ed2d2b9ab30e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b2a920dbc506df2426232990e31fbb60d255c00be54941826a0f094af4a92f0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d516f278f362d7c45e7149de82fa835997ac198f1165c04352910d51fb674d69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c820d8bea16cff0d463ce68d2f2acb72d7a7c73372757ff104e14fcec1025a05\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b8cdf54407cc95783919eba8e92f093bbc09eac5e908c4ba0685629948c71bd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-06T09:58:27Z\\\",\\\"message\\\":\\\"arting RequestHeaderAuthRequestController\\\\nI1006 09:58:27.156944 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1006 09:58:27.156783 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1006 09:58:27.157213 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-432254888/tls.crt::/tmp/serving-cert-432254888/tls.key\\\\\\\"\\\\nI1006 09:58:27.157407 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1006 09:58:27.157443 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1006 09:58:27.157410 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-432254888/tls.crt::/tmp/serving-cert-432254888/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759744705\\\\\\\\\\\\\\\" (2025-10-06 09:58:25 +0000 UTC to 2025-11-05 09:58:26 +0000 UTC (now=2025-10-06 09:58:27.157390076 +0000 UTC))\\\\\\\"\\\\nI1006 09:58:27.157632 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1006 09:58:27.158158 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759744707\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759744707\\\\\\\\\\\\\\\" (2025-10-06 08:58:26 +0000 UTC to 2026-10-06 08:58:26 +0000 UTC (now=2025-10-06 09:58:27.158090663 +0000 UTC))\\\\\\\"\\\\nI1006 09:58:27.158220 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1006 09:58:27.158270 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1006 09:58:27.158310 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nF1006 09:58:27.158617 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://413822c99ba2866521177cce4f1894aef17c160aebb2662da91824a04827d510\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e37e8c7400f7fd14a5c5390a5985e3d8d277e9c5ca60d5e9e9757d07d1778570\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e37e8c7400f7fd14a5c5390a5985e3d8d277e9c5ca60d5e9e9757d07d1778570\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:31Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:31 crc kubenswrapper[4824]: I1006 09:59:31.054543 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d739433-0a38-42fe-b884-b8433011967e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://464e279c5903770c34d80f31ed231b194d5628b39055c4058908039d10690059\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://efa1b62728466ef7aac0e745230348ce920eb9ec9b07fb2cd5a276150dac13fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fedc40f0217626c56d155fc7e78e0592556fef9464db2aaf09c5facc486aad8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78fac300cbe2ac52f5788306b8fc0011f6ba3a22c3546974e70998f2d394987a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:31Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:31 crc kubenswrapper[4824]: I1006 09:59:31.074228 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:31Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:31 crc kubenswrapper[4824]: I1006 09:59:31.097025 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:31 crc kubenswrapper[4824]: I1006 09:59:31.097079 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:31 crc kubenswrapper[4824]: I1006 09:59:31.097099 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:31 crc kubenswrapper[4824]: I1006 09:59:31.097128 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:31 crc kubenswrapper[4824]: I1006 09:59:31.097148 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:31Z","lastTransitionTime":"2025-10-06T09:59:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:31 crc kubenswrapper[4824]: I1006 09:59:31.097959 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-szn8q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b65abf9f-4e88-4571-960d-3ca997d9c344\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:59:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:59:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c80f7c0d85c4c76aa3c4ff4e078900ccc3be12fa5bb75f595ead6eeacf39e3f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://664bf42e1324a9c9529f36f08d61e26999b3386bdd1fac77851fb7e7411e8934\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-06T09:59:19Z\\\",\\\"message\\\":\\\"2025-10-06T09:58:34+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_cffdfe89-8b06-4a09-9c9b-7ac6a5561d38\\\\n2025-10-06T09:58:34+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_cffdfe89-8b06-4a09-9c9b-7ac6a5561d38 to /host/opt/cni/bin/\\\\n2025-10-06T09:58:34Z [verbose] multus-daemon started\\\\n2025-10-06T09:58:34Z [verbose] Readiness Indicator file check\\\\n2025-10-06T09:59:19Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:59:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49s5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-szn8q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:31Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:31 crc kubenswrapper[4824]: I1006 09:59:31.128520 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f511ece-25eb-465f-b85c-9e5f3f886c21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10bfed3e523b7422ce1e981f98166eca78bd12dd4a7cf0634915e1f4ecc3de2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43199f86ad4a1f02bc57a926cd96b2c3c0e2ea46fb36a3f495668706be4c8041\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d6064ad3bcdafebb4c66e7d91f183cc787e9ce13660545c71a722b3ff04ce75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6213164d366243c2139d5971041d6f076cf38fd0cce4c0ca5495111963487ec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa6d59fac987f60a4340ab35d84aa059f2b7085a06362e7207a79702d17eee6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09638fbc9d163ad0b63fa57d24ea5298207ad1f55bab8671be311960be2627a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c88940bd9c33c0ee0feef797ad549f874cad42942352072dba2b4eaf1e8b803\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33caf222d92beb8a42c203fc61d69e804ec3a22d6889f8ead2ce783237d8b464\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-06T09:59:01Z\\\",\\\"message\\\":\\\"nt-go/informers/factory.go:160\\\\nI1006 09:59:01.318023 6499 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1006 09:59:01.317995 6499 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1006 09:59:01.318223 6499 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1006 09:59:01.318391 6499 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1006 09:59:01.319128 6499 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1006 09:59:01.319521 6499 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1006 09:59:01.319562 6499 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1006 09:59:01.319610 6499 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1006 09:59:01.319617 6499 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1006 09:59:01.319629 6499 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1006 09:59:01.319666 6499 factory.go:656] Stopping watch factory\\\\nI1006 09:59:01.319698 6499 ovnkube.go:599] Stopped ovnkube\\\\nI1006 09:59:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T09:59:00Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c88940bd9c33c0ee0feef797ad549f874cad42942352072dba2b4eaf1e8b803\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-06T09:59:30Z\\\",\\\"message\\\":\\\"flector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1006 09:59:30.228084 6864 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1006 09:59:30.228172 6864 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1006 09:59:30.228389 6864 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1006 09:59:30.228646 6864 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1006 09:59:30.228673 6864 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1006 09:59:30.228695 6864 handler.go:208] Removed *v1.Node event handler 2\\\\nI1006 09:59:30.228716 6864 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1006 09:59:30.228730 6864 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1006 09:59:30.228754 6864 factory.go:656] Stopping watch factory\\\\nI1006 09:59:30.228781 6864 ovnkube.go:599] Stopped ovnkube\\\\nI1006 09:59:30.228783 6864 handler.go:208] Removed *v1.Node event handler 7\\\\nI1006 09:59:30.228805 6864 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1006 09\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T09:59:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0644c7db28c7d26aa8d69cd5b9fbcff4d4c06b96df113137dea22a72d9fecd85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6da4c07be4d507a6f20bd8b42be6e6ea4712a956f57ef164036cfded7142a87e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6da4c07be4d507a6f20bd8b42be6e6ea4712a956f57ef164036cfded7142a87e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4vwms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:31Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:31 crc kubenswrapper[4824]: I1006 09:59:31.149964 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lszvg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0cf70ab4-4cfe-41c5-8db7-035451bafcfa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7b80b5b375b3b32132501ee0d661e06e199654c696ac76ad56e7ba96918cc77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fkbf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://519653c6b241cd89ef3ece003667d420a7a55cdf513f4577a69e14ade5fb4c76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fkbf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lszvg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:31Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:31 crc kubenswrapper[4824]: I1006 09:59:31.169129 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gnw94" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7cf1d355-64cb-48a6-acfd-7d258b7afa10\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-psbq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-psbq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:46Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gnw94\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:31Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:31 crc kubenswrapper[4824]: I1006 09:59:31.188814 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b67b743d-7d7e-4a7f-b648-c721f50bb3a9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:59:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:59:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b5bb30b2f9b1a35bf1dfd2274c7e3d280757902687d9ae79b710fbd5f07f398\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0658bdad936b07337e1eb58a4d62c1d7fc98fb6b3b115d9be4c1d5d79ecbc8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://268983a8eccc2af9242e8f5ebea3722e30cd66086272ec6f681ff96ed9a0e7fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6f4a5101060399a25f1d98b0495745ab0a679b88cc6bee5ea1bf0caff7b6767\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6f4a5101060399a25f1d98b0495745ab0a679b88cc6bee5ea1bf0caff7b6767\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:11Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:31Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:31 crc kubenswrapper[4824]: I1006 09:59:31.200687 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:31 crc kubenswrapper[4824]: I1006 09:59:31.200752 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:31 crc kubenswrapper[4824]: I1006 09:59:31.200772 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:31 crc kubenswrapper[4824]: I1006 09:59:31.200802 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:31 crc kubenswrapper[4824]: I1006 09:59:31.200822 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:31Z","lastTransitionTime":"2025-10-06T09:59:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:31 crc kubenswrapper[4824]: I1006 09:59:31.213196 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-7c22q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7eaaa68-d971-4a1c-b6c5-551ded8ff27f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f5150c8c6cfc65ecdcd0e7bd52199dd60e2472c96874f103b49919c81b5e5f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53336c7d3c509b4e022708998aedc5dee9fdaf7136ae3ee69b4181394c3a22fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53336c7d3c509b4e022708998aedc5dee9fdaf7136ae3ee69b4181394c3a22fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://652ab0a13a2bb421872edda3f7188dce8e296247052550c937f8447fea206756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://652ab0a13a2bb421872edda3f7188dce8e296247052550c937f8447fea206756\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3514dec0386bb8c016a7f555d7191ea420d5058b0ee98f97d965dd857c5376e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3514dec0386bb8c016a7f555d7191ea420d5058b0ee98f97d965dd857c5376e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b035e0bf34e3429400ac8942124a3e9834231899487d03d15609fe5624747a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b035e0bf34e3429400ac8942124a3e9834231899487d03d15609fe5624747a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6743e027a11864ad3f433d110e60451fb36d49791e03b22d437b70f89d056ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6743e027a11864ad3f433d110e60451fb36d49791e03b22d437b70f89d056ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8ea48b6886fd39a1c5e5b29ca1fffc2bd4552d61dc621eb2fba66fb464cf120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8ea48b6886fd39a1c5e5b29ca1fffc2bd4552d61dc621eb2fba66fb464cf120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-7c22q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:31Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:31 crc kubenswrapper[4824]: I1006 09:59:31.236313 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2035278cf272827b39a45aa41ecd5b1e41aafc1d9eb8bccd853660fb6e587159\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff3152fceadf36a55578f587623f37d748f90f2afeda5702a05504b372d399be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:31Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:31 crc kubenswrapper[4824]: I1006 09:59:31.256670 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xcgwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74317f75-9ebc-4327-b88c-598c2241240f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8def09a126f9bb910affad9b7fb32303b960add9fbfd5bb709d59b4aeba7d245\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v6sg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xcgwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:31Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:31 crc kubenswrapper[4824]: I1006 09:59:31.276359 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1140c8c-93f1-4fce-8c06-a69561ff8a6d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5462cb77b7d61aed7d4b4be2fbc6ba1451d267a66afd6dede2f259c15230e7c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gq49w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5512be36013baeca611ee92645a1d339da237b70d63aac43334beafe24513d44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gq49w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-khgzw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:31Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:31 crc kubenswrapper[4824]: I1006 09:59:31.299207 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b67b743d-7d7e-4a7f-b648-c721f50bb3a9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:59:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:59:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b5bb30b2f9b1a35bf1dfd2274c7e3d280757902687d9ae79b710fbd5f07f398\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0658bdad936b07337e1eb58a4d62c1d7fc98fb6b3b115d9be4c1d5d79ecbc8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://268983a8eccc2af9242e8f5ebea3722e30cd66086272ec6f681ff96ed9a0e7fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6f4a5101060399a25f1d98b0495745ab0a679b88cc6bee5ea1bf0caff7b6767\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6f4a5101060399a25f1d98b0495745ab0a679b88cc6bee5ea1bf0caff7b6767\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:11Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:31Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:31 crc kubenswrapper[4824]: I1006 09:59:31.304861 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:31 crc kubenswrapper[4824]: I1006 09:59:31.304907 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:31 crc kubenswrapper[4824]: I1006 09:59:31.304927 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:31 crc kubenswrapper[4824]: I1006 09:59:31.304955 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:31 crc kubenswrapper[4824]: I1006 09:59:31.305008 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:31Z","lastTransitionTime":"2025-10-06T09:59:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:31 crc kubenswrapper[4824]: I1006 09:59:31.325652 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-7c22q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7eaaa68-d971-4a1c-b6c5-551ded8ff27f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f5150c8c6cfc65ecdcd0e7bd52199dd60e2472c96874f103b49919c81b5e5f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53336c7d3c509b4e022708998aedc5dee9fdaf7136ae3ee69b4181394c3a22fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53336c7d3c509b4e022708998aedc5dee9fdaf7136ae3ee69b4181394c3a22fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://652ab0a13a2bb421872edda3f7188dce8e296247052550c937f8447fea206756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://652ab0a13a2bb421872edda3f7188dce8e296247052550c937f8447fea206756\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3514dec0386bb8c016a7f555d7191ea420d5058b0ee98f97d965dd857c5376e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3514dec0386bb8c016a7f555d7191ea420d5058b0ee98f97d965dd857c5376e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b035e0bf34e3429400ac8942124a3e9834231899487d03d15609fe5624747a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b035e0bf34e3429400ac8942124a3e9834231899487d03d15609fe5624747a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6743e027a11864ad3f433d110e60451fb36d49791e03b22d437b70f89d056ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6743e027a11864ad3f433d110e60451fb36d49791e03b22d437b70f89d056ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8ea48b6886fd39a1c5e5b29ca1fffc2bd4552d61dc621eb2fba66fb464cf120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8ea48b6886fd39a1c5e5b29ca1fffc2bd4552d61dc621eb2fba66fb464cf120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-7c22q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:31Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:31 crc kubenswrapper[4824]: I1006 09:59:31.347385 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2035278cf272827b39a45aa41ecd5b1e41aafc1d9eb8bccd853660fb6e587159\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff3152fceadf36a55578f587623f37d748f90f2afeda5702a05504b372d399be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:31Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:31 crc kubenswrapper[4824]: I1006 09:59:31.365135 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xcgwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74317f75-9ebc-4327-b88c-598c2241240f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8def09a126f9bb910affad9b7fb32303b960add9fbfd5bb709d59b4aeba7d245\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v6sg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xcgwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:31Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:31 crc kubenswrapper[4824]: I1006 09:59:31.384970 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1140c8c-93f1-4fce-8c06-a69561ff8a6d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5462cb77b7d61aed7d4b4be2fbc6ba1451d267a66afd6dede2f259c15230e7c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gq49w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5512be36013baeca611ee92645a1d339da237b70d63aac43334beafe24513d44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gq49w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-khgzw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:31Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:31 crc kubenswrapper[4824]: I1006 09:59:31.404361 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab36ed34d6ca02b59043c46e151019620ce31733209d072c21394a79fe36be4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:31Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:31 crc kubenswrapper[4824]: I1006 09:59:31.409078 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:31 crc kubenswrapper[4824]: I1006 09:59:31.409139 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:31 crc kubenswrapper[4824]: I1006 09:59:31.409158 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:31 crc kubenswrapper[4824]: I1006 09:59:31.409184 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:31 crc kubenswrapper[4824]: I1006 09:59:31.409202 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:31Z","lastTransitionTime":"2025-10-06T09:59:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:31 crc kubenswrapper[4824]: I1006 09:59:31.421091 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5rx6l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0c320fc-94d9-4d82-81ee-ccccfd125efc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f8be7bad4e9644494dfb42380c7f5481e1039a1d27ad379c623515948d87392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njtbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5rx6l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:31Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:31 crc kubenswrapper[4824]: I1006 09:59:31.438141 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da385c16-418f-4c00-a25d-04fd6506a1c6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79b92b42226cd9bba974d119a8e7f3d1d822e8a21cf1e6cdfff48dcbe7a5a898\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://507540c1b777046fb6b90bb1de444f904c92fbd7b6e598217bb844b25495c683\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://507540c1b777046fb6b90bb1de444f904c92fbd7b6e598217bb844b25495c683\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:31Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:31 crc kubenswrapper[4824]: I1006 09:59:31.459405 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f9011f6673310dffea3acd7bdc2a56f2af91a87641631490fdc63da610f047a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:31Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:31 crc kubenswrapper[4824]: I1006 09:59:31.479303 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:31Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:31 crc kubenswrapper[4824]: I1006 09:59:31.500428 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:31Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:31 crc kubenswrapper[4824]: I1006 09:59:31.512279 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:31 crc kubenswrapper[4824]: I1006 09:59:31.512326 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:31 crc kubenswrapper[4824]: I1006 09:59:31.512347 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:31 crc kubenswrapper[4824]: I1006 09:59:31.512373 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:31 crc kubenswrapper[4824]: I1006 09:59:31.512393 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:31Z","lastTransitionTime":"2025-10-06T09:59:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:31 crc kubenswrapper[4824]: I1006 09:59:31.519594 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lszvg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0cf70ab4-4cfe-41c5-8db7-035451bafcfa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7b80b5b375b3b32132501ee0d661e06e199654c696ac76ad56e7ba96918cc77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fkbf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://519653c6b241cd89ef3ece003667d420a7a55cdf513f4577a69e14ade5fb4c76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fkbf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lszvg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:31Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:31 crc kubenswrapper[4824]: I1006 09:59:31.537640 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gnw94" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7cf1d355-64cb-48a6-acfd-7d258b7afa10\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-psbq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-psbq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:46Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gnw94\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:31Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:31 crc kubenswrapper[4824]: I1006 09:59:31.560834 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b665cda3-da4e-482e-8248-023abec382f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9235c742d8c45bf667f32979c2891fece50f4ee1bd0c7cf6a753ed2d2b9ab30e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b2a920dbc506df2426232990e31fbb60d255c00be54941826a0f094af4a92f0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d516f278f362d7c45e7149de82fa835997ac198f1165c04352910d51fb674d69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c820d8bea16cff0d463ce68d2f2acb72d7a7c73372757ff104e14fcec1025a05\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b8cdf54407cc95783919eba8e92f093bbc09eac5e908c4ba0685629948c71bd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-06T09:58:27Z\\\",\\\"message\\\":\\\"arting RequestHeaderAuthRequestController\\\\nI1006 09:58:27.156944 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1006 09:58:27.156783 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1006 09:58:27.157213 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-432254888/tls.crt::/tmp/serving-cert-432254888/tls.key\\\\\\\"\\\\nI1006 09:58:27.157407 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1006 09:58:27.157443 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1006 09:58:27.157410 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-432254888/tls.crt::/tmp/serving-cert-432254888/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759744705\\\\\\\\\\\\\\\" (2025-10-06 09:58:25 +0000 UTC to 2025-11-05 09:58:26 +0000 UTC (now=2025-10-06 09:58:27.157390076 +0000 UTC))\\\\\\\"\\\\nI1006 09:58:27.157632 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1006 09:58:27.158158 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759744707\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759744707\\\\\\\\\\\\\\\" (2025-10-06 08:58:26 +0000 UTC to 2026-10-06 08:58:26 +0000 UTC (now=2025-10-06 09:58:27.158090663 +0000 UTC))\\\\\\\"\\\\nI1006 09:58:27.158220 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1006 09:58:27.158270 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1006 09:58:27.158310 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nF1006 09:58:27.158617 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://413822c99ba2866521177cce4f1894aef17c160aebb2662da91824a04827d510\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e37e8c7400f7fd14a5c5390a5985e3d8d277e9c5ca60d5e9e9757d07d1778570\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e37e8c7400f7fd14a5c5390a5985e3d8d277e9c5ca60d5e9e9757d07d1778570\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:31Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:31 crc kubenswrapper[4824]: I1006 09:59:31.583077 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d739433-0a38-42fe-b884-b8433011967e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://464e279c5903770c34d80f31ed231b194d5628b39055c4058908039d10690059\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://efa1b62728466ef7aac0e745230348ce920eb9ec9b07fb2cd5a276150dac13fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fedc40f0217626c56d155fc7e78e0592556fef9464db2aaf09c5facc486aad8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78fac300cbe2ac52f5788306b8fc0011f6ba3a22c3546974e70998f2d394987a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:31Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:31 crc kubenswrapper[4824]: I1006 09:59:31.605318 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:31Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:31 crc kubenswrapper[4824]: I1006 09:59:31.615768 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:31 crc kubenswrapper[4824]: I1006 09:59:31.615833 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:31 crc kubenswrapper[4824]: I1006 09:59:31.615853 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:31 crc kubenswrapper[4824]: I1006 09:59:31.615877 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:31 crc kubenswrapper[4824]: I1006 09:59:31.615898 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:31Z","lastTransitionTime":"2025-10-06T09:59:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:31 crc kubenswrapper[4824]: I1006 09:59:31.629146 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-szn8q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b65abf9f-4e88-4571-960d-3ca997d9c344\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:59:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:59:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c80f7c0d85c4c76aa3c4ff4e078900ccc3be12fa5bb75f595ead6eeacf39e3f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://664bf42e1324a9c9529f36f08d61e26999b3386bdd1fac77851fb7e7411e8934\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-06T09:59:19Z\\\",\\\"message\\\":\\\"2025-10-06T09:58:34+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_cffdfe89-8b06-4a09-9c9b-7ac6a5561d38\\\\n2025-10-06T09:58:34+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_cffdfe89-8b06-4a09-9c9b-7ac6a5561d38 to /host/opt/cni/bin/\\\\n2025-10-06T09:58:34Z [verbose] multus-daemon started\\\\n2025-10-06T09:58:34Z [verbose] Readiness Indicator file check\\\\n2025-10-06T09:59:19Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:59:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49s5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-szn8q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:31Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:31 crc kubenswrapper[4824]: I1006 09:59:31.661628 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f511ece-25eb-465f-b85c-9e5f3f886c21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10bfed3e523b7422ce1e981f98166eca78bd12dd4a7cf0634915e1f4ecc3de2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43199f86ad4a1f02bc57a926cd96b2c3c0e2ea46fb36a3f495668706be4c8041\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d6064ad3bcdafebb4c66e7d91f183cc787e9ce13660545c71a722b3ff04ce75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6213164d366243c2139d5971041d6f076cf38fd0cce4c0ca5495111963487ec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa6d59fac987f60a4340ab35d84aa059f2b7085a06362e7207a79702d17eee6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09638fbc9d163ad0b63fa57d24ea5298207ad1f55bab8671be311960be2627a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c88940bd9c33c0ee0feef797ad549f874cad42942352072dba2b4eaf1e8b803\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://33caf222d92beb8a42c203fc61d69e804ec3a22d6889f8ead2ce783237d8b464\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-06T09:59:01Z\\\",\\\"message\\\":\\\"nt-go/informers/factory.go:160\\\\nI1006 09:59:01.318023 6499 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1006 09:59:01.317995 6499 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1006 09:59:01.318223 6499 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1006 09:59:01.318391 6499 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1006 09:59:01.319128 6499 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1006 09:59:01.319521 6499 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1006 09:59:01.319562 6499 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1006 09:59:01.319610 6499 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1006 09:59:01.319617 6499 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1006 09:59:01.319629 6499 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1006 09:59:01.319666 6499 factory.go:656] Stopping watch factory\\\\nI1006 09:59:01.319698 6499 ovnkube.go:599] Stopped ovnkube\\\\nI1006 09:59:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T09:59:00Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c88940bd9c33c0ee0feef797ad549f874cad42942352072dba2b4eaf1e8b803\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-06T09:59:30Z\\\",\\\"message\\\":\\\"flector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1006 09:59:30.228084 6864 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1006 09:59:30.228172 6864 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1006 09:59:30.228389 6864 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1006 09:59:30.228646 6864 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1006 09:59:30.228673 6864 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1006 09:59:30.228695 6864 handler.go:208] Removed *v1.Node event handler 2\\\\nI1006 09:59:30.228716 6864 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1006 09:59:30.228730 6864 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1006 09:59:30.228754 6864 factory.go:656] Stopping watch factory\\\\nI1006 09:59:30.228781 6864 ovnkube.go:599] Stopped ovnkube\\\\nI1006 09:59:30.228783 6864 handler.go:208] Removed *v1.Node event handler 7\\\\nI1006 09:59:30.228805 6864 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1006 09\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T09:59:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0644c7db28c7d26aa8d69cd5b9fbcff4d4c06b96df113137dea22a72d9fecd85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6da4c07be4d507a6f20bd8b42be6e6ea4712a956f57ef164036cfded7142a87e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6da4c07be4d507a6f20bd8b42be6e6ea4712a956f57ef164036cfded7142a87e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4vwms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:31Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:31 crc kubenswrapper[4824]: I1006 09:59:31.718824 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:31 crc kubenswrapper[4824]: I1006 09:59:31.718888 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:31 crc kubenswrapper[4824]: I1006 09:59:31.718906 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:31 crc kubenswrapper[4824]: I1006 09:59:31.718939 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:31 crc kubenswrapper[4824]: I1006 09:59:31.718962 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:31Z","lastTransitionTime":"2025-10-06T09:59:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:31 crc kubenswrapper[4824]: I1006 09:59:31.822730 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:31 crc kubenswrapper[4824]: I1006 09:59:31.822800 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:31 crc kubenswrapper[4824]: I1006 09:59:31.822818 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:31 crc kubenswrapper[4824]: I1006 09:59:31.822849 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:31 crc kubenswrapper[4824]: I1006 09:59:31.822871 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:31Z","lastTransitionTime":"2025-10-06T09:59:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:31 crc kubenswrapper[4824]: I1006 09:59:31.901756 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-4vwms_2f511ece-25eb-465f-b85c-9e5f3f886c21/ovnkube-controller/3.log" Oct 06 09:59:31 crc kubenswrapper[4824]: I1006 09:59:31.908213 4824 scope.go:117] "RemoveContainer" containerID="9c88940bd9c33c0ee0feef797ad549f874cad42942352072dba2b4eaf1e8b803" Oct 06 09:59:31 crc kubenswrapper[4824]: E1006 09:59:31.908509 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-4vwms_openshift-ovn-kubernetes(2f511ece-25eb-465f-b85c-9e5f3f886c21)\"" pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" podUID="2f511ece-25eb-465f-b85c-9e5f3f886c21" Oct 06 09:59:31 crc kubenswrapper[4824]: I1006 09:59:31.926868 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:31 crc kubenswrapper[4824]: I1006 09:59:31.926955 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:31 crc kubenswrapper[4824]: I1006 09:59:31.927048 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:31 crc kubenswrapper[4824]: I1006 09:59:31.927118 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:31 crc kubenswrapper[4824]: I1006 09:59:31.927145 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:31Z","lastTransitionTime":"2025-10-06T09:59:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:31 crc kubenswrapper[4824]: I1006 09:59:31.929457 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2035278cf272827b39a45aa41ecd5b1e41aafc1d9eb8bccd853660fb6e587159\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff3152fceadf36a55578f587623f37d748f90f2afeda5702a05504b372d399be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:31Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:31 crc kubenswrapper[4824]: I1006 09:59:31.949100 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xcgwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74317f75-9ebc-4327-b88c-598c2241240f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8def09a126f9bb910affad9b7fb32303b960add9fbfd5bb709d59b4aeba7d245\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7v6sg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xcgwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:31Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:31 crc kubenswrapper[4824]: I1006 09:59:31.968158 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1140c8c-93f1-4fce-8c06-a69561ff8a6d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5462cb77b7d61aed7d4b4be2fbc6ba1451d267a66afd6dede2f259c15230e7c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gq49w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5512be36013baeca611ee92645a1d339da237b70d63aac43334beafe24513d44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gq49w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-khgzw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:31Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:31 crc kubenswrapper[4824]: I1006 09:59:31.988851 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:31Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:32 crc kubenswrapper[4824]: I1006 09:59:32.013100 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:32Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:32 crc kubenswrapper[4824]: I1006 09:59:32.031569 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:32 crc kubenswrapper[4824]: I1006 09:59:32.031645 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:32 crc kubenswrapper[4824]: I1006 09:59:32.031663 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:32 crc kubenswrapper[4824]: I1006 09:59:32.031693 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:32 crc kubenswrapper[4824]: I1006 09:59:32.031712 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:32Z","lastTransitionTime":"2025-10-06T09:59:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:32 crc kubenswrapper[4824]: I1006 09:59:32.034637 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab36ed34d6ca02b59043c46e151019620ce31733209d072c21394a79fe36be4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:32Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:32 crc kubenswrapper[4824]: I1006 09:59:32.052809 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5rx6l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0c320fc-94d9-4d82-81ee-ccccfd125efc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f8be7bad4e9644494dfb42380c7f5481e1039a1d27ad379c623515948d87392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njtbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5rx6l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:32Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:32 crc kubenswrapper[4824]: I1006 09:59:32.069223 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da385c16-418f-4c00-a25d-04fd6506a1c6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79b92b42226cd9bba974d119a8e7f3d1d822e8a21cf1e6cdfff48dcbe7a5a898\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://507540c1b777046fb6b90bb1de444f904c92fbd7b6e598217bb844b25495c683\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://507540c1b777046fb6b90bb1de444f904c92fbd7b6e598217bb844b25495c683\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:32Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:32 crc kubenswrapper[4824]: I1006 09:59:32.089971 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f9011f6673310dffea3acd7bdc2a56f2af91a87641631490fdc63da610f047a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:32Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:32 crc kubenswrapper[4824]: I1006 09:59:32.112770 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-szn8q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b65abf9f-4e88-4571-960d-3ca997d9c344\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:59:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:59:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c80f7c0d85c4c76aa3c4ff4e078900ccc3be12fa5bb75f595ead6eeacf39e3f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://664bf42e1324a9c9529f36f08d61e26999b3386bdd1fac77851fb7e7411e8934\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-06T09:59:19Z\\\",\\\"message\\\":\\\"2025-10-06T09:58:34+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_cffdfe89-8b06-4a09-9c9b-7ac6a5561d38\\\\n2025-10-06T09:58:34+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_cffdfe89-8b06-4a09-9c9b-7ac6a5561d38 to /host/opt/cni/bin/\\\\n2025-10-06T09:58:34Z [verbose] multus-daemon started\\\\n2025-10-06T09:58:34Z [verbose] Readiness Indicator file check\\\\n2025-10-06T09:59:19Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:59:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49s5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-szn8q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:32Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:32 crc kubenswrapper[4824]: I1006 09:59:32.136060 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:32 crc kubenswrapper[4824]: I1006 09:59:32.136139 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:32 crc kubenswrapper[4824]: I1006 09:59:32.136163 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:32 crc kubenswrapper[4824]: I1006 09:59:32.136196 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:32 crc kubenswrapper[4824]: I1006 09:59:32.136220 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:32Z","lastTransitionTime":"2025-10-06T09:59:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:32 crc kubenswrapper[4824]: I1006 09:59:32.148571 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f511ece-25eb-465f-b85c-9e5f3f886c21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10bfed3e523b7422ce1e981f98166eca78bd12dd4a7cf0634915e1f4ecc3de2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43199f86ad4a1f02bc57a926cd96b2c3c0e2ea46fb36a3f495668706be4c8041\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d6064ad3bcdafebb4c66e7d91f183cc787e9ce13660545c71a722b3ff04ce75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6213164d366243c2139d5971041d6f076cf38fd0cce4c0ca5495111963487ec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa6d59fac987f60a4340ab35d84aa059f2b7085a06362e7207a79702d17eee6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09638fbc9d163ad0b63fa57d24ea5298207ad1f55bab8671be311960be2627a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c88940bd9c33c0ee0feef797ad549f874cad42942352072dba2b4eaf1e8b803\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c88940bd9c33c0ee0feef797ad549f874cad42942352072dba2b4eaf1e8b803\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-06T09:59:30Z\\\",\\\"message\\\":\\\"flector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1006 09:59:30.228084 6864 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1006 09:59:30.228172 6864 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1006 09:59:30.228389 6864 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1006 09:59:30.228646 6864 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1006 09:59:30.228673 6864 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1006 09:59:30.228695 6864 handler.go:208] Removed *v1.Node event handler 2\\\\nI1006 09:59:30.228716 6864 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1006 09:59:30.228730 6864 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1006 09:59:30.228754 6864 factory.go:656] Stopping watch factory\\\\nI1006 09:59:30.228781 6864 ovnkube.go:599] Stopped ovnkube\\\\nI1006 09:59:30.228783 6864 handler.go:208] Removed *v1.Node event handler 7\\\\nI1006 09:59:30.228805 6864 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1006 09\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T09:59:29Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-4vwms_openshift-ovn-kubernetes(2f511ece-25eb-465f-b85c-9e5f3f886c21)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0644c7db28c7d26aa8d69cd5b9fbcff4d4c06b96df113137dea22a72d9fecd85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6da4c07be4d507a6f20bd8b42be6e6ea4712a956f57ef164036cfded7142a87e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6da4c07be4d507a6f20bd8b42be6e6ea4712a956f57ef164036cfded7142a87e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mc4k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4vwms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:32Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:32 crc kubenswrapper[4824]: I1006 09:59:32.168594 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lszvg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0cf70ab4-4cfe-41c5-8db7-035451bafcfa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7b80b5b375b3b32132501ee0d661e06e199654c696ac76ad56e7ba96918cc77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fkbf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://519653c6b241cd89ef3ece003667d420a7a55cdf513f4577a69e14ade5fb4c76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fkbf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lszvg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:32Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:32 crc kubenswrapper[4824]: I1006 09:59:32.187757 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gnw94" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7cf1d355-64cb-48a6-acfd-7d258b7afa10\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-psbq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-psbq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:46Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gnw94\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:32Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:32 crc kubenswrapper[4824]: I1006 09:59:32.211654 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b665cda3-da4e-482e-8248-023abec382f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9235c742d8c45bf667f32979c2891fece50f4ee1bd0c7cf6a753ed2d2b9ab30e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b2a920dbc506df2426232990e31fbb60d255c00be54941826a0f094af4a92f0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d516f278f362d7c45e7149de82fa835997ac198f1165c04352910d51fb674d69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c820d8bea16cff0d463ce68d2f2acb72d7a7c73372757ff104e14fcec1025a05\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b8cdf54407cc95783919eba8e92f093bbc09eac5e908c4ba0685629948c71bd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-06T09:58:27Z\\\",\\\"message\\\":\\\"arting RequestHeaderAuthRequestController\\\\nI1006 09:58:27.156944 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1006 09:58:27.156783 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1006 09:58:27.157213 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-432254888/tls.crt::/tmp/serving-cert-432254888/tls.key\\\\\\\"\\\\nI1006 09:58:27.157407 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1006 09:58:27.157443 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1006 09:58:27.157410 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-432254888/tls.crt::/tmp/serving-cert-432254888/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759744705\\\\\\\\\\\\\\\" (2025-10-06 09:58:25 +0000 UTC to 2025-11-05 09:58:26 +0000 UTC (now=2025-10-06 09:58:27.157390076 +0000 UTC))\\\\\\\"\\\\nI1006 09:58:27.157632 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1006 09:58:27.158158 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759744707\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759744707\\\\\\\\\\\\\\\" (2025-10-06 08:58:26 +0000 UTC to 2026-10-06 08:58:26 +0000 UTC (now=2025-10-06 09:58:27.158090663 +0000 UTC))\\\\\\\"\\\\nI1006 09:58:27.158220 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1006 09:58:27.158270 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1006 09:58:27.158310 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nF1006 09:58:27.158617 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://413822c99ba2866521177cce4f1894aef17c160aebb2662da91824a04827d510\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:14Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e37e8c7400f7fd14a5c5390a5985e3d8d277e9c5ca60d5e9e9757d07d1778570\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e37e8c7400f7fd14a5c5390a5985e3d8d277e9c5ca60d5e9e9757d07d1778570\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:11Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:32Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:32 crc kubenswrapper[4824]: I1006 09:59:32.234782 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d739433-0a38-42fe-b884-b8433011967e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://464e279c5903770c34d80f31ed231b194d5628b39055c4058908039d10690059\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://efa1b62728466ef7aac0e745230348ce920eb9ec9b07fb2cd5a276150dac13fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fedc40f0217626c56d155fc7e78e0592556fef9464db2aaf09c5facc486aad8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78fac300cbe2ac52f5788306b8fc0011f6ba3a22c3546974e70998f2d394987a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:11Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:32Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:32 crc kubenswrapper[4824]: I1006 09:59:32.239856 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:32 crc kubenswrapper[4824]: I1006 09:59:32.239918 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:32 crc kubenswrapper[4824]: I1006 09:59:32.239937 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:32 crc kubenswrapper[4824]: I1006 09:59:32.239968 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:32 crc kubenswrapper[4824]: I1006 09:59:32.240067 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:32Z","lastTransitionTime":"2025-10-06T09:59:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:32 crc kubenswrapper[4824]: I1006 09:59:32.261359 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:30Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:32Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:32 crc kubenswrapper[4824]: I1006 09:59:32.273957 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 09:59:32 crc kubenswrapper[4824]: I1006 09:59:32.274031 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 09:59:32 crc kubenswrapper[4824]: E1006 09:59:32.274155 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 06 09:59:32 crc kubenswrapper[4824]: I1006 09:59:32.274206 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 09:59:32 crc kubenswrapper[4824]: I1006 09:59:32.274206 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gnw94" Oct 06 09:59:32 crc kubenswrapper[4824]: E1006 09:59:32.274460 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 06 09:59:32 crc kubenswrapper[4824]: E1006 09:59:32.274543 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 06 09:59:32 crc kubenswrapper[4824]: E1006 09:59:32.274673 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gnw94" podUID="7cf1d355-64cb-48a6-acfd-7d258b7afa10" Oct 06 09:59:32 crc kubenswrapper[4824]: I1006 09:59:32.282502 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b67b743d-7d7e-4a7f-b648-c721f50bb3a9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:59:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:59:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b5bb30b2f9b1a35bf1dfd2274c7e3d280757902687d9ae79b710fbd5f07f398\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0658bdad936b07337e1eb58a4d62c1d7fc98fb6b3b115d9be4c1d5d79ecbc8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://268983a8eccc2af9242e8f5ebea3722e30cd66086272ec6f681ff96ed9a0e7fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6f4a5101060399a25f1d98b0495745ab0a679b88cc6bee5ea1bf0caff7b6767\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6f4a5101060399a25f1d98b0495745ab0a679b88cc6bee5ea1bf0caff7b6767\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:12Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:11Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:32Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:32 crc kubenswrapper[4824]: I1006 09:59:32.307955 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-7c22q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7eaaa68-d971-4a1c-b6c5-551ded8ff27f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T09:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f5150c8c6cfc65ecdcd0e7bd52199dd60e2472c96874f103b49919c81b5e5f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T09:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53336c7d3c509b4e022708998aedc5dee9fdaf7136ae3ee69b4181394c3a22fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53336c7d3c509b4e022708998aedc5dee9fdaf7136ae3ee69b4181394c3a22fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://652ab0a13a2bb421872edda3f7188dce8e296247052550c937f8447fea206756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://652ab0a13a2bb421872edda3f7188dce8e296247052550c937f8447fea206756\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3514dec0386bb8c016a7f555d7191ea420d5058b0ee98f97d965dd857c5376e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3514dec0386bb8c016a7f555d7191ea420d5058b0ee98f97d965dd857c5376e6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b035e0bf34e3429400ac8942124a3e9834231899487d03d15609fe5624747a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b035e0bf34e3429400ac8942124a3e9834231899487d03d15609fe5624747a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6743e027a11864ad3f433d110e60451fb36d49791e03b22d437b70f89d056ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6743e027a11864ad3f433d110e60451fb36d49791e03b22d437b70f89d056ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8ea48b6886fd39a1c5e5b29ca1fffc2bd4552d61dc621eb2fba66fb464cf120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8ea48b6886fd39a1c5e5b29ca1fffc2bd4552d61dc621eb2fba66fb464cf120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T09:58:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T09:58:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z7pmj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T09:58:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-7c22q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T09:59:32Z is after 2025-08-24T17:21:41Z" Oct 06 09:59:32 crc kubenswrapper[4824]: I1006 09:59:32.343204 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:32 crc kubenswrapper[4824]: I1006 09:59:32.343246 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:32 crc kubenswrapper[4824]: I1006 09:59:32.343258 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:32 crc kubenswrapper[4824]: I1006 09:59:32.343280 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:32 crc kubenswrapper[4824]: I1006 09:59:32.343294 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:32Z","lastTransitionTime":"2025-10-06T09:59:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:32 crc kubenswrapper[4824]: I1006 09:59:32.446635 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:32 crc kubenswrapper[4824]: I1006 09:59:32.446694 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:32 crc kubenswrapper[4824]: I1006 09:59:32.446712 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:32 crc kubenswrapper[4824]: I1006 09:59:32.446740 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:32 crc kubenswrapper[4824]: I1006 09:59:32.446760 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:32Z","lastTransitionTime":"2025-10-06T09:59:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:32 crc kubenswrapper[4824]: I1006 09:59:32.550593 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:32 crc kubenswrapper[4824]: I1006 09:59:32.550647 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:32 crc kubenswrapper[4824]: I1006 09:59:32.550667 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:32 crc kubenswrapper[4824]: I1006 09:59:32.550693 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:32 crc kubenswrapper[4824]: I1006 09:59:32.550710 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:32Z","lastTransitionTime":"2025-10-06T09:59:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:32 crc kubenswrapper[4824]: I1006 09:59:32.654257 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:32 crc kubenswrapper[4824]: I1006 09:59:32.654326 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:32 crc kubenswrapper[4824]: I1006 09:59:32.654354 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:32 crc kubenswrapper[4824]: I1006 09:59:32.654389 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:32 crc kubenswrapper[4824]: I1006 09:59:32.654418 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:32Z","lastTransitionTime":"2025-10-06T09:59:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:32 crc kubenswrapper[4824]: I1006 09:59:32.758091 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:32 crc kubenswrapper[4824]: I1006 09:59:32.758188 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:32 crc kubenswrapper[4824]: I1006 09:59:32.758210 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:32 crc kubenswrapper[4824]: I1006 09:59:32.758237 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:32 crc kubenswrapper[4824]: I1006 09:59:32.758255 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:32Z","lastTransitionTime":"2025-10-06T09:59:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:32 crc kubenswrapper[4824]: I1006 09:59:32.862238 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:32 crc kubenswrapper[4824]: I1006 09:59:32.862310 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:32 crc kubenswrapper[4824]: I1006 09:59:32.862334 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:32 crc kubenswrapper[4824]: I1006 09:59:32.862365 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:32 crc kubenswrapper[4824]: I1006 09:59:32.862396 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:32Z","lastTransitionTime":"2025-10-06T09:59:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:32 crc kubenswrapper[4824]: I1006 09:59:32.966327 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:32 crc kubenswrapper[4824]: I1006 09:59:32.966396 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:32 crc kubenswrapper[4824]: I1006 09:59:32.966415 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:32 crc kubenswrapper[4824]: I1006 09:59:32.966444 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:32 crc kubenswrapper[4824]: I1006 09:59:32.966464 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:32Z","lastTransitionTime":"2025-10-06T09:59:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:33 crc kubenswrapper[4824]: I1006 09:59:33.069495 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:33 crc kubenswrapper[4824]: I1006 09:59:33.069591 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:33 crc kubenswrapper[4824]: I1006 09:59:33.069614 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:33 crc kubenswrapper[4824]: I1006 09:59:33.069641 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:33 crc kubenswrapper[4824]: I1006 09:59:33.069658 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:33Z","lastTransitionTime":"2025-10-06T09:59:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:33 crc kubenswrapper[4824]: I1006 09:59:33.172778 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:33 crc kubenswrapper[4824]: I1006 09:59:33.172823 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:33 crc kubenswrapper[4824]: I1006 09:59:33.172836 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:33 crc kubenswrapper[4824]: I1006 09:59:33.172857 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:33 crc kubenswrapper[4824]: I1006 09:59:33.172874 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:33Z","lastTransitionTime":"2025-10-06T09:59:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:33 crc kubenswrapper[4824]: I1006 09:59:33.275609 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:33 crc kubenswrapper[4824]: I1006 09:59:33.275696 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:33 crc kubenswrapper[4824]: I1006 09:59:33.275716 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:33 crc kubenswrapper[4824]: I1006 09:59:33.275744 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:33 crc kubenswrapper[4824]: I1006 09:59:33.275763 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:33Z","lastTransitionTime":"2025-10-06T09:59:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:33 crc kubenswrapper[4824]: I1006 09:59:33.379092 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:33 crc kubenswrapper[4824]: I1006 09:59:33.379180 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:33 crc kubenswrapper[4824]: I1006 09:59:33.379207 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:33 crc kubenswrapper[4824]: I1006 09:59:33.379240 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:33 crc kubenswrapper[4824]: I1006 09:59:33.379259 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:33Z","lastTransitionTime":"2025-10-06T09:59:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:33 crc kubenswrapper[4824]: I1006 09:59:33.483156 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:33 crc kubenswrapper[4824]: I1006 09:59:33.483234 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:33 crc kubenswrapper[4824]: I1006 09:59:33.483252 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:33 crc kubenswrapper[4824]: I1006 09:59:33.483281 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:33 crc kubenswrapper[4824]: I1006 09:59:33.483300 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:33Z","lastTransitionTime":"2025-10-06T09:59:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:33 crc kubenswrapper[4824]: I1006 09:59:33.588081 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:33 crc kubenswrapper[4824]: I1006 09:59:33.588162 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:33 crc kubenswrapper[4824]: I1006 09:59:33.588180 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:33 crc kubenswrapper[4824]: I1006 09:59:33.588209 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:33 crc kubenswrapper[4824]: I1006 09:59:33.588227 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:33Z","lastTransitionTime":"2025-10-06T09:59:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:33 crc kubenswrapper[4824]: I1006 09:59:33.692547 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:33 crc kubenswrapper[4824]: I1006 09:59:33.692622 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:33 crc kubenswrapper[4824]: I1006 09:59:33.692640 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:33 crc kubenswrapper[4824]: I1006 09:59:33.692670 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:33 crc kubenswrapper[4824]: I1006 09:59:33.692693 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:33Z","lastTransitionTime":"2025-10-06T09:59:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:33 crc kubenswrapper[4824]: I1006 09:59:33.796940 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:33 crc kubenswrapper[4824]: I1006 09:59:33.797023 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:33 crc kubenswrapper[4824]: I1006 09:59:33.797034 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:33 crc kubenswrapper[4824]: I1006 09:59:33.797054 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:33 crc kubenswrapper[4824]: I1006 09:59:33.797068 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:33Z","lastTransitionTime":"2025-10-06T09:59:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:33 crc kubenswrapper[4824]: I1006 09:59:33.900581 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:33 crc kubenswrapper[4824]: I1006 09:59:33.900627 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:33 crc kubenswrapper[4824]: I1006 09:59:33.900639 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:33 crc kubenswrapper[4824]: I1006 09:59:33.900661 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:33 crc kubenswrapper[4824]: I1006 09:59:33.900675 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:33Z","lastTransitionTime":"2025-10-06T09:59:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:34 crc kubenswrapper[4824]: I1006 09:59:34.006639 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:34 crc kubenswrapper[4824]: I1006 09:59:34.006702 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:34 crc kubenswrapper[4824]: I1006 09:59:34.006721 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:34 crc kubenswrapper[4824]: I1006 09:59:34.006747 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:34 crc kubenswrapper[4824]: I1006 09:59:34.006833 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:34Z","lastTransitionTime":"2025-10-06T09:59:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:34 crc kubenswrapper[4824]: I1006 09:59:34.110259 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:34 crc kubenswrapper[4824]: I1006 09:59:34.110325 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:34 crc kubenswrapper[4824]: I1006 09:59:34.110343 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:34 crc kubenswrapper[4824]: I1006 09:59:34.110370 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:34 crc kubenswrapper[4824]: I1006 09:59:34.110391 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:34Z","lastTransitionTime":"2025-10-06T09:59:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:34 crc kubenswrapper[4824]: I1006 09:59:34.213766 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:34 crc kubenswrapper[4824]: I1006 09:59:34.213840 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:34 crc kubenswrapper[4824]: I1006 09:59:34.213859 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:34 crc kubenswrapper[4824]: I1006 09:59:34.213888 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:34 crc kubenswrapper[4824]: I1006 09:59:34.213909 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:34Z","lastTransitionTime":"2025-10-06T09:59:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:34 crc kubenswrapper[4824]: I1006 09:59:34.214118 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 09:59:34 crc kubenswrapper[4824]: I1006 09:59:34.214311 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 09:59:34 crc kubenswrapper[4824]: E1006 09:59:34.214445 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-06 10:00:38.214407364 +0000 UTC m=+147.578830255 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 09:59:34 crc kubenswrapper[4824]: E1006 09:59:34.214521 4824 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 06 09:59:34 crc kubenswrapper[4824]: I1006 09:59:34.214540 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 09:59:34 crc kubenswrapper[4824]: E1006 09:59:34.214607 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-06 10:00:38.214579558 +0000 UTC m=+147.579002449 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 06 09:59:34 crc kubenswrapper[4824]: E1006 09:59:34.214683 4824 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 06 09:59:34 crc kubenswrapper[4824]: E1006 09:59:34.214768 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-06 10:00:38.214741552 +0000 UTC m=+147.579164613 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 06 09:59:34 crc kubenswrapper[4824]: I1006 09:59:34.274022 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 09:59:34 crc kubenswrapper[4824]: I1006 09:59:34.274065 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gnw94" Oct 06 09:59:34 crc kubenswrapper[4824]: I1006 09:59:34.274241 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 09:59:34 crc kubenswrapper[4824]: E1006 09:59:34.274231 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 06 09:59:34 crc kubenswrapper[4824]: I1006 09:59:34.274409 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 09:59:34 crc kubenswrapper[4824]: E1006 09:59:34.274652 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 06 09:59:34 crc kubenswrapper[4824]: E1006 09:59:34.274797 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 06 09:59:34 crc kubenswrapper[4824]: E1006 09:59:34.274868 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gnw94" podUID="7cf1d355-64cb-48a6-acfd-7d258b7afa10" Oct 06 09:59:34 crc kubenswrapper[4824]: I1006 09:59:34.315970 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 09:59:34 crc kubenswrapper[4824]: I1006 09:59:34.316097 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 09:59:34 crc kubenswrapper[4824]: E1006 09:59:34.316324 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 06 09:59:34 crc kubenswrapper[4824]: E1006 09:59:34.316373 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 06 09:59:34 crc kubenswrapper[4824]: E1006 09:59:34.316394 4824 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 06 09:59:34 crc kubenswrapper[4824]: E1006 09:59:34.316401 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 06 09:59:34 crc kubenswrapper[4824]: E1006 09:59:34.316435 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 06 09:59:34 crc kubenswrapper[4824]: E1006 09:59:34.316457 4824 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 06 09:59:34 crc kubenswrapper[4824]: E1006 09:59:34.316496 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-06 10:00:38.31646429 +0000 UTC m=+147.680887181 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 06 09:59:34 crc kubenswrapper[4824]: E1006 09:59:34.316539 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-06 10:00:38.316512471 +0000 UTC m=+147.680935372 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 06 09:59:34 crc kubenswrapper[4824]: I1006 09:59:34.318055 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:34 crc kubenswrapper[4824]: I1006 09:59:34.318155 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:34 crc kubenswrapper[4824]: I1006 09:59:34.318175 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:34 crc kubenswrapper[4824]: I1006 09:59:34.318200 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:34 crc kubenswrapper[4824]: I1006 09:59:34.318219 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:34Z","lastTransitionTime":"2025-10-06T09:59:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:34 crc kubenswrapper[4824]: I1006 09:59:34.421187 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:34 crc kubenswrapper[4824]: I1006 09:59:34.421247 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:34 crc kubenswrapper[4824]: I1006 09:59:34.421265 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:34 crc kubenswrapper[4824]: I1006 09:59:34.421290 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:34 crc kubenswrapper[4824]: I1006 09:59:34.421307 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:34Z","lastTransitionTime":"2025-10-06T09:59:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:34 crc kubenswrapper[4824]: I1006 09:59:34.526146 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:34 crc kubenswrapper[4824]: I1006 09:59:34.526234 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:34 crc kubenswrapper[4824]: I1006 09:59:34.526252 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:34 crc kubenswrapper[4824]: I1006 09:59:34.526278 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:34 crc kubenswrapper[4824]: I1006 09:59:34.526297 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:34Z","lastTransitionTime":"2025-10-06T09:59:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:34 crc kubenswrapper[4824]: I1006 09:59:34.629945 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:34 crc kubenswrapper[4824]: I1006 09:59:34.630076 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:34 crc kubenswrapper[4824]: I1006 09:59:34.630137 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:34 crc kubenswrapper[4824]: I1006 09:59:34.630160 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:34 crc kubenswrapper[4824]: I1006 09:59:34.630178 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:34Z","lastTransitionTime":"2025-10-06T09:59:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:34 crc kubenswrapper[4824]: I1006 09:59:34.733542 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:34 crc kubenswrapper[4824]: I1006 09:59:34.733606 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:34 crc kubenswrapper[4824]: I1006 09:59:34.733625 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:34 crc kubenswrapper[4824]: I1006 09:59:34.733654 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:34 crc kubenswrapper[4824]: I1006 09:59:34.733674 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:34Z","lastTransitionTime":"2025-10-06T09:59:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:34 crc kubenswrapper[4824]: I1006 09:59:34.837366 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:34 crc kubenswrapper[4824]: I1006 09:59:34.837443 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:34 crc kubenswrapper[4824]: I1006 09:59:34.837463 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:34 crc kubenswrapper[4824]: I1006 09:59:34.837492 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:34 crc kubenswrapper[4824]: I1006 09:59:34.837510 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:34Z","lastTransitionTime":"2025-10-06T09:59:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:34 crc kubenswrapper[4824]: I1006 09:59:34.940492 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:34 crc kubenswrapper[4824]: I1006 09:59:34.940566 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:34 crc kubenswrapper[4824]: I1006 09:59:34.940586 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:34 crc kubenswrapper[4824]: I1006 09:59:34.940617 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:34 crc kubenswrapper[4824]: I1006 09:59:34.940640 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:34Z","lastTransitionTime":"2025-10-06T09:59:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:35 crc kubenswrapper[4824]: I1006 09:59:35.043857 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:35 crc kubenswrapper[4824]: I1006 09:59:35.043917 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:35 crc kubenswrapper[4824]: I1006 09:59:35.043935 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:35 crc kubenswrapper[4824]: I1006 09:59:35.044015 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:35 crc kubenswrapper[4824]: I1006 09:59:35.044040 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:35Z","lastTransitionTime":"2025-10-06T09:59:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:35 crc kubenswrapper[4824]: I1006 09:59:35.147179 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:35 crc kubenswrapper[4824]: I1006 09:59:35.147248 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:35 crc kubenswrapper[4824]: I1006 09:59:35.147267 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:35 crc kubenswrapper[4824]: I1006 09:59:35.147298 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:35 crc kubenswrapper[4824]: I1006 09:59:35.147320 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:35Z","lastTransitionTime":"2025-10-06T09:59:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:35 crc kubenswrapper[4824]: I1006 09:59:35.249708 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:35 crc kubenswrapper[4824]: I1006 09:59:35.249793 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:35 crc kubenswrapper[4824]: I1006 09:59:35.249808 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:35 crc kubenswrapper[4824]: I1006 09:59:35.249830 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:35 crc kubenswrapper[4824]: I1006 09:59:35.249850 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:35Z","lastTransitionTime":"2025-10-06T09:59:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:35 crc kubenswrapper[4824]: I1006 09:59:35.353297 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:35 crc kubenswrapper[4824]: I1006 09:59:35.353364 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:35 crc kubenswrapper[4824]: I1006 09:59:35.353382 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:35 crc kubenswrapper[4824]: I1006 09:59:35.353407 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:35 crc kubenswrapper[4824]: I1006 09:59:35.353429 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:35Z","lastTransitionTime":"2025-10-06T09:59:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:35 crc kubenswrapper[4824]: I1006 09:59:35.456413 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:35 crc kubenswrapper[4824]: I1006 09:59:35.456477 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:35 crc kubenswrapper[4824]: I1006 09:59:35.456497 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:35 crc kubenswrapper[4824]: I1006 09:59:35.456523 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:35 crc kubenswrapper[4824]: I1006 09:59:35.456546 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:35Z","lastTransitionTime":"2025-10-06T09:59:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:35 crc kubenswrapper[4824]: I1006 09:59:35.559673 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:35 crc kubenswrapper[4824]: I1006 09:59:35.559736 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:35 crc kubenswrapper[4824]: I1006 09:59:35.559756 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:35 crc kubenswrapper[4824]: I1006 09:59:35.559783 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:35 crc kubenswrapper[4824]: I1006 09:59:35.559804 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:35Z","lastTransitionTime":"2025-10-06T09:59:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:35 crc kubenswrapper[4824]: I1006 09:59:35.663151 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:35 crc kubenswrapper[4824]: I1006 09:59:35.663509 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:35 crc kubenswrapper[4824]: I1006 09:59:35.663527 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:35 crc kubenswrapper[4824]: I1006 09:59:35.663558 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:35 crc kubenswrapper[4824]: I1006 09:59:35.663578 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:35Z","lastTransitionTime":"2025-10-06T09:59:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:35 crc kubenswrapper[4824]: I1006 09:59:35.766673 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:35 crc kubenswrapper[4824]: I1006 09:59:35.766746 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:35 crc kubenswrapper[4824]: I1006 09:59:35.766764 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:35 crc kubenswrapper[4824]: I1006 09:59:35.766800 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:35 crc kubenswrapper[4824]: I1006 09:59:35.766823 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:35Z","lastTransitionTime":"2025-10-06T09:59:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:35 crc kubenswrapper[4824]: I1006 09:59:35.869923 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:35 crc kubenswrapper[4824]: I1006 09:59:35.870024 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:35 crc kubenswrapper[4824]: I1006 09:59:35.870050 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:35 crc kubenswrapper[4824]: I1006 09:59:35.870087 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:35 crc kubenswrapper[4824]: I1006 09:59:35.870116 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:35Z","lastTransitionTime":"2025-10-06T09:59:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:35 crc kubenswrapper[4824]: I1006 09:59:35.973718 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:35 crc kubenswrapper[4824]: I1006 09:59:35.973793 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:35 crc kubenswrapper[4824]: I1006 09:59:35.973820 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:35 crc kubenswrapper[4824]: I1006 09:59:35.973851 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:35 crc kubenswrapper[4824]: I1006 09:59:35.973871 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:35Z","lastTransitionTime":"2025-10-06T09:59:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:36 crc kubenswrapper[4824]: I1006 09:59:36.077381 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:36 crc kubenswrapper[4824]: I1006 09:59:36.077455 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:36 crc kubenswrapper[4824]: I1006 09:59:36.077475 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:36 crc kubenswrapper[4824]: I1006 09:59:36.077505 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:36 crc kubenswrapper[4824]: I1006 09:59:36.077529 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:36Z","lastTransitionTime":"2025-10-06T09:59:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:36 crc kubenswrapper[4824]: I1006 09:59:36.181098 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:36 crc kubenswrapper[4824]: I1006 09:59:36.181167 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:36 crc kubenswrapper[4824]: I1006 09:59:36.181185 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:36 crc kubenswrapper[4824]: I1006 09:59:36.181213 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:36 crc kubenswrapper[4824]: I1006 09:59:36.181236 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:36Z","lastTransitionTime":"2025-10-06T09:59:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:36 crc kubenswrapper[4824]: I1006 09:59:36.273482 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gnw94" Oct 06 09:59:36 crc kubenswrapper[4824]: I1006 09:59:36.273552 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 09:59:36 crc kubenswrapper[4824]: I1006 09:59:36.273600 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 09:59:36 crc kubenswrapper[4824]: E1006 09:59:36.273776 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gnw94" podUID="7cf1d355-64cb-48a6-acfd-7d258b7afa10" Oct 06 09:59:36 crc kubenswrapper[4824]: I1006 09:59:36.273803 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 09:59:36 crc kubenswrapper[4824]: E1006 09:59:36.273973 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 06 09:59:36 crc kubenswrapper[4824]: E1006 09:59:36.274216 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 06 09:59:36 crc kubenswrapper[4824]: E1006 09:59:36.274333 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 06 09:59:36 crc kubenswrapper[4824]: I1006 09:59:36.284262 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:36 crc kubenswrapper[4824]: I1006 09:59:36.284306 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:36 crc kubenswrapper[4824]: I1006 09:59:36.284326 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:36 crc kubenswrapper[4824]: I1006 09:59:36.284349 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:36 crc kubenswrapper[4824]: I1006 09:59:36.284368 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:36Z","lastTransitionTime":"2025-10-06T09:59:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:36 crc kubenswrapper[4824]: I1006 09:59:36.387340 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:36 crc kubenswrapper[4824]: I1006 09:59:36.387424 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:36 crc kubenswrapper[4824]: I1006 09:59:36.387449 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:36 crc kubenswrapper[4824]: I1006 09:59:36.387485 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:36 crc kubenswrapper[4824]: I1006 09:59:36.387509 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:36Z","lastTransitionTime":"2025-10-06T09:59:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:36 crc kubenswrapper[4824]: I1006 09:59:36.491013 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:36 crc kubenswrapper[4824]: I1006 09:59:36.491082 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:36 crc kubenswrapper[4824]: I1006 09:59:36.491101 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:36 crc kubenswrapper[4824]: I1006 09:59:36.491128 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:36 crc kubenswrapper[4824]: I1006 09:59:36.491148 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:36Z","lastTransitionTime":"2025-10-06T09:59:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:36 crc kubenswrapper[4824]: I1006 09:59:36.594927 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:36 crc kubenswrapper[4824]: I1006 09:59:36.595032 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:36 crc kubenswrapper[4824]: I1006 09:59:36.595054 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:36 crc kubenswrapper[4824]: I1006 09:59:36.595083 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:36 crc kubenswrapper[4824]: I1006 09:59:36.595101 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:36Z","lastTransitionTime":"2025-10-06T09:59:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:36 crc kubenswrapper[4824]: I1006 09:59:36.698585 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:36 crc kubenswrapper[4824]: I1006 09:59:36.698652 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:36 crc kubenswrapper[4824]: I1006 09:59:36.698670 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:36 crc kubenswrapper[4824]: I1006 09:59:36.698699 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:36 crc kubenswrapper[4824]: I1006 09:59:36.698724 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:36Z","lastTransitionTime":"2025-10-06T09:59:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:36 crc kubenswrapper[4824]: I1006 09:59:36.802376 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:36 crc kubenswrapper[4824]: I1006 09:59:36.802437 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:36 crc kubenswrapper[4824]: I1006 09:59:36.802455 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:36 crc kubenswrapper[4824]: I1006 09:59:36.802480 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:36 crc kubenswrapper[4824]: I1006 09:59:36.802498 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:36Z","lastTransitionTime":"2025-10-06T09:59:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:36 crc kubenswrapper[4824]: I1006 09:59:36.905807 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:36 crc kubenswrapper[4824]: I1006 09:59:36.905894 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:36 crc kubenswrapper[4824]: I1006 09:59:36.905918 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:36 crc kubenswrapper[4824]: I1006 09:59:36.905946 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:36 crc kubenswrapper[4824]: I1006 09:59:36.905967 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:36Z","lastTransitionTime":"2025-10-06T09:59:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:37 crc kubenswrapper[4824]: I1006 09:59:37.009585 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:37 crc kubenswrapper[4824]: I1006 09:59:37.009662 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:37 crc kubenswrapper[4824]: I1006 09:59:37.009682 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:37 crc kubenswrapper[4824]: I1006 09:59:37.009711 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:37 crc kubenswrapper[4824]: I1006 09:59:37.009731 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:37Z","lastTransitionTime":"2025-10-06T09:59:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:37 crc kubenswrapper[4824]: I1006 09:59:37.113540 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:37 crc kubenswrapper[4824]: I1006 09:59:37.113611 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:37 crc kubenswrapper[4824]: I1006 09:59:37.113633 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:37 crc kubenswrapper[4824]: I1006 09:59:37.113661 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:37 crc kubenswrapper[4824]: I1006 09:59:37.113682 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:37Z","lastTransitionTime":"2025-10-06T09:59:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:37 crc kubenswrapper[4824]: I1006 09:59:37.217250 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:37 crc kubenswrapper[4824]: I1006 09:59:37.217323 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:37 crc kubenswrapper[4824]: I1006 09:59:37.217342 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:37 crc kubenswrapper[4824]: I1006 09:59:37.217373 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:37 crc kubenswrapper[4824]: I1006 09:59:37.217394 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:37Z","lastTransitionTime":"2025-10-06T09:59:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:37 crc kubenswrapper[4824]: I1006 09:59:37.320806 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:37 crc kubenswrapper[4824]: I1006 09:59:37.320878 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:37 crc kubenswrapper[4824]: I1006 09:59:37.320897 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:37 crc kubenswrapper[4824]: I1006 09:59:37.320929 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:37 crc kubenswrapper[4824]: I1006 09:59:37.320949 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:37Z","lastTransitionTime":"2025-10-06T09:59:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:37 crc kubenswrapper[4824]: I1006 09:59:37.424489 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:37 crc kubenswrapper[4824]: I1006 09:59:37.424560 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:37 crc kubenswrapper[4824]: I1006 09:59:37.424583 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:37 crc kubenswrapper[4824]: I1006 09:59:37.424611 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:37 crc kubenswrapper[4824]: I1006 09:59:37.424630 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:37Z","lastTransitionTime":"2025-10-06T09:59:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:37 crc kubenswrapper[4824]: I1006 09:59:37.527417 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:37 crc kubenswrapper[4824]: I1006 09:59:37.527488 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:37 crc kubenswrapper[4824]: I1006 09:59:37.527507 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:37 crc kubenswrapper[4824]: I1006 09:59:37.527542 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:37 crc kubenswrapper[4824]: I1006 09:59:37.527562 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:37Z","lastTransitionTime":"2025-10-06T09:59:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:37 crc kubenswrapper[4824]: I1006 09:59:37.631056 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:37 crc kubenswrapper[4824]: I1006 09:59:37.631123 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:37 crc kubenswrapper[4824]: I1006 09:59:37.631141 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:37 crc kubenswrapper[4824]: I1006 09:59:37.631170 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:37 crc kubenswrapper[4824]: I1006 09:59:37.631189 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:37Z","lastTransitionTime":"2025-10-06T09:59:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:37 crc kubenswrapper[4824]: I1006 09:59:37.734315 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:37 crc kubenswrapper[4824]: I1006 09:59:37.734382 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:37 crc kubenswrapper[4824]: I1006 09:59:37.734401 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:37 crc kubenswrapper[4824]: I1006 09:59:37.734428 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:37 crc kubenswrapper[4824]: I1006 09:59:37.734450 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:37Z","lastTransitionTime":"2025-10-06T09:59:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:37 crc kubenswrapper[4824]: I1006 09:59:37.837854 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:37 crc kubenswrapper[4824]: I1006 09:59:37.837906 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:37 crc kubenswrapper[4824]: I1006 09:59:37.837924 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:37 crc kubenswrapper[4824]: I1006 09:59:37.837950 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:37 crc kubenswrapper[4824]: I1006 09:59:37.837973 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:37Z","lastTransitionTime":"2025-10-06T09:59:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:37 crc kubenswrapper[4824]: I1006 09:59:37.940731 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:37 crc kubenswrapper[4824]: I1006 09:59:37.941155 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:37 crc kubenswrapper[4824]: I1006 09:59:37.941312 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:37 crc kubenswrapper[4824]: I1006 09:59:37.941461 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:37 crc kubenswrapper[4824]: I1006 09:59:37.941593 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:37Z","lastTransitionTime":"2025-10-06T09:59:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:38 crc kubenswrapper[4824]: I1006 09:59:38.045256 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:38 crc kubenswrapper[4824]: I1006 09:59:38.045318 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:38 crc kubenswrapper[4824]: I1006 09:59:38.045341 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:38 crc kubenswrapper[4824]: I1006 09:59:38.045370 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:38 crc kubenswrapper[4824]: I1006 09:59:38.045389 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:38Z","lastTransitionTime":"2025-10-06T09:59:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:38 crc kubenswrapper[4824]: I1006 09:59:38.149180 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:38 crc kubenswrapper[4824]: I1006 09:59:38.149241 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:38 crc kubenswrapper[4824]: I1006 09:59:38.149260 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:38 crc kubenswrapper[4824]: I1006 09:59:38.149288 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:38 crc kubenswrapper[4824]: I1006 09:59:38.149310 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:38Z","lastTransitionTime":"2025-10-06T09:59:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:38 crc kubenswrapper[4824]: I1006 09:59:38.249269 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 09:59:38 crc kubenswrapper[4824]: I1006 09:59:38.249335 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 09:59:38 crc kubenswrapper[4824]: I1006 09:59:38.249353 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 09:59:38 crc kubenswrapper[4824]: I1006 09:59:38.249379 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 09:59:38 crc kubenswrapper[4824]: I1006 09:59:38.249397 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T09:59:38Z","lastTransitionTime":"2025-10-06T09:59:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 09:59:38 crc kubenswrapper[4824]: I1006 09:59:38.273249 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 09:59:38 crc kubenswrapper[4824]: I1006 09:59:38.273317 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 09:59:38 crc kubenswrapper[4824]: I1006 09:59:38.273366 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gnw94" Oct 06 09:59:38 crc kubenswrapper[4824]: E1006 09:59:38.273423 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 06 09:59:38 crc kubenswrapper[4824]: I1006 09:59:38.273435 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 09:59:38 crc kubenswrapper[4824]: E1006 09:59:38.273594 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gnw94" podUID="7cf1d355-64cb-48a6-acfd-7d258b7afa10" Oct 06 09:59:38 crc kubenswrapper[4824]: E1006 09:59:38.273730 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 06 09:59:38 crc kubenswrapper[4824]: E1006 09:59:38.273836 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 06 09:59:38 crc kubenswrapper[4824]: I1006 09:59:38.695205 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-sn4f7"] Oct 06 09:59:38 crc kubenswrapper[4824]: I1006 09:59:38.696143 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-sn4f7" Oct 06 09:59:38 crc kubenswrapper[4824]: I1006 09:59:38.699798 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Oct 06 09:59:38 crc kubenswrapper[4824]: I1006 09:59:38.700611 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Oct 06 09:59:38 crc kubenswrapper[4824]: I1006 09:59:38.701358 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Oct 06 09:59:38 crc kubenswrapper[4824]: I1006 09:59:38.703150 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Oct 06 09:59:38 crc kubenswrapper[4824]: I1006 09:59:38.741045 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=15.741005324 podStartE2EDuration="15.741005324s" podCreationTimestamp="2025-10-06 09:59:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 09:59:38.718277835 +0000 UTC m=+88.082700736" watchObservedRunningTime="2025-10-06 09:59:38.741005324 +0000 UTC m=+88.105428245" Oct 06 09:59:38 crc kubenswrapper[4824]: I1006 09:59:38.771787 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c8ac88b0-596c-4a40-857b-f800ab05c42c-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-sn4f7\" (UID: \"c8ac88b0-596c-4a40-857b-f800ab05c42c\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-sn4f7" Oct 06 09:59:38 crc kubenswrapper[4824]: I1006 09:59:38.771879 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/c8ac88b0-596c-4a40-857b-f800ab05c42c-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-sn4f7\" (UID: \"c8ac88b0-596c-4a40-857b-f800ab05c42c\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-sn4f7" Oct 06 09:59:38 crc kubenswrapper[4824]: I1006 09:59:38.771948 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/c8ac88b0-596c-4a40-857b-f800ab05c42c-service-ca\") pod \"cluster-version-operator-5c965bbfc6-sn4f7\" (UID: \"c8ac88b0-596c-4a40-857b-f800ab05c42c\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-sn4f7" Oct 06 09:59:38 crc kubenswrapper[4824]: I1006 09:59:38.772048 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c8ac88b0-596c-4a40-857b-f800ab05c42c-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-sn4f7\" (UID: \"c8ac88b0-596c-4a40-857b-f800ab05c42c\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-sn4f7" Oct 06 09:59:38 crc kubenswrapper[4824]: I1006 09:59:38.772106 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/c8ac88b0-596c-4a40-857b-f800ab05c42c-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-sn4f7\" (UID: \"c8ac88b0-596c-4a40-857b-f800ab05c42c\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-sn4f7" Oct 06 09:59:38 crc kubenswrapper[4824]: I1006 09:59:38.847615 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-5rx6l" podStartSLOduration=66.847584083 podStartE2EDuration="1m6.847584083s" podCreationTimestamp="2025-10-06 09:58:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 09:59:38.825645844 +0000 UTC m=+88.190068715" watchObservedRunningTime="2025-10-06 09:59:38.847584083 +0000 UTC m=+88.212006954" Oct 06 09:59:38 crc kubenswrapper[4824]: I1006 09:59:38.872999 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/c8ac88b0-596c-4a40-857b-f800ab05c42c-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-sn4f7\" (UID: \"c8ac88b0-596c-4a40-857b-f800ab05c42c\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-sn4f7" Oct 06 09:59:38 crc kubenswrapper[4824]: I1006 09:59:38.873100 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c8ac88b0-596c-4a40-857b-f800ab05c42c-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-sn4f7\" (UID: \"c8ac88b0-596c-4a40-857b-f800ab05c42c\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-sn4f7" Oct 06 09:59:38 crc kubenswrapper[4824]: I1006 09:59:38.873136 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/c8ac88b0-596c-4a40-857b-f800ab05c42c-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-sn4f7\" (UID: \"c8ac88b0-596c-4a40-857b-f800ab05c42c\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-sn4f7" Oct 06 09:59:38 crc kubenswrapper[4824]: I1006 09:59:38.873194 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/c8ac88b0-596c-4a40-857b-f800ab05c42c-service-ca\") pod \"cluster-version-operator-5c965bbfc6-sn4f7\" (UID: \"c8ac88b0-596c-4a40-857b-f800ab05c42c\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-sn4f7" Oct 06 09:59:38 crc kubenswrapper[4824]: I1006 09:59:38.873219 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c8ac88b0-596c-4a40-857b-f800ab05c42c-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-sn4f7\" (UID: \"c8ac88b0-596c-4a40-857b-f800ab05c42c\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-sn4f7" Oct 06 09:59:38 crc kubenswrapper[4824]: I1006 09:59:38.873528 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/c8ac88b0-596c-4a40-857b-f800ab05c42c-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-sn4f7\" (UID: \"c8ac88b0-596c-4a40-857b-f800ab05c42c\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-sn4f7" Oct 06 09:59:38 crc kubenswrapper[4824]: I1006 09:59:38.873701 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/c8ac88b0-596c-4a40-857b-f800ab05c42c-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-sn4f7\" (UID: \"c8ac88b0-596c-4a40-857b-f800ab05c42c\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-sn4f7" Oct 06 09:59:38 crc kubenswrapper[4824]: I1006 09:59:38.874452 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/c8ac88b0-596c-4a40-857b-f800ab05c42c-service-ca\") pod \"cluster-version-operator-5c965bbfc6-sn4f7\" (UID: \"c8ac88b0-596c-4a40-857b-f800ab05c42c\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-sn4f7" Oct 06 09:59:38 crc kubenswrapper[4824]: I1006 09:59:38.884932 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c8ac88b0-596c-4a40-857b-f800ab05c42c-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-sn4f7\" (UID: \"c8ac88b0-596c-4a40-857b-f800ab05c42c\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-sn4f7" Oct 06 09:59:38 crc kubenswrapper[4824]: I1006 09:59:38.898144 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=68.898116208 podStartE2EDuration="1m8.898116208s" podCreationTimestamp="2025-10-06 09:58:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 09:59:38.897860382 +0000 UTC m=+88.262283253" watchObservedRunningTime="2025-10-06 09:59:38.898116208 +0000 UTC m=+88.262539089" Oct 06 09:59:38 crc kubenswrapper[4824]: I1006 09:59:38.898301 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=68.898293862 podStartE2EDuration="1m8.898293862s" podCreationTimestamp="2025-10-06 09:58:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 09:59:38.870467166 +0000 UTC m=+88.234890047" watchObservedRunningTime="2025-10-06 09:59:38.898293862 +0000 UTC m=+88.262716733" Oct 06 09:59:38 crc kubenswrapper[4824]: I1006 09:59:38.901297 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c8ac88b0-596c-4a40-857b-f800ab05c42c-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-sn4f7\" (UID: \"c8ac88b0-596c-4a40-857b-f800ab05c42c\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-sn4f7" Oct 06 09:59:38 crc kubenswrapper[4824]: I1006 09:59:38.939647 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-szn8q" podStartSLOduration=67.939614377 podStartE2EDuration="1m7.939614377s" podCreationTimestamp="2025-10-06 09:58:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 09:59:38.939204697 +0000 UTC m=+88.303627568" watchObservedRunningTime="2025-10-06 09:59:38.939614377 +0000 UTC m=+88.304037278" Oct 06 09:59:38 crc kubenswrapper[4824]: I1006 09:59:38.991646 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lszvg" podStartSLOduration=66.991591249 podStartE2EDuration="1m6.991591249s" podCreationTimestamp="2025-10-06 09:58:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 09:59:38.990164613 +0000 UTC m=+88.354587534" watchObservedRunningTime="2025-10-06 09:59:38.991591249 +0000 UTC m=+88.356014150" Oct 06 09:59:39 crc kubenswrapper[4824]: I1006 09:59:39.018827 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-sn4f7" Oct 06 09:59:39 crc kubenswrapper[4824]: I1006 09:59:39.033941 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=36.033908278 podStartE2EDuration="36.033908278s" podCreationTimestamp="2025-10-06 09:59:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 09:59:39.031932149 +0000 UTC m=+88.396355020" watchObservedRunningTime="2025-10-06 09:59:39.033908278 +0000 UTC m=+88.398331179" Oct 06 09:59:39 crc kubenswrapper[4824]: I1006 09:59:39.058217 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-7c22q" podStartSLOduration=68.058185896 podStartE2EDuration="1m8.058185896s" podCreationTimestamp="2025-10-06 09:58:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 09:59:39.058037493 +0000 UTC m=+88.422460394" watchObservedRunningTime="2025-10-06 09:59:39.058185896 +0000 UTC m=+88.422608767" Oct 06 09:59:39 crc kubenswrapper[4824]: I1006 09:59:39.099966 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-xcgwc" podStartSLOduration=68.099940342 podStartE2EDuration="1m8.099940342s" podCreationTimestamp="2025-10-06 09:58:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 09:59:39.099919251 +0000 UTC m=+88.464342122" watchObservedRunningTime="2025-10-06 09:59:39.099940342 +0000 UTC m=+88.464363193" Oct 06 09:59:39 crc kubenswrapper[4824]: I1006 09:59:39.118324 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podStartSLOduration=68.118308842 podStartE2EDuration="1m8.118308842s" podCreationTimestamp="2025-10-06 09:58:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 09:59:39.117232975 +0000 UTC m=+88.481655846" watchObservedRunningTime="2025-10-06 09:59:39.118308842 +0000 UTC m=+88.482731703" Oct 06 09:59:39 crc kubenswrapper[4824]: I1006 09:59:39.940821 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-sn4f7" event={"ID":"c8ac88b0-596c-4a40-857b-f800ab05c42c","Type":"ContainerStarted","Data":"64c7addaaeb8466584eaa6ebd281e09aef5458e266e376397c22a59b016d9d04"} Oct 06 09:59:39 crc kubenswrapper[4824]: I1006 09:59:39.941264 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-sn4f7" event={"ID":"c8ac88b0-596c-4a40-857b-f800ab05c42c","Type":"ContainerStarted","Data":"b151dc07e9b5f974607b6f0fca301c2e04bc515dc2bef8f9b0f04df88bff0630"} Oct 06 09:59:39 crc kubenswrapper[4824]: I1006 09:59:39.964485 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-sn4f7" podStartSLOduration=68.964454018 podStartE2EDuration="1m8.964454018s" podCreationTimestamp="2025-10-06 09:58:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 09:59:39.96295421 +0000 UTC m=+89.327377111" watchObservedRunningTime="2025-10-06 09:59:39.964454018 +0000 UTC m=+89.328876919" Oct 06 09:59:40 crc kubenswrapper[4824]: I1006 09:59:40.273402 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 09:59:40 crc kubenswrapper[4824]: I1006 09:59:40.273465 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 09:59:40 crc kubenswrapper[4824]: I1006 09:59:40.273497 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 09:59:40 crc kubenswrapper[4824]: E1006 09:59:40.273863 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 06 09:59:40 crc kubenswrapper[4824]: I1006 09:59:40.273660 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gnw94" Oct 06 09:59:40 crc kubenswrapper[4824]: E1006 09:59:40.274305 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 06 09:59:40 crc kubenswrapper[4824]: E1006 09:59:40.274519 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 06 09:59:40 crc kubenswrapper[4824]: E1006 09:59:40.274728 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gnw94" podUID="7cf1d355-64cb-48a6-acfd-7d258b7afa10" Oct 06 09:59:42 crc kubenswrapper[4824]: I1006 09:59:42.273398 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 09:59:42 crc kubenswrapper[4824]: I1006 09:59:42.273496 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gnw94" Oct 06 09:59:42 crc kubenswrapper[4824]: I1006 09:59:42.273401 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 09:59:42 crc kubenswrapper[4824]: E1006 09:59:42.273587 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 06 09:59:42 crc kubenswrapper[4824]: I1006 09:59:42.273642 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 09:59:42 crc kubenswrapper[4824]: E1006 09:59:42.274093 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 06 09:59:42 crc kubenswrapper[4824]: E1006 09:59:42.274277 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 06 09:59:42 crc kubenswrapper[4824]: E1006 09:59:42.274426 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gnw94" podUID="7cf1d355-64cb-48a6-acfd-7d258b7afa10" Oct 06 09:59:44 crc kubenswrapper[4824]: I1006 09:59:44.273104 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 09:59:44 crc kubenswrapper[4824]: I1006 09:59:44.273172 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gnw94" Oct 06 09:59:44 crc kubenswrapper[4824]: I1006 09:59:44.273127 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 09:59:44 crc kubenswrapper[4824]: E1006 09:59:44.273281 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 06 09:59:44 crc kubenswrapper[4824]: I1006 09:59:44.273314 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 09:59:44 crc kubenswrapper[4824]: E1006 09:59:44.273489 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gnw94" podUID="7cf1d355-64cb-48a6-acfd-7d258b7afa10" Oct 06 09:59:44 crc kubenswrapper[4824]: E1006 09:59:44.273604 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 06 09:59:44 crc kubenswrapper[4824]: E1006 09:59:44.273681 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 06 09:59:46 crc kubenswrapper[4824]: I1006 09:59:46.273380 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 09:59:46 crc kubenswrapper[4824]: I1006 09:59:46.273615 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gnw94" Oct 06 09:59:46 crc kubenswrapper[4824]: I1006 09:59:46.273721 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 09:59:46 crc kubenswrapper[4824]: E1006 09:59:46.273835 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 06 09:59:46 crc kubenswrapper[4824]: I1006 09:59:46.273857 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 09:59:46 crc kubenswrapper[4824]: E1006 09:59:46.274428 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 06 09:59:46 crc kubenswrapper[4824]: E1006 09:59:46.274636 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gnw94" podUID="7cf1d355-64cb-48a6-acfd-7d258b7afa10" Oct 06 09:59:46 crc kubenswrapper[4824]: E1006 09:59:46.274849 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 06 09:59:46 crc kubenswrapper[4824]: I1006 09:59:46.275273 4824 scope.go:117] "RemoveContainer" containerID="9c88940bd9c33c0ee0feef797ad549f874cad42942352072dba2b4eaf1e8b803" Oct 06 09:59:46 crc kubenswrapper[4824]: E1006 09:59:46.275553 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-4vwms_openshift-ovn-kubernetes(2f511ece-25eb-465f-b85c-9e5f3f886c21)\"" pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" podUID="2f511ece-25eb-465f-b85c-9e5f3f886c21" Oct 06 09:59:48 crc kubenswrapper[4824]: I1006 09:59:48.273819 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 09:59:48 crc kubenswrapper[4824]: I1006 09:59:48.273855 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 09:59:48 crc kubenswrapper[4824]: I1006 09:59:48.273956 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gnw94" Oct 06 09:59:48 crc kubenswrapper[4824]: E1006 09:59:48.274736 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 06 09:59:48 crc kubenswrapper[4824]: E1006 09:59:48.274387 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 06 09:59:48 crc kubenswrapper[4824]: E1006 09:59:48.274847 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gnw94" podUID="7cf1d355-64cb-48a6-acfd-7d258b7afa10" Oct 06 09:59:48 crc kubenswrapper[4824]: I1006 09:59:48.274236 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 09:59:48 crc kubenswrapper[4824]: E1006 09:59:48.275029 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 06 09:59:50 crc kubenswrapper[4824]: I1006 09:59:50.210477 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7cf1d355-64cb-48a6-acfd-7d258b7afa10-metrics-certs\") pod \"network-metrics-daemon-gnw94\" (UID: \"7cf1d355-64cb-48a6-acfd-7d258b7afa10\") " pod="openshift-multus/network-metrics-daemon-gnw94" Oct 06 09:59:50 crc kubenswrapper[4824]: E1006 09:59:50.210692 4824 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 06 09:59:50 crc kubenswrapper[4824]: E1006 09:59:50.210772 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7cf1d355-64cb-48a6-acfd-7d258b7afa10-metrics-certs podName:7cf1d355-64cb-48a6-acfd-7d258b7afa10 nodeName:}" failed. No retries permitted until 2025-10-06 10:00:54.210748411 +0000 UTC m=+163.575171282 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/7cf1d355-64cb-48a6-acfd-7d258b7afa10-metrics-certs") pod "network-metrics-daemon-gnw94" (UID: "7cf1d355-64cb-48a6-acfd-7d258b7afa10") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 06 09:59:50 crc kubenswrapper[4824]: I1006 09:59:50.273939 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 09:59:50 crc kubenswrapper[4824]: I1006 09:59:50.274045 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 09:59:50 crc kubenswrapper[4824]: E1006 09:59:50.274074 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 06 09:59:50 crc kubenswrapper[4824]: I1006 09:59:50.274071 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gnw94" Oct 06 09:59:50 crc kubenswrapper[4824]: E1006 09:59:50.274189 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 06 09:59:50 crc kubenswrapper[4824]: E1006 09:59:50.274297 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gnw94" podUID="7cf1d355-64cb-48a6-acfd-7d258b7afa10" Oct 06 09:59:50 crc kubenswrapper[4824]: I1006 09:59:50.274410 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 09:59:50 crc kubenswrapper[4824]: E1006 09:59:50.274671 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 06 09:59:52 crc kubenswrapper[4824]: I1006 09:59:52.273334 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 09:59:52 crc kubenswrapper[4824]: I1006 09:59:52.273421 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 09:59:52 crc kubenswrapper[4824]: I1006 09:59:52.273461 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gnw94" Oct 06 09:59:52 crc kubenswrapper[4824]: E1006 09:59:52.273660 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 06 09:59:52 crc kubenswrapper[4824]: I1006 09:59:52.273770 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 09:59:52 crc kubenswrapper[4824]: E1006 09:59:52.273857 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 06 09:59:52 crc kubenswrapper[4824]: E1006 09:59:52.273960 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 06 09:59:52 crc kubenswrapper[4824]: E1006 09:59:52.274296 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gnw94" podUID="7cf1d355-64cb-48a6-acfd-7d258b7afa10" Oct 06 09:59:54 crc kubenswrapper[4824]: I1006 09:59:54.273053 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gnw94" Oct 06 09:59:54 crc kubenswrapper[4824]: I1006 09:59:54.273063 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 09:59:54 crc kubenswrapper[4824]: I1006 09:59:54.273079 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 09:59:54 crc kubenswrapper[4824]: E1006 09:59:54.273843 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 06 09:59:54 crc kubenswrapper[4824]: E1006 09:59:54.273643 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gnw94" podUID="7cf1d355-64cb-48a6-acfd-7d258b7afa10" Oct 06 09:59:54 crc kubenswrapper[4824]: I1006 09:59:54.273114 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 09:59:54 crc kubenswrapper[4824]: E1006 09:59:54.273954 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 06 09:59:54 crc kubenswrapper[4824]: E1006 09:59:54.274056 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 06 09:59:55 crc kubenswrapper[4824]: I1006 09:59:55.300781 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Oct 06 09:59:56 crc kubenswrapper[4824]: I1006 09:59:56.273402 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gnw94" Oct 06 09:59:56 crc kubenswrapper[4824]: I1006 09:59:56.273461 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 09:59:56 crc kubenswrapper[4824]: I1006 09:59:56.273489 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 09:59:56 crc kubenswrapper[4824]: E1006 09:59:56.273596 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gnw94" podUID="7cf1d355-64cb-48a6-acfd-7d258b7afa10" Oct 06 09:59:56 crc kubenswrapper[4824]: I1006 09:59:56.273430 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 09:59:56 crc kubenswrapper[4824]: E1006 09:59:56.273780 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 06 09:59:56 crc kubenswrapper[4824]: E1006 09:59:56.274019 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 06 09:59:56 crc kubenswrapper[4824]: E1006 09:59:56.274136 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 06 09:59:57 crc kubenswrapper[4824]: I1006 09:59:57.274472 4824 scope.go:117] "RemoveContainer" containerID="9c88940bd9c33c0ee0feef797ad549f874cad42942352072dba2b4eaf1e8b803" Oct 06 09:59:57 crc kubenswrapper[4824]: E1006 09:59:57.274677 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-4vwms_openshift-ovn-kubernetes(2f511ece-25eb-465f-b85c-9e5f3f886c21)\"" pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" podUID="2f511ece-25eb-465f-b85c-9e5f3f886c21" Oct 06 09:59:58 crc kubenswrapper[4824]: I1006 09:59:58.273948 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 09:59:58 crc kubenswrapper[4824]: I1006 09:59:58.274109 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gnw94" Oct 06 09:59:58 crc kubenswrapper[4824]: E1006 09:59:58.274113 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 06 09:59:58 crc kubenswrapper[4824]: E1006 09:59:58.274312 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gnw94" podUID="7cf1d355-64cb-48a6-acfd-7d258b7afa10" Oct 06 09:59:58 crc kubenswrapper[4824]: I1006 09:59:58.274681 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 09:59:58 crc kubenswrapper[4824]: I1006 09:59:58.274827 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 09:59:58 crc kubenswrapper[4824]: E1006 09:59:58.275355 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 06 09:59:58 crc kubenswrapper[4824]: E1006 09:59:58.275555 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 06 10:00:00 crc kubenswrapper[4824]: I1006 10:00:00.273809 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 10:00:00 crc kubenswrapper[4824]: I1006 10:00:00.273862 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gnw94" Oct 06 10:00:00 crc kubenswrapper[4824]: E1006 10:00:00.274072 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 06 10:00:00 crc kubenswrapper[4824]: I1006 10:00:00.274094 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 10:00:00 crc kubenswrapper[4824]: I1006 10:00:00.274122 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 10:00:00 crc kubenswrapper[4824]: E1006 10:00:00.274294 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gnw94" podUID="7cf1d355-64cb-48a6-acfd-7d258b7afa10" Oct 06 10:00:00 crc kubenswrapper[4824]: E1006 10:00:00.274424 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 06 10:00:00 crc kubenswrapper[4824]: E1006 10:00:00.274539 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 06 10:00:01 crc kubenswrapper[4824]: I1006 10:00:01.316884 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=6.316859485 podStartE2EDuration="6.316859485s" podCreationTimestamp="2025-10-06 09:59:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:00:01.314921867 +0000 UTC m=+110.679344728" watchObservedRunningTime="2025-10-06 10:00:01.316859485 +0000 UTC m=+110.681282346" Oct 06 10:00:02 crc kubenswrapper[4824]: I1006 10:00:02.273296 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gnw94" Oct 06 10:00:02 crc kubenswrapper[4824]: I1006 10:00:02.273385 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 10:00:02 crc kubenswrapper[4824]: I1006 10:00:02.273297 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 10:00:02 crc kubenswrapper[4824]: I1006 10:00:02.273434 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 10:00:02 crc kubenswrapper[4824]: E1006 10:00:02.273501 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gnw94" podUID="7cf1d355-64cb-48a6-acfd-7d258b7afa10" Oct 06 10:00:02 crc kubenswrapper[4824]: E1006 10:00:02.273572 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 06 10:00:02 crc kubenswrapper[4824]: E1006 10:00:02.273622 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 06 10:00:02 crc kubenswrapper[4824]: E1006 10:00:02.273905 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 06 10:00:04 crc kubenswrapper[4824]: I1006 10:00:04.273523 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 10:00:04 crc kubenswrapper[4824]: I1006 10:00:04.273601 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gnw94" Oct 06 10:00:04 crc kubenswrapper[4824]: I1006 10:00:04.273544 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 10:00:04 crc kubenswrapper[4824]: I1006 10:00:04.273767 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 10:00:04 crc kubenswrapper[4824]: E1006 10:00:04.273750 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 06 10:00:04 crc kubenswrapper[4824]: E1006 10:00:04.273911 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gnw94" podUID="7cf1d355-64cb-48a6-acfd-7d258b7afa10" Oct 06 10:00:04 crc kubenswrapper[4824]: E1006 10:00:04.274096 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 06 10:00:04 crc kubenswrapper[4824]: E1006 10:00:04.274203 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 06 10:00:06 crc kubenswrapper[4824]: I1006 10:00:06.044412 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-szn8q_b65abf9f-4e88-4571-960d-3ca997d9c344/kube-multus/1.log" Oct 06 10:00:06 crc kubenswrapper[4824]: I1006 10:00:06.044874 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-szn8q_b65abf9f-4e88-4571-960d-3ca997d9c344/kube-multus/0.log" Oct 06 10:00:06 crc kubenswrapper[4824]: I1006 10:00:06.044918 4824 generic.go:334] "Generic (PLEG): container finished" podID="b65abf9f-4e88-4571-960d-3ca997d9c344" containerID="c80f7c0d85c4c76aa3c4ff4e078900ccc3be12fa5bb75f595ead6eeacf39e3f4" exitCode=1 Oct 06 10:00:06 crc kubenswrapper[4824]: I1006 10:00:06.044962 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-szn8q" event={"ID":"b65abf9f-4e88-4571-960d-3ca997d9c344","Type":"ContainerDied","Data":"c80f7c0d85c4c76aa3c4ff4e078900ccc3be12fa5bb75f595ead6eeacf39e3f4"} Oct 06 10:00:06 crc kubenswrapper[4824]: I1006 10:00:06.045046 4824 scope.go:117] "RemoveContainer" containerID="664bf42e1324a9c9529f36f08d61e26999b3386bdd1fac77851fb7e7411e8934" Oct 06 10:00:06 crc kubenswrapper[4824]: I1006 10:00:06.046621 4824 scope.go:117] "RemoveContainer" containerID="c80f7c0d85c4c76aa3c4ff4e078900ccc3be12fa5bb75f595ead6eeacf39e3f4" Oct 06 10:00:06 crc kubenswrapper[4824]: E1006 10:00:06.050621 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-szn8q_openshift-multus(b65abf9f-4e88-4571-960d-3ca997d9c344)\"" pod="openshift-multus/multus-szn8q" podUID="b65abf9f-4e88-4571-960d-3ca997d9c344" Oct 06 10:00:06 crc kubenswrapper[4824]: I1006 10:00:06.273209 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 10:00:06 crc kubenswrapper[4824]: I1006 10:00:06.273312 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 10:00:06 crc kubenswrapper[4824]: I1006 10:00:06.273428 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gnw94" Oct 06 10:00:06 crc kubenswrapper[4824]: E1006 10:00:06.273431 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 06 10:00:06 crc kubenswrapper[4824]: I1006 10:00:06.273471 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 10:00:06 crc kubenswrapper[4824]: E1006 10:00:06.273624 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 06 10:00:06 crc kubenswrapper[4824]: E1006 10:00:06.273817 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gnw94" podUID="7cf1d355-64cb-48a6-acfd-7d258b7afa10" Oct 06 10:00:06 crc kubenswrapper[4824]: E1006 10:00:06.273897 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 06 10:00:07 crc kubenswrapper[4824]: I1006 10:00:07.051233 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-szn8q_b65abf9f-4e88-4571-960d-3ca997d9c344/kube-multus/1.log" Oct 06 10:00:08 crc kubenswrapper[4824]: I1006 10:00:08.273517 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gnw94" Oct 06 10:00:08 crc kubenswrapper[4824]: I1006 10:00:08.273569 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 10:00:08 crc kubenswrapper[4824]: I1006 10:00:08.273672 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 10:00:08 crc kubenswrapper[4824]: I1006 10:00:08.273818 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 10:00:08 crc kubenswrapper[4824]: E1006 10:00:08.274067 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 06 10:00:08 crc kubenswrapper[4824]: E1006 10:00:08.273873 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gnw94" podUID="7cf1d355-64cb-48a6-acfd-7d258b7afa10" Oct 06 10:00:08 crc kubenswrapper[4824]: E1006 10:00:08.274299 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 06 10:00:08 crc kubenswrapper[4824]: E1006 10:00:08.274383 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 06 10:00:09 crc kubenswrapper[4824]: I1006 10:00:09.275423 4824 scope.go:117] "RemoveContainer" containerID="9c88940bd9c33c0ee0feef797ad549f874cad42942352072dba2b4eaf1e8b803" Oct 06 10:00:09 crc kubenswrapper[4824]: E1006 10:00:09.275647 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-4vwms_openshift-ovn-kubernetes(2f511ece-25eb-465f-b85c-9e5f3f886c21)\"" pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" podUID="2f511ece-25eb-465f-b85c-9e5f3f886c21" Oct 06 10:00:10 crc kubenswrapper[4824]: I1006 10:00:10.273834 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 10:00:10 crc kubenswrapper[4824]: I1006 10:00:10.273955 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 10:00:10 crc kubenswrapper[4824]: I1006 10:00:10.274095 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gnw94" Oct 06 10:00:10 crc kubenswrapper[4824]: E1006 10:00:10.274024 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 06 10:00:10 crc kubenswrapper[4824]: I1006 10:00:10.274123 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 10:00:10 crc kubenswrapper[4824]: E1006 10:00:10.274247 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 06 10:00:10 crc kubenswrapper[4824]: E1006 10:00:10.274382 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gnw94" podUID="7cf1d355-64cb-48a6-acfd-7d258b7afa10" Oct 06 10:00:10 crc kubenswrapper[4824]: E1006 10:00:10.274843 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 06 10:00:11 crc kubenswrapper[4824]: E1006 10:00:11.280233 4824 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Oct 06 10:00:11 crc kubenswrapper[4824]: E1006 10:00:11.385513 4824 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 06 10:00:12 crc kubenswrapper[4824]: I1006 10:00:12.273843 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gnw94" Oct 06 10:00:12 crc kubenswrapper[4824]: I1006 10:00:12.273923 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 10:00:12 crc kubenswrapper[4824]: I1006 10:00:12.273923 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 10:00:12 crc kubenswrapper[4824]: I1006 10:00:12.274494 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 10:00:12 crc kubenswrapper[4824]: E1006 10:00:12.274709 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gnw94" podUID="7cf1d355-64cb-48a6-acfd-7d258b7afa10" Oct 06 10:00:12 crc kubenswrapper[4824]: E1006 10:00:12.274884 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 06 10:00:12 crc kubenswrapper[4824]: E1006 10:00:12.275026 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 06 10:00:12 crc kubenswrapper[4824]: E1006 10:00:12.275169 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 06 10:00:14 crc kubenswrapper[4824]: I1006 10:00:14.273483 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 10:00:14 crc kubenswrapper[4824]: I1006 10:00:14.273541 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gnw94" Oct 06 10:00:14 crc kubenswrapper[4824]: E1006 10:00:14.274308 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 06 10:00:14 crc kubenswrapper[4824]: I1006 10:00:14.273637 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 10:00:14 crc kubenswrapper[4824]: E1006 10:00:14.274381 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gnw94" podUID="7cf1d355-64cb-48a6-acfd-7d258b7afa10" Oct 06 10:00:14 crc kubenswrapper[4824]: I1006 10:00:14.273591 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 10:00:14 crc kubenswrapper[4824]: E1006 10:00:14.274418 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 06 10:00:14 crc kubenswrapper[4824]: E1006 10:00:14.274497 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 06 10:00:16 crc kubenswrapper[4824]: I1006 10:00:16.273096 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 10:00:16 crc kubenswrapper[4824]: E1006 10:00:16.274402 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 06 10:00:16 crc kubenswrapper[4824]: I1006 10:00:16.273195 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 10:00:16 crc kubenswrapper[4824]: E1006 10:00:16.274565 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 06 10:00:16 crc kubenswrapper[4824]: I1006 10:00:16.273113 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 10:00:16 crc kubenswrapper[4824]: E1006 10:00:16.274655 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 06 10:00:16 crc kubenswrapper[4824]: I1006 10:00:16.273188 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gnw94" Oct 06 10:00:16 crc kubenswrapper[4824]: E1006 10:00:16.274780 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gnw94" podUID="7cf1d355-64cb-48a6-acfd-7d258b7afa10" Oct 06 10:00:16 crc kubenswrapper[4824]: E1006 10:00:16.387435 4824 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 06 10:00:18 crc kubenswrapper[4824]: I1006 10:00:18.273102 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 10:00:18 crc kubenswrapper[4824]: I1006 10:00:18.273220 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 10:00:18 crc kubenswrapper[4824]: I1006 10:00:18.273101 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 10:00:18 crc kubenswrapper[4824]: E1006 10:00:18.273317 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 06 10:00:18 crc kubenswrapper[4824]: E1006 10:00:18.273430 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 06 10:00:18 crc kubenswrapper[4824]: I1006 10:00:18.273465 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gnw94" Oct 06 10:00:18 crc kubenswrapper[4824]: E1006 10:00:18.273539 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 06 10:00:18 crc kubenswrapper[4824]: E1006 10:00:18.273732 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gnw94" podUID="7cf1d355-64cb-48a6-acfd-7d258b7afa10" Oct 06 10:00:19 crc kubenswrapper[4824]: I1006 10:00:19.274114 4824 scope.go:117] "RemoveContainer" containerID="c80f7c0d85c4c76aa3c4ff4e078900ccc3be12fa5bb75f595ead6eeacf39e3f4" Oct 06 10:00:20 crc kubenswrapper[4824]: I1006 10:00:20.103171 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-szn8q_b65abf9f-4e88-4571-960d-3ca997d9c344/kube-multus/1.log" Oct 06 10:00:20 crc kubenswrapper[4824]: I1006 10:00:20.103701 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-szn8q" event={"ID":"b65abf9f-4e88-4571-960d-3ca997d9c344","Type":"ContainerStarted","Data":"238ede7256cac751995c87da2f05e25727cc9496ddf838e9d57f365f50acf4ba"} Oct 06 10:00:20 crc kubenswrapper[4824]: I1006 10:00:20.274100 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 10:00:20 crc kubenswrapper[4824]: E1006 10:00:20.274533 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 06 10:00:20 crc kubenswrapper[4824]: I1006 10:00:20.274136 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gnw94" Oct 06 10:00:20 crc kubenswrapper[4824]: E1006 10:00:20.275202 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gnw94" podUID="7cf1d355-64cb-48a6-acfd-7d258b7afa10" Oct 06 10:00:20 crc kubenswrapper[4824]: I1006 10:00:20.274097 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 10:00:20 crc kubenswrapper[4824]: E1006 10:00:20.275431 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 06 10:00:20 crc kubenswrapper[4824]: I1006 10:00:20.274169 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 10:00:20 crc kubenswrapper[4824]: E1006 10:00:20.275695 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 06 10:00:21 crc kubenswrapper[4824]: E1006 10:00:21.388290 4824 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 06 10:00:22 crc kubenswrapper[4824]: I1006 10:00:22.273425 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 10:00:22 crc kubenswrapper[4824]: I1006 10:00:22.273511 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 10:00:22 crc kubenswrapper[4824]: I1006 10:00:22.273581 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 10:00:22 crc kubenswrapper[4824]: I1006 10:00:22.273626 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gnw94" Oct 06 10:00:22 crc kubenswrapper[4824]: E1006 10:00:22.273769 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 06 10:00:22 crc kubenswrapper[4824]: E1006 10:00:22.273881 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 06 10:00:22 crc kubenswrapper[4824]: E1006 10:00:22.273942 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 06 10:00:22 crc kubenswrapper[4824]: E1006 10:00:22.274021 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gnw94" podUID="7cf1d355-64cb-48a6-acfd-7d258b7afa10" Oct 06 10:00:24 crc kubenswrapper[4824]: I1006 10:00:24.273214 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 10:00:24 crc kubenswrapper[4824]: E1006 10:00:24.273399 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 06 10:00:24 crc kubenswrapper[4824]: I1006 10:00:24.273232 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 10:00:24 crc kubenswrapper[4824]: E1006 10:00:24.273655 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 06 10:00:24 crc kubenswrapper[4824]: I1006 10:00:24.273799 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gnw94" Oct 06 10:00:24 crc kubenswrapper[4824]: E1006 10:00:24.273863 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gnw94" podUID="7cf1d355-64cb-48a6-acfd-7d258b7afa10" Oct 06 10:00:24 crc kubenswrapper[4824]: I1006 10:00:24.274011 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 10:00:24 crc kubenswrapper[4824]: E1006 10:00:24.274064 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 06 10:00:24 crc kubenswrapper[4824]: I1006 10:00:24.277107 4824 scope.go:117] "RemoveContainer" containerID="9c88940bd9c33c0ee0feef797ad549f874cad42942352072dba2b4eaf1e8b803" Oct 06 10:00:25 crc kubenswrapper[4824]: I1006 10:00:25.129423 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-4vwms_2f511ece-25eb-465f-b85c-9e5f3f886c21/ovnkube-controller/3.log" Oct 06 10:00:25 crc kubenswrapper[4824]: I1006 10:00:25.133028 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" event={"ID":"2f511ece-25eb-465f-b85c-9e5f3f886c21","Type":"ContainerStarted","Data":"3dffdccf2aac654a5a0f629b8c5b2233c4561cb036262068f2289f6e95b46575"} Oct 06 10:00:25 crc kubenswrapper[4824]: I1006 10:00:25.133452 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" Oct 06 10:00:25 crc kubenswrapper[4824]: I1006 10:00:25.180712 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" podStartSLOduration=114.18069653 podStartE2EDuration="1m54.18069653s" podCreationTimestamp="2025-10-06 09:58:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:00:25.179364478 +0000 UTC m=+134.543787339" watchObservedRunningTime="2025-10-06 10:00:25.18069653 +0000 UTC m=+134.545119381" Oct 06 10:00:25 crc kubenswrapper[4824]: I1006 10:00:25.208019 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-gnw94"] Oct 06 10:00:25 crc kubenswrapper[4824]: I1006 10:00:25.208452 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gnw94" Oct 06 10:00:25 crc kubenswrapper[4824]: E1006 10:00:25.208620 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gnw94" podUID="7cf1d355-64cb-48a6-acfd-7d258b7afa10" Oct 06 10:00:26 crc kubenswrapper[4824]: I1006 10:00:26.274125 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 10:00:26 crc kubenswrapper[4824]: I1006 10:00:26.274123 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 10:00:26 crc kubenswrapper[4824]: I1006 10:00:26.274143 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 10:00:26 crc kubenswrapper[4824]: E1006 10:00:26.275734 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 06 10:00:26 crc kubenswrapper[4824]: E1006 10:00:26.275897 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 06 10:00:26 crc kubenswrapper[4824]: E1006 10:00:26.276174 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 06 10:00:26 crc kubenswrapper[4824]: E1006 10:00:26.389831 4824 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 06 10:00:27 crc kubenswrapper[4824]: I1006 10:00:27.273153 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gnw94" Oct 06 10:00:27 crc kubenswrapper[4824]: E1006 10:00:27.273642 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gnw94" podUID="7cf1d355-64cb-48a6-acfd-7d258b7afa10" Oct 06 10:00:28 crc kubenswrapper[4824]: I1006 10:00:28.273831 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 10:00:28 crc kubenswrapper[4824]: E1006 10:00:28.274018 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 06 10:00:28 crc kubenswrapper[4824]: I1006 10:00:28.273844 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 10:00:28 crc kubenswrapper[4824]: E1006 10:00:28.274232 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 06 10:00:28 crc kubenswrapper[4824]: I1006 10:00:28.275024 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 10:00:28 crc kubenswrapper[4824]: E1006 10:00:28.275259 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 06 10:00:29 crc kubenswrapper[4824]: I1006 10:00:29.273871 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gnw94" Oct 06 10:00:29 crc kubenswrapper[4824]: E1006 10:00:29.274184 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gnw94" podUID="7cf1d355-64cb-48a6-acfd-7d258b7afa10" Oct 06 10:00:30 crc kubenswrapper[4824]: I1006 10:00:30.273430 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 10:00:30 crc kubenswrapper[4824]: I1006 10:00:30.273431 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 10:00:30 crc kubenswrapper[4824]: I1006 10:00:30.273521 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 10:00:30 crc kubenswrapper[4824]: E1006 10:00:30.274710 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 06 10:00:30 crc kubenswrapper[4824]: E1006 10:00:30.274900 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 06 10:00:30 crc kubenswrapper[4824]: E1006 10:00:30.274928 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 06 10:00:31 crc kubenswrapper[4824]: I1006 10:00:31.273674 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gnw94" Oct 06 10:00:31 crc kubenswrapper[4824]: E1006 10:00:31.274817 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gnw94" podUID="7cf1d355-64cb-48a6-acfd-7d258b7afa10" Oct 06 10:00:32 crc kubenswrapper[4824]: I1006 10:00:32.273524 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 10:00:32 crc kubenswrapper[4824]: I1006 10:00:32.273665 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 10:00:32 crc kubenswrapper[4824]: I1006 10:00:32.274120 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 10:00:32 crc kubenswrapper[4824]: I1006 10:00:32.277117 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Oct 06 10:00:32 crc kubenswrapper[4824]: I1006 10:00:32.277892 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Oct 06 10:00:32 crc kubenswrapper[4824]: I1006 10:00:32.278298 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Oct 06 10:00:32 crc kubenswrapper[4824]: I1006 10:00:32.278399 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Oct 06 10:00:33 crc kubenswrapper[4824]: I1006 10:00:33.273362 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gnw94" Oct 06 10:00:33 crc kubenswrapper[4824]: I1006 10:00:33.277467 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Oct 06 10:00:33 crc kubenswrapper[4824]: I1006 10:00:33.277499 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Oct 06 10:00:38 crc kubenswrapper[4824]: I1006 10:00:38.285308 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 10:00:38 crc kubenswrapper[4824]: E1006 10:00:38.285574 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-06 10:02:40.28553179 +0000 UTC m=+269.649954721 (durationBeforeRetry 2m2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 10:00:38 crc kubenswrapper[4824]: I1006 10:00:38.285646 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 10:00:38 crc kubenswrapper[4824]: I1006 10:00:38.285838 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 10:00:38 crc kubenswrapper[4824]: I1006 10:00:38.287477 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 10:00:38 crc kubenswrapper[4824]: I1006 10:00:38.296761 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 10:00:38 crc kubenswrapper[4824]: I1006 10:00:38.328667 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 10:00:38 crc kubenswrapper[4824]: I1006 10:00:38.388182 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 10:00:38 crc kubenswrapper[4824]: I1006 10:00:38.388286 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 10:00:38 crc kubenswrapper[4824]: I1006 10:00:38.396220 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 10:00:38 crc kubenswrapper[4824]: I1006 10:00:38.397630 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 10:00:38 crc kubenswrapper[4824]: W1006 10:00:38.596119 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5fe485a1_e14f_4c09_b5b9_f252bc42b7e8.slice/crio-c25b92f70e12093c740539b0f67576281002bb5bc644a7a7f3382104910f7df2 WatchSource:0}: Error finding container c25b92f70e12093c740539b0f67576281002bb5bc644a7a7f3382104910f7df2: Status 404 returned error can't find the container with id c25b92f70e12093c740539b0f67576281002bb5bc644a7a7f3382104910f7df2 Oct 06 10:00:38 crc kubenswrapper[4824]: I1006 10:00:38.599098 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 10:00:38 crc kubenswrapper[4824]: I1006 10:00:38.617009 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 10:00:38 crc kubenswrapper[4824]: W1006 10:00:38.875996 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9d751cbb_f2e2_430d_9754_c882a5e924a5.slice/crio-4b5f227e5a97f99a214f92c9c1446f19b82a4d6d541f89ab533a043fe0d6e0a7 WatchSource:0}: Error finding container 4b5f227e5a97f99a214f92c9c1446f19b82a4d6d541f89ab533a043fe0d6e0a7: Status 404 returned error can't find the container with id 4b5f227e5a97f99a214f92c9c1446f19b82a4d6d541f89ab533a043fe0d6e0a7 Oct 06 10:00:38 crc kubenswrapper[4824]: W1006 10:00:38.905850 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3b6479f0_333b_4a96_9adf_2099afdc2447.slice/crio-fd9d4d67ffa5147a6141e77452c086c249a77c8b9159102fc49d58733796c209 WatchSource:0}: Error finding container fd9d4d67ffa5147a6141e77452c086c249a77c8b9159102fc49d58733796c209: Status 404 returned error can't find the container with id fd9d4d67ffa5147a6141e77452c086c249a77c8b9159102fc49d58733796c209 Oct 06 10:00:39 crc kubenswrapper[4824]: I1006 10:00:39.200669 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"960debf0388e88fba70511886124062fd38e79fb6335f0651a5afce9feaa90ab"} Oct 06 10:00:39 crc kubenswrapper[4824]: I1006 10:00:39.201367 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"fd9d4d67ffa5147a6141e77452c086c249a77c8b9159102fc49d58733796c209"} Oct 06 10:00:39 crc kubenswrapper[4824]: I1006 10:00:39.201683 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 10:00:39 crc kubenswrapper[4824]: I1006 10:00:39.203130 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"706b8b21b432848d3c2648c5e77b5b1f76503c35a3a53e77b9b31481ff106f8a"} Oct 06 10:00:39 crc kubenswrapper[4824]: I1006 10:00:39.203239 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"c25b92f70e12093c740539b0f67576281002bb5bc644a7a7f3382104910f7df2"} Oct 06 10:00:39 crc kubenswrapper[4824]: I1006 10:00:39.206252 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"2656ce2837f8251ad492a7cd959843d6ee01fddc852c5c49b5580d1f737776f9"} Oct 06 10:00:39 crc kubenswrapper[4824]: I1006 10:00:39.206301 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"4b5f227e5a97f99a214f92c9c1446f19b82a4d6d541f89ab533a043fe0d6e0a7"} Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.195754 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.245346 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-bjzzk"] Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.245808 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-bjzzk" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.247829 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-6kn7b"] Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.248609 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-6kn7b" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.251238 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-s4ht9"] Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.251659 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-s4ht9" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.255698 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-q76nt"] Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.256526 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-4hw5p"] Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.257182 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-4hw5p" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.257687 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-q76nt" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.257825 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wlbrp"] Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.258396 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wlbrp" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.260294 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-v6gnw"] Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.261282 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-fjdwv"] Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.261574 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-v6gnw" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.262726 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-7qwsv"] Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.263076 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-mdv87"] Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.263466 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-mdv87" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.263700 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-wk4bd"] Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.263837 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-fjdwv" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.264267 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wk4bd" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.264563 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-7qwsv" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.267002 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-7p7wb"] Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.267828 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7p7wb" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.274875 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-x4fdc"] Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.275439 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-x4fdc" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.277363 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-fpxws"] Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.277818 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.278071 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.278112 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-fpxws" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.278301 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.278391 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-dklh2"] Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.278537 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.279014 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-dklh2" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.287729 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.300588 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-prthm"] Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.301823 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-prthm" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.301709 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-8l8gd"] Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.307334 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.307492 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.308488 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.308942 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.309122 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.309308 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.308888 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.310118 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.310318 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.310418 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.310454 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.310728 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-8l8gd" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.311043 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.311257 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.311453 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.311799 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.312273 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.312299 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-nqmr6"] Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.312412 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.312500 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.312531 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.312605 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.312667 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.312681 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.312763 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.312852 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.313074 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.313341 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-nqmr6" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.313747 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.315318 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.315655 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.315936 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.316569 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.316935 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.317075 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.317200 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.317330 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.317375 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.317374 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.317498 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.317529 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.317618 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-j2m8k"] Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.317652 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.317711 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.317799 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.317812 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.317854 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.317961 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.318029 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.318038 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.318091 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-sb9c8"] Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.318169 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.318179 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.318274 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.318276 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.318399 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-hqhlr"] Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.318441 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.318471 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.318181 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.318636 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.318711 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.318907 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.318935 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-hqhlr" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.319190 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-j2m8k" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.319260 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-sb9c8" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.319719 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.319826 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.319855 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.319961 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.320159 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.320204 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.320398 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.320462 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.320583 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.320644 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.320815 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.320833 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.320400 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.320964 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.321018 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.321107 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.321176 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.321287 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.322240 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.322563 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-8z68k"] Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.322869 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.323668 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-8z68k" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.323919 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-vvtbh"] Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.327357 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.328077 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.328201 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.330274 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-vn26d"] Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.331582 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vvtbh" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.332435 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.332753 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.332812 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-vn26d" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.332897 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.332938 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-xr4mg"] Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.358483 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.358899 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-xr4mg" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.359914 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.358791 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-wp6kv"] Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.360091 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.361828 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.363583 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.363935 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-56948"] Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.364452 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-56948" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.364626 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-wp6kv" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.365041 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.367603 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.368537 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.370366 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.370535 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.370667 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.370922 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.371170 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.371345 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.371446 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-fn7dj"] Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.371850 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-fn7dj" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.374553 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.375309 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.376266 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-zzrfg"] Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.377210 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-2xznh"] Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.377550 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zzrfg" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.377781 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-2xznh" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.377682 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6tt2t"] Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.379146 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6tt2t" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.380116 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-s8j95"] Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.380545 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-px5fp"] Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.381304 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-qpxrp"] Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.381828 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-qpxrp" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.382615 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-s8j95" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.383060 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-px5fp" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.384451 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.386213 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-6kn7b"] Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.387177 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-s4ht9"] Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.387849 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-8rxsm"] Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.388553 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-8rxsm" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.392828 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29329080-9kpdv"] Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.393717 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.393867 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29329080-9kpdv" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.396198 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-qq4zz"] Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.397028 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-qq4zz" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.400084 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-5blgj"] Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.400814 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-5blgj" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.408177 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-bjzzk"] Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.411370 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wlbrp"] Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.412018 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jmr46\" (UniqueName: \"kubernetes.io/projected/171abefb-ea19-40fa-8a22-4b38956d5e4b-kube-api-access-jmr46\") pod \"downloads-7954f5f757-s4ht9\" (UID: \"171abefb-ea19-40fa-8a22-4b38956d5e4b\") " pod="openshift-console/downloads-7954f5f757-s4ht9" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.421508 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-q76nt"] Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.429114 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.431408 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.441011 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-v6gnw"] Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.452760 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-fdbvz"] Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.453839 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-wk4bd"] Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.453924 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-prthm"] Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.454107 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-fdbvz" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.454955 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-7p7wb"] Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.459591 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-dklh2"] Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.459671 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-x4fdc"] Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.459740 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-nqmr6"] Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.464682 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-nk67f"] Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.483840 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.485415 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-sb9c8"] Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.485458 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-8rxsm"] Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.485476 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6tt2t"] Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.485582 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-nk67f" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.486160 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.486208 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-56948"] Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.488324 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.488686 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-s8j95"] Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.488744 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-8z68k"] Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.493292 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-fpxws"] Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.494623 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-mdv87"] Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.496440 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-fjdwv"] Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.496888 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-xr4mg"] Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.498027 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-hqhlr"] Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.500326 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-5blgj"] Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.500369 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-8l8gd"] Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.501169 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-zzrfg"] Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.502810 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-vvtbh"] Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.503848 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29329080-9kpdv"] Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.504905 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-qpxrp"] Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.505964 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-fn7dj"] Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.507334 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-j2m8k"] Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.508365 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-qq4zz"] Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.509034 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-px5fp"] Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.510489 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.514412 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jmr46\" (UniqueName: \"kubernetes.io/projected/171abefb-ea19-40fa-8a22-4b38956d5e4b-kube-api-access-jmr46\") pod \"downloads-7954f5f757-s4ht9\" (UID: \"171abefb-ea19-40fa-8a22-4b38956d5e4b\") " pod="openshift-console/downloads-7954f5f757-s4ht9" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.515314 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-vn26d"] Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.516367 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-2xznh"] Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.521019 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-wp6kv"] Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.522517 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-vpg8w"] Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.524483 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-dr4mr"] Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.525162 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-dr4mr" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.525289 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-vpg8w" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.526609 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-nk67f"] Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.527457 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-vpg8w"] Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.529301 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-dr4mr"] Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.530147 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.549094 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.588933 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.609586 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.628559 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.648088 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.669648 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.713142 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.721015 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.728906 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.748419 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.767763 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.789353 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.809534 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.840415 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.849447 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.868695 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.889179 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.908698 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.929332 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.949137 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.968806 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Oct 06 10:00:40 crc kubenswrapper[4824]: I1006 10:00:40.988934 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Oct 06 10:00:41 crc kubenswrapper[4824]: I1006 10:00:41.009440 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Oct 06 10:00:41 crc kubenswrapper[4824]: I1006 10:00:41.028877 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Oct 06 10:00:41 crc kubenswrapper[4824]: I1006 10:00:41.049972 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Oct 06 10:00:41 crc kubenswrapper[4824]: I1006 10:00:41.069187 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Oct 06 10:00:41 crc kubenswrapper[4824]: I1006 10:00:41.089536 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Oct 06 10:00:41 crc kubenswrapper[4824]: I1006 10:00:41.109410 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Oct 06 10:00:41 crc kubenswrapper[4824]: I1006 10:00:41.129458 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Oct 06 10:00:41 crc kubenswrapper[4824]: I1006 10:00:41.148275 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Oct 06 10:00:41 crc kubenswrapper[4824]: I1006 10:00:41.168359 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Oct 06 10:00:41 crc kubenswrapper[4824]: I1006 10:00:41.190394 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Oct 06 10:00:41 crc kubenswrapper[4824]: I1006 10:00:41.209094 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Oct 06 10:00:41 crc kubenswrapper[4824]: I1006 10:00:41.229583 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Oct 06 10:00:41 crc kubenswrapper[4824]: I1006 10:00:41.249234 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Oct 06 10:00:41 crc kubenswrapper[4824]: I1006 10:00:41.269432 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Oct 06 10:00:41 crc kubenswrapper[4824]: I1006 10:00:41.288680 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Oct 06 10:00:41 crc kubenswrapper[4824]: I1006 10:00:41.308648 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Oct 06 10:00:41 crc kubenswrapper[4824]: I1006 10:00:41.329541 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Oct 06 10:00:41 crc kubenswrapper[4824]: I1006 10:00:41.349854 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Oct 06 10:00:41 crc kubenswrapper[4824]: I1006 10:00:41.369374 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Oct 06 10:00:41 crc kubenswrapper[4824]: I1006 10:00:41.386862 4824 request.go:700] Waited for 1.007040501s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-operator-lifecycle-manager/configmaps?fieldSelector=metadata.name%3Dopenshift-service-ca.crt&limit=500&resourceVersion=0 Oct 06 10:00:41 crc kubenswrapper[4824]: I1006 10:00:41.388945 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Oct 06 10:00:41 crc kubenswrapper[4824]: I1006 10:00:41.409092 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Oct 06 10:00:41 crc kubenswrapper[4824]: I1006 10:00:41.428125 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Oct 06 10:00:41 crc kubenswrapper[4824]: I1006 10:00:41.450215 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Oct 06 10:00:41 crc kubenswrapper[4824]: I1006 10:00:41.469731 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Oct 06 10:00:41 crc kubenswrapper[4824]: I1006 10:00:41.488543 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Oct 06 10:00:41 crc kubenswrapper[4824]: I1006 10:00:41.509297 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Oct 06 10:00:41 crc kubenswrapper[4824]: I1006 10:00:41.529201 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Oct 06 10:00:41 crc kubenswrapper[4824]: I1006 10:00:41.549130 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Oct 06 10:00:41 crc kubenswrapper[4824]: I1006 10:00:41.569586 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Oct 06 10:00:41 crc kubenswrapper[4824]: I1006 10:00:41.589318 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Oct 06 10:00:41 crc kubenswrapper[4824]: I1006 10:00:41.610323 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Oct 06 10:00:41 crc kubenswrapper[4824]: I1006 10:00:41.629054 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Oct 06 10:00:41 crc kubenswrapper[4824]: I1006 10:00:41.649490 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Oct 06 10:00:41 crc kubenswrapper[4824]: I1006 10:00:41.670017 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Oct 06 10:00:41 crc kubenswrapper[4824]: I1006 10:00:41.689668 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Oct 06 10:00:41 crc kubenswrapper[4824]: I1006 10:00:41.710276 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Oct 06 10:00:41 crc kubenswrapper[4824]: I1006 10:00:41.729089 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Oct 06 10:00:41 crc kubenswrapper[4824]: I1006 10:00:41.749064 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Oct 06 10:00:41 crc kubenswrapper[4824]: I1006 10:00:41.768213 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Oct 06 10:00:41 crc kubenswrapper[4824]: I1006 10:00:41.808788 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 06 10:00:41 crc kubenswrapper[4824]: I1006 10:00:41.829078 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 06 10:00:41 crc kubenswrapper[4824]: I1006 10:00:41.849614 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Oct 06 10:00:41 crc kubenswrapper[4824]: I1006 10:00:41.869199 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Oct 06 10:00:41 crc kubenswrapper[4824]: I1006 10:00:41.888600 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Oct 06 10:00:41 crc kubenswrapper[4824]: I1006 10:00:41.909232 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Oct 06 10:00:41 crc kubenswrapper[4824]: I1006 10:00:41.937251 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Oct 06 10:00:41 crc kubenswrapper[4824]: I1006 10:00:41.948352 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Oct 06 10:00:41 crc kubenswrapper[4824]: I1006 10:00:41.968631 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Oct 06 10:00:41 crc kubenswrapper[4824]: I1006 10:00:41.988859 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.009128 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.028655 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.050194 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.074417 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.089963 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.108170 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.129556 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.149533 4824 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.190243 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.205075 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jmr46\" (UniqueName: \"kubernetes.io/projected/171abefb-ea19-40fa-8a22-4b38956d5e4b-kube-api-access-jmr46\") pod \"downloads-7954f5f757-s4ht9\" (UID: \"171abefb-ea19-40fa-8a22-4b38956d5e4b\") " pod="openshift-console/downloads-7954f5f757-s4ht9" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.208890 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.229933 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.249007 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.271091 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.291860 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.309205 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.339473 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2qj4r\" (UniqueName: \"kubernetes.io/projected/dd50f0ac-ac97-413f-a350-7496681f7f0a-kube-api-access-2qj4r\") pod \"route-controller-manager-6576b87f9c-wk4bd\" (UID: \"dd50f0ac-ac97-413f-a350-7496681f7f0a\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wk4bd" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.339544 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ft4p7\" (UniqueName: \"kubernetes.io/projected/4d7b1b9b-33de-4fca-b629-6790c3a5cb7d-kube-api-access-ft4p7\") pod \"cluster-samples-operator-665b6dd947-hqhlr\" (UID: \"4d7b1b9b-33de-4fca-b629-6790c3a5cb7d\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-hqhlr" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.339588 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/7532dafc-fb61-4066-bb24-2470e4b2cfe6-metrics-tls\") pod \"dns-operator-744455d44c-6kn7b\" (UID: \"7532dafc-fb61-4066-bb24-2470e4b2cfe6\") " pod="openshift-dns-operator/dns-operator-744455d44c-6kn7b" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.339625 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/45682655-0e98-4a5e-b5e8-308f4d0f7d9a-config\") pod \"kube-apiserver-operator-766d6c64bb-prthm\" (UID: \"45682655-0e98-4a5e-b5e8-308f4d0f7d9a\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-prthm" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.339659 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-48kwg\" (UniqueName: \"kubernetes.io/projected/b349b0af-0ba8-409f-a915-7fb06b2e3774-kube-api-access-48kwg\") pod \"machine-api-operator-5694c8668f-fjdwv\" (UID: \"b349b0af-0ba8-409f-a915-7fb06b2e3774\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-fjdwv" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.339731 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8l8gd\" (UID: \"c2f6e84b-a28c-4ead-97f5-978b0639e5bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-8l8gd" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.339770 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/c2f6e84b-a28c-4ead-97f5-978b0639e5bf-ca-trust-extracted\") pod \"image-registry-697d97f7c8-8l8gd\" (UID: \"c2f6e84b-a28c-4ead-97f5-978b0639e5bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-8l8gd" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.339813 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4bceefc5-1905-48bd-b214-ab5824f0a302-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-mdv87\" (UID: \"4bceefc5-1905-48bd-b214-ab5824f0a302\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-mdv87" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.339847 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c8fc8c20-e06a-4039-80c0-82f84ea81391-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-sb9c8\" (UID: \"c8fc8c20-e06a-4039-80c0-82f84ea81391\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-sb9c8" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.339938 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/323631a8-107c-4fd0-8b57-04550f0fc7d1-service-ca-bundle\") pod \"authentication-operator-69f744f599-j2m8k\" (UID: \"323631a8-107c-4fd0-8b57-04550f0fc7d1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-j2m8k" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.340129 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4bceefc5-1905-48bd-b214-ab5824f0a302-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-mdv87\" (UID: \"4bceefc5-1905-48bd-b214-ab5824f0a302\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-mdv87" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.340172 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/a47e3600-79be-46c3-a1b8-6678d8da80c5-etcd-ca\") pod \"etcd-operator-b45778765-x4fdc\" (UID: \"a47e3600-79be-46c3-a1b8-6678d8da80c5\") " pod="openshift-etcd-operator/etcd-operator-b45778765-x4fdc" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.340244 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/dd50f0ac-ac97-413f-a350-7496681f7f0a-client-ca\") pod \"route-controller-manager-6576b87f9c-wk4bd\" (UID: \"dd50f0ac-ac97-413f-a350-7496681f7f0a\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wk4bd" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.340310 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/722b7a46-52a1-49a7-a053-803334dc5964-service-ca\") pod \"console-f9d7485db-fpxws\" (UID: \"722b7a46-52a1-49a7-a053-803334dc5964\") " pod="openshift-console/console-f9d7485db-fpxws" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.340356 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z8t7q\" (UniqueName: \"kubernetes.io/projected/323631a8-107c-4fd0-8b57-04550f0fc7d1-kube-api-access-z8t7q\") pod \"authentication-operator-69f744f599-j2m8k\" (UID: \"323631a8-107c-4fd0-8b57-04550f0fc7d1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-j2m8k" Oct 06 10:00:42 crc kubenswrapper[4824]: E1006 10:00:42.340402 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-06 10:00:42.840366571 +0000 UTC m=+152.204789472 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8l8gd" (UID: "c2f6e84b-a28c-4ead-97f5-978b0639e5bf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.340446 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/722b7a46-52a1-49a7-a053-803334dc5964-console-serving-cert\") pod \"console-f9d7485db-fpxws\" (UID: \"722b7a46-52a1-49a7-a053-803334dc5964\") " pod="openshift-console/console-f9d7485db-fpxws" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.340493 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9da4a453-8e05-4360-b74c-ad1c4706b005-config\") pod \"machine-approver-56656f9798-4hw5p\" (UID: \"9da4a453-8e05-4360-b74c-ad1c4706b005\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-4hw5p" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.340530 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2c9tb\" (UniqueName: \"kubernetes.io/projected/d21f4737-c582-41e9-846c-dd75ec173afd-kube-api-access-2c9tb\") pod \"cluster-image-registry-operator-dc59b4c8b-dklh2\" (UID: \"d21f4737-c582-41e9-846c-dd75ec173afd\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-dklh2" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.340562 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/911b7e6b-1df0-4ef3-822e-84925bd6bbdc-image-import-ca\") pod \"apiserver-76f77b778f-q76nt\" (UID: \"911b7e6b-1df0-4ef3-822e-84925bd6bbdc\") " pod="openshift-apiserver/apiserver-76f77b778f-q76nt" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.340600 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b6f7d3cb-418e-4f73-95a0-2477e37976c3-serving-cert\") pod \"controller-manager-879f6c89f-nqmr6\" (UID: \"b6f7d3cb-418e-4f73-95a0-2477e37976c3\") " pod="openshift-controller-manager/controller-manager-879f6c89f-nqmr6" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.340638 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/323631a8-107c-4fd0-8b57-04550f0fc7d1-config\") pod \"authentication-operator-69f744f599-j2m8k\" (UID: \"323631a8-107c-4fd0-8b57-04550f0fc7d1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-j2m8k" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.340672 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/a47e3600-79be-46c3-a1b8-6678d8da80c5-etcd-client\") pod \"etcd-operator-b45778765-x4fdc\" (UID: \"a47e3600-79be-46c3-a1b8-6678d8da80c5\") " pod="openshift-etcd-operator/etcd-operator-b45778765-x4fdc" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.340711 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gvtfv\" (UniqueName: \"kubernetes.io/projected/7532dafc-fb61-4066-bb24-2470e4b2cfe6-kube-api-access-gvtfv\") pod \"dns-operator-744455d44c-6kn7b\" (UID: \"7532dafc-fb61-4066-bb24-2470e4b2cfe6\") " pod="openshift-dns-operator/dns-operator-744455d44c-6kn7b" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.340744 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/22545b78-29bb-45aa-a98c-fd7ce9703239-trusted-ca\") pod \"console-operator-58897d9998-bjzzk\" (UID: \"22545b78-29bb-45aa-a98c-fd7ce9703239\") " pod="openshift-console-operator/console-operator-58897d9998-bjzzk" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.340779 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d6e255a4-d3e7-4cef-bf40-f98aa11b3e3b-metrics-certs\") pod \"router-default-5444994796-7qwsv\" (UID: \"d6e255a4-d3e7-4cef-bf40-f98aa11b3e3b\") " pod="openshift-ingress/router-default-5444994796-7qwsv" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.340814 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b6f7d3cb-418e-4f73-95a0-2477e37976c3-client-ca\") pod \"controller-manager-879f6c89f-nqmr6\" (UID: \"b6f7d3cb-418e-4f73-95a0-2477e37976c3\") " pod="openshift-controller-manager/controller-manager-879f6c89f-nqmr6" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.340853 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a47e3600-79be-46c3-a1b8-6678d8da80c5-config\") pod \"etcd-operator-b45778765-x4fdc\" (UID: \"a47e3600-79be-46c3-a1b8-6678d8da80c5\") " pod="openshift-etcd-operator/etcd-operator-b45778765-x4fdc" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.340885 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/722b7a46-52a1-49a7-a053-803334dc5964-console-config\") pod \"console-f9d7485db-fpxws\" (UID: \"722b7a46-52a1-49a7-a053-803334dc5964\") " pod="openshift-console/console-f9d7485db-fpxws" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.340921 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6t6pf\" (UniqueName: \"kubernetes.io/projected/9b1f3faf-72a1-44e2-814b-964902e78d3e-kube-api-access-6t6pf\") pod \"openshift-controller-manager-operator-756b6f6bc6-wlbrp\" (UID: \"9b1f3faf-72a1-44e2-814b-964902e78d3e\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wlbrp" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.340959 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d21f4737-c582-41e9-846c-dd75ec173afd-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-dklh2\" (UID: \"d21f4737-c582-41e9-846c-dd75ec173afd\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-dklh2" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.341043 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9b1f3faf-72a1-44e2-814b-964902e78d3e-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-wlbrp\" (UID: \"9b1f3faf-72a1-44e2-814b-964902e78d3e\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wlbrp" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.341084 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/c2f6e84b-a28c-4ead-97f5-978b0639e5bf-registry-certificates\") pod \"image-registry-697d97f7c8-8l8gd\" (UID: \"c2f6e84b-a28c-4ead-97f5-978b0639e5bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-8l8gd" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.341119 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dd50f0ac-ac97-413f-a350-7496681f7f0a-serving-cert\") pod \"route-controller-manager-6576b87f9c-wk4bd\" (UID: \"dd50f0ac-ac97-413f-a350-7496681f7f0a\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wk4bd" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.341156 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22545b78-29bb-45aa-a98c-fd7ce9703239-config\") pod \"console-operator-58897d9998-bjzzk\" (UID: \"22545b78-29bb-45aa-a98c-fd7ce9703239\") " pod="openshift-console-operator/console-operator-58897d9998-bjzzk" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.341192 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/911b7e6b-1df0-4ef3-822e-84925bd6bbdc-etcd-serving-ca\") pod \"apiserver-76f77b778f-q76nt\" (UID: \"911b7e6b-1df0-4ef3-822e-84925bd6bbdc\") " pod="openshift-apiserver/apiserver-76f77b778f-q76nt" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.341226 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/3bdcba24-670d-427a-bfd6-4372c9f944a2-bound-sa-token\") pod \"ingress-operator-5b745b69d9-7p7wb\" (UID: \"3bdcba24-670d-427a-bfd6-4372c9f944a2\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7p7wb" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.341256 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/911b7e6b-1df0-4ef3-822e-84925bd6bbdc-serving-cert\") pod \"apiserver-76f77b778f-q76nt\" (UID: \"911b7e6b-1df0-4ef3-822e-84925bd6bbdc\") " pod="openshift-apiserver/apiserver-76f77b778f-q76nt" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.341293 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/911b7e6b-1df0-4ef3-822e-84925bd6bbdc-trusted-ca-bundle\") pod \"apiserver-76f77b778f-q76nt\" (UID: \"911b7e6b-1df0-4ef3-822e-84925bd6bbdc\") " pod="openshift-apiserver/apiserver-76f77b778f-q76nt" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.341357 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3bdcba24-670d-427a-bfd6-4372c9f944a2-trusted-ca\") pod \"ingress-operator-5b745b69d9-7p7wb\" (UID: \"3bdcba24-670d-427a-bfd6-4372c9f944a2\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7p7wb" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.341406 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/911b7e6b-1df0-4ef3-822e-84925bd6bbdc-config\") pod \"apiserver-76f77b778f-q76nt\" (UID: \"911b7e6b-1df0-4ef3-822e-84925bd6bbdc\") " pod="openshift-apiserver/apiserver-76f77b778f-q76nt" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.341447 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/c2f6e84b-a28c-4ead-97f5-978b0639e5bf-installation-pull-secrets\") pod \"image-registry-697d97f7c8-8l8gd\" (UID: \"c2f6e84b-a28c-4ead-97f5-978b0639e5bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-8l8gd" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.341503 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jnc4f\" (UniqueName: \"kubernetes.io/projected/a47e3600-79be-46c3-a1b8-6678d8da80c5-kube-api-access-jnc4f\") pod \"etcd-operator-b45778765-x4fdc\" (UID: \"a47e3600-79be-46c3-a1b8-6678d8da80c5\") " pod="openshift-etcd-operator/etcd-operator-b45778765-x4fdc" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.341535 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/3bdcba24-670d-427a-bfd6-4372c9f944a2-metrics-tls\") pod \"ingress-operator-5b745b69d9-7p7wb\" (UID: \"3bdcba24-670d-427a-bfd6-4372c9f944a2\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7p7wb" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.341624 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nqdpp\" (UniqueName: \"kubernetes.io/projected/722b7a46-52a1-49a7-a053-803334dc5964-kube-api-access-nqdpp\") pod \"console-f9d7485db-fpxws\" (UID: \"722b7a46-52a1-49a7-a053-803334dc5964\") " pod="openshift-console/console-f9d7485db-fpxws" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.341670 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/911b7e6b-1df0-4ef3-822e-84925bd6bbdc-node-pullsecrets\") pod \"apiserver-76f77b778f-q76nt\" (UID: \"911b7e6b-1df0-4ef3-822e-84925bd6bbdc\") " pod="openshift-apiserver/apiserver-76f77b778f-q76nt" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.341709 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/911b7e6b-1df0-4ef3-822e-84925bd6bbdc-audit\") pod \"apiserver-76f77b778f-q76nt\" (UID: \"911b7e6b-1df0-4ef3-822e-84925bd6bbdc\") " pod="openshift-apiserver/apiserver-76f77b778f-q76nt" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.341753 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/323631a8-107c-4fd0-8b57-04550f0fc7d1-serving-cert\") pod \"authentication-operator-69f744f599-j2m8k\" (UID: \"323631a8-107c-4fd0-8b57-04550f0fc7d1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-j2m8k" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.341789 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dd50f0ac-ac97-413f-a350-7496681f7f0a-config\") pod \"route-controller-manager-6576b87f9c-wk4bd\" (UID: \"dd50f0ac-ac97-413f-a350-7496681f7f0a\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wk4bd" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.341823 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d21f4737-c582-41e9-846c-dd75ec173afd-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-dklh2\" (UID: \"d21f4737-c582-41e9-846c-dd75ec173afd\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-dklh2" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.341935 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-66bhb\" (UniqueName: \"kubernetes.io/projected/c2f6e84b-a28c-4ead-97f5-978b0639e5bf-kube-api-access-66bhb\") pod \"image-registry-697d97f7c8-8l8gd\" (UID: \"c2f6e84b-a28c-4ead-97f5-978b0639e5bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-8l8gd" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.341974 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b6f7d3cb-418e-4f73-95a0-2477e37976c3-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-nqmr6\" (UID: \"b6f7d3cb-418e-4f73-95a0-2477e37976c3\") " pod="openshift-controller-manager/controller-manager-879f6c89f-nqmr6" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.342053 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/c2f6e84b-a28c-4ead-97f5-978b0639e5bf-registry-tls\") pod \"image-registry-697d97f7c8-8l8gd\" (UID: \"c2f6e84b-a28c-4ead-97f5-978b0639e5bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-8l8gd" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.342091 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a47e3600-79be-46c3-a1b8-6678d8da80c5-serving-cert\") pod \"etcd-operator-b45778765-x4fdc\" (UID: \"a47e3600-79be-46c3-a1b8-6678d8da80c5\") " pod="openshift-etcd-operator/etcd-operator-b45778765-x4fdc" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.342140 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/722b7a46-52a1-49a7-a053-803334dc5964-trusted-ca-bundle\") pod \"console-f9d7485db-fpxws\" (UID: \"722b7a46-52a1-49a7-a053-803334dc5964\") " pod="openshift-console/console-f9d7485db-fpxws" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.342282 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/22545b78-29bb-45aa-a98c-fd7ce9703239-serving-cert\") pod \"console-operator-58897d9998-bjzzk\" (UID: \"22545b78-29bb-45aa-a98c-fd7ce9703239\") " pod="openshift-console-operator/console-operator-58897d9998-bjzzk" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.342346 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-57vkp\" (UniqueName: \"kubernetes.io/projected/911b7e6b-1df0-4ef3-822e-84925bd6bbdc-kube-api-access-57vkp\") pod \"apiserver-76f77b778f-q76nt\" (UID: \"911b7e6b-1df0-4ef3-822e-84925bd6bbdc\") " pod="openshift-apiserver/apiserver-76f77b778f-q76nt" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.342410 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4bceefc5-1905-48bd-b214-ab5824f0a302-config\") pod \"kube-controller-manager-operator-78b949d7b-mdv87\" (UID: \"4bceefc5-1905-48bd-b214-ab5824f0a302\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-mdv87" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.342476 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9b1f3faf-72a1-44e2-814b-964902e78d3e-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-wlbrp\" (UID: \"9b1f3faf-72a1-44e2-814b-964902e78d3e\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wlbrp" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.342582 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7c497a24-d8cb-4560-993f-fc73b3af8c04-serving-cert\") pod \"openshift-config-operator-7777fb866f-v6gnw\" (UID: \"7c497a24-d8cb-4560-993f-fc73b3af8c04\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-v6gnw" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.342669 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-94hbm\" (UniqueName: \"kubernetes.io/projected/7c497a24-d8cb-4560-993f-fc73b3af8c04-kube-api-access-94hbm\") pod \"openshift-config-operator-7777fb866f-v6gnw\" (UID: \"7c497a24-d8cb-4560-993f-fc73b3af8c04\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-v6gnw" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.342778 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fmfcl\" (UniqueName: \"kubernetes.io/projected/c8fc8c20-e06a-4039-80c0-82f84ea81391-kube-api-access-fmfcl\") pod \"openshift-apiserver-operator-796bbdcf4f-sb9c8\" (UID: \"c8fc8c20-e06a-4039-80c0-82f84ea81391\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-sb9c8" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.342825 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/b349b0af-0ba8-409f-a915-7fb06b2e3774-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-fjdwv\" (UID: \"b349b0af-0ba8-409f-a915-7fb06b2e3774\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-fjdwv" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.342865 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/a47e3600-79be-46c3-a1b8-6678d8da80c5-etcd-service-ca\") pod \"etcd-operator-b45778765-x4fdc\" (UID: \"a47e3600-79be-46c3-a1b8-6678d8da80c5\") " pod="openshift-etcd-operator/etcd-operator-b45778765-x4fdc" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.342901 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4lbvd\" (UniqueName: \"kubernetes.io/projected/3bdcba24-670d-427a-bfd6-4372c9f944a2-kube-api-access-4lbvd\") pod \"ingress-operator-5b745b69d9-7p7wb\" (UID: \"3bdcba24-670d-427a-bfd6-4372c9f944a2\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7p7wb" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.342935 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/722b7a46-52a1-49a7-a053-803334dc5964-oauth-serving-cert\") pod \"console-f9d7485db-fpxws\" (UID: \"722b7a46-52a1-49a7-a053-803334dc5964\") " pod="openshift-console/console-f9d7485db-fpxws" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.342967 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/911b7e6b-1df0-4ef3-822e-84925bd6bbdc-etcd-client\") pod \"apiserver-76f77b778f-q76nt\" (UID: \"911b7e6b-1df0-4ef3-822e-84925bd6bbdc\") " pod="openshift-apiserver/apiserver-76f77b778f-q76nt" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.343031 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/4d7b1b9b-33de-4fca-b629-6790c3a5cb7d-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-hqhlr\" (UID: \"4d7b1b9b-33de-4fca-b629-6790c3a5cb7d\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-hqhlr" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.343063 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/722b7a46-52a1-49a7-a053-803334dc5964-console-oauth-config\") pod \"console-f9d7485db-fpxws\" (UID: \"722b7a46-52a1-49a7-a053-803334dc5964\") " pod="openshift-console/console-f9d7485db-fpxws" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.343100 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/9da4a453-8e05-4360-b74c-ad1c4706b005-auth-proxy-config\") pod \"machine-approver-56656f9798-4hw5p\" (UID: \"9da4a453-8e05-4360-b74c-ad1c4706b005\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-4hw5p" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.343133 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z44hr\" (UniqueName: \"kubernetes.io/projected/9da4a453-8e05-4360-b74c-ad1c4706b005-kube-api-access-z44hr\") pod \"machine-approver-56656f9798-4hw5p\" (UID: \"9da4a453-8e05-4360-b74c-ad1c4706b005\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-4hw5p" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.343167 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b349b0af-0ba8-409f-a915-7fb06b2e3774-config\") pod \"machine-api-operator-5694c8668f-fjdwv\" (UID: \"b349b0af-0ba8-409f-a915-7fb06b2e3774\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-fjdwv" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.343200 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/911b7e6b-1df0-4ef3-822e-84925bd6bbdc-audit-dir\") pod \"apiserver-76f77b778f-q76nt\" (UID: \"911b7e6b-1df0-4ef3-822e-84925bd6bbdc\") " pod="openshift-apiserver/apiserver-76f77b778f-q76nt" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.343239 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/7c497a24-d8cb-4560-993f-fc73b3af8c04-available-featuregates\") pod \"openshift-config-operator-7777fb866f-v6gnw\" (UID: \"7c497a24-d8cb-4560-993f-fc73b3af8c04\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-v6gnw" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.343275 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/d6e255a4-d3e7-4cef-bf40-f98aa11b3e3b-default-certificate\") pod \"router-default-5444994796-7qwsv\" (UID: \"d6e255a4-d3e7-4cef-bf40-f98aa11b3e3b\") " pod="openshift-ingress/router-default-5444994796-7qwsv" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.343315 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/323631a8-107c-4fd0-8b57-04550f0fc7d1-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-j2m8k\" (UID: \"323631a8-107c-4fd0-8b57-04550f0fc7d1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-j2m8k" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.343350 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/45682655-0e98-4a5e-b5e8-308f4d0f7d9a-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-prthm\" (UID: \"45682655-0e98-4a5e-b5e8-308f4d0f7d9a\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-prthm" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.343386 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/d21f4737-c582-41e9-846c-dd75ec173afd-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-dklh2\" (UID: \"d21f4737-c582-41e9-846c-dd75ec173afd\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-dklh2" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.343423 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d6e255a4-d3e7-4cef-bf40-f98aa11b3e3b-service-ca-bundle\") pod \"router-default-5444994796-7qwsv\" (UID: \"d6e255a4-d3e7-4cef-bf40-f98aa11b3e3b\") " pod="openshift-ingress/router-default-5444994796-7qwsv" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.343464 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/b349b0af-0ba8-409f-a915-7fb06b2e3774-images\") pod \"machine-api-operator-5694c8668f-fjdwv\" (UID: \"b349b0af-0ba8-409f-a915-7fb06b2e3774\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-fjdwv" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.343498 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l2n5j\" (UniqueName: \"kubernetes.io/projected/d6e255a4-d3e7-4cef-bf40-f98aa11b3e3b-kube-api-access-l2n5j\") pod \"router-default-5444994796-7qwsv\" (UID: \"d6e255a4-d3e7-4cef-bf40-f98aa11b3e3b\") " pod="openshift-ingress/router-default-5444994796-7qwsv" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.343540 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vz27c\" (UniqueName: \"kubernetes.io/projected/22545b78-29bb-45aa-a98c-fd7ce9703239-kube-api-access-vz27c\") pod \"console-operator-58897d9998-bjzzk\" (UID: \"22545b78-29bb-45aa-a98c-fd7ce9703239\") " pod="openshift-console-operator/console-operator-58897d9998-bjzzk" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.343576 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/d6e255a4-d3e7-4cef-bf40-f98aa11b3e3b-stats-auth\") pod \"router-default-5444994796-7qwsv\" (UID: \"d6e255a4-d3e7-4cef-bf40-f98aa11b3e3b\") " pod="openshift-ingress/router-default-5444994796-7qwsv" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.343614 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b6f7d3cb-418e-4f73-95a0-2477e37976c3-config\") pod \"controller-manager-879f6c89f-nqmr6\" (UID: \"b6f7d3cb-418e-4f73-95a0-2477e37976c3\") " pod="openshift-controller-manager/controller-manager-879f6c89f-nqmr6" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.343703 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c2f6e84b-a28c-4ead-97f5-978b0639e5bf-trusted-ca\") pod \"image-registry-697d97f7c8-8l8gd\" (UID: \"c2f6e84b-a28c-4ead-97f5-978b0639e5bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-8l8gd" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.343754 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c2f6e84b-a28c-4ead-97f5-978b0639e5bf-bound-sa-token\") pod \"image-registry-697d97f7c8-8l8gd\" (UID: \"c2f6e84b-a28c-4ead-97f5-978b0639e5bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-8l8gd" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.343784 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c8fc8c20-e06a-4039-80c0-82f84ea81391-config\") pod \"openshift-apiserver-operator-796bbdcf4f-sb9c8\" (UID: \"c8fc8c20-e06a-4039-80c0-82f84ea81391\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-sb9c8" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.343852 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/45682655-0e98-4a5e-b5e8-308f4d0f7d9a-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-prthm\" (UID: \"45682655-0e98-4a5e-b5e8-308f4d0f7d9a\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-prthm" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.343903 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/9da4a453-8e05-4360-b74c-ad1c4706b005-machine-approver-tls\") pod \"machine-approver-56656f9798-4hw5p\" (UID: \"9da4a453-8e05-4360-b74c-ad1c4706b005\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-4hw5p" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.343947 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2h2fc\" (UniqueName: \"kubernetes.io/projected/b6f7d3cb-418e-4f73-95a0-2477e37976c3-kube-api-access-2h2fc\") pod \"controller-manager-879f6c89f-nqmr6\" (UID: \"b6f7d3cb-418e-4f73-95a0-2477e37976c3\") " pod="openshift-controller-manager/controller-manager-879f6c89f-nqmr6" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.344034 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/911b7e6b-1df0-4ef3-822e-84925bd6bbdc-encryption-config\") pod \"apiserver-76f77b778f-q76nt\" (UID: \"911b7e6b-1df0-4ef3-822e-84925bd6bbdc\") " pod="openshift-apiserver/apiserver-76f77b778f-q76nt" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.419809 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-s4ht9" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.444662 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 10:00:42 crc kubenswrapper[4824]: E1006 10:00:42.445047 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-06 10:00:42.944824064 +0000 UTC m=+152.309246975 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.445153 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t2gcn\" (UniqueName: \"kubernetes.io/projected/d0546787-f9de-4b06-ab4f-0e1632d36230-kube-api-access-t2gcn\") pod \"catalog-operator-68c6474976-2xznh\" (UID: \"d0546787-f9de-4b06-ab4f-0e1632d36230\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-2xznh" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.445267 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/982129a4-11b4-4024-afb0-ac6024165c7a-profile-collector-cert\") pod \"olm-operator-6b444d44fb-6tt2t\" (UID: \"982129a4-11b4-4024-afb0-ac6024165c7a\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6tt2t" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.445346 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/2a0286c5-9866-4a1c-9855-49cea5d04d69-proxy-tls\") pod \"machine-config-operator-74547568cd-vvtbh\" (UID: \"2a0286c5-9866-4a1c-9855-49cea5d04d69\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vvtbh" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.445407 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-66bhb\" (UniqueName: \"kubernetes.io/projected/c2f6e84b-a28c-4ead-97f5-978b0639e5bf-kube-api-access-66bhb\") pod \"image-registry-697d97f7c8-8l8gd\" (UID: \"c2f6e84b-a28c-4ead-97f5-978b0639e5bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-8l8gd" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.445508 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b6f7d3cb-418e-4f73-95a0-2477e37976c3-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-nqmr6\" (UID: \"b6f7d3cb-418e-4f73-95a0-2477e37976c3\") " pod="openshift-controller-manager/controller-manager-879f6c89f-nqmr6" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.445612 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/af8fbf5a-f90e-4d22-ae65-05d8d8253308-audit-dir\") pod \"oauth-openshift-558db77b4-8z68k\" (UID: \"af8fbf5a-f90e-4d22-ae65-05d8d8253308\") " pod="openshift-authentication/oauth-openshift-558db77b4-8z68k" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.445684 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/c2f6e84b-a28c-4ead-97f5-978b0639e5bf-registry-tls\") pod \"image-registry-697d97f7c8-8l8gd\" (UID: \"c2f6e84b-a28c-4ead-97f5-978b0639e5bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-8l8gd" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.445747 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a47e3600-79be-46c3-a1b8-6678d8da80c5-serving-cert\") pod \"etcd-operator-b45778765-x4fdc\" (UID: \"a47e3600-79be-46c3-a1b8-6678d8da80c5\") " pod="openshift-etcd-operator/etcd-operator-b45778765-x4fdc" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.445811 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/722b7a46-52a1-49a7-a053-803334dc5964-trusted-ca-bundle\") pod \"console-f9d7485db-fpxws\" (UID: \"722b7a46-52a1-49a7-a053-803334dc5964\") " pod="openshift-console/console-f9d7485db-fpxws" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.445869 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-57vkp\" (UniqueName: \"kubernetes.io/projected/911b7e6b-1df0-4ef3-822e-84925bd6bbdc-kube-api-access-57vkp\") pod \"apiserver-76f77b778f-q76nt\" (UID: \"911b7e6b-1df0-4ef3-822e-84925bd6bbdc\") " pod="openshift-apiserver/apiserver-76f77b778f-q76nt" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.445937 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/d0546787-f9de-4b06-ab4f-0e1632d36230-profile-collector-cert\") pod \"catalog-operator-68c6474976-2xznh\" (UID: \"d0546787-f9de-4b06-ab4f-0e1632d36230\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-2xznh" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.446030 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/af8fbf5a-f90e-4d22-ae65-05d8d8253308-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-8z68k\" (UID: \"af8fbf5a-f90e-4d22-ae65-05d8d8253308\") " pod="openshift-authentication/oauth-openshift-558db77b4-8z68k" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.446094 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/a1d6c12e-ff5a-4687-80ca-5699dd1bb9fe-mountpoint-dir\") pod \"csi-hostpathplugin-nk67f\" (UID: \"a1d6c12e-ff5a-4687-80ca-5699dd1bb9fe\") " pod="hostpath-provisioner/csi-hostpathplugin-nk67f" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.446162 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/af8fbf5a-f90e-4d22-ae65-05d8d8253308-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-8z68k\" (UID: \"af8fbf5a-f90e-4d22-ae65-05d8d8253308\") " pod="openshift-authentication/oauth-openshift-558db77b4-8z68k" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.446227 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fmfcl\" (UniqueName: \"kubernetes.io/projected/c8fc8c20-e06a-4039-80c0-82f84ea81391-kube-api-access-fmfcl\") pod \"openshift-apiserver-operator-796bbdcf4f-sb9c8\" (UID: \"c8fc8c20-e06a-4039-80c0-82f84ea81391\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-sb9c8" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.446279 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d40e2672-6c82-45e6-a292-21615712032d-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-wp6kv\" (UID: \"d40e2672-6c82-45e6-a292-21615712032d\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-wp6kv" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.446339 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/911b7e6b-1df0-4ef3-822e-84925bd6bbdc-etcd-client\") pod \"apiserver-76f77b778f-q76nt\" (UID: \"911b7e6b-1df0-4ef3-822e-84925bd6bbdc\") " pod="openshift-apiserver/apiserver-76f77b778f-q76nt" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.446387 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/ffceea22-78f2-4032-a23a-71b4fb7ea83e-tmpfs\") pod \"packageserver-d55dfcdfc-s8j95\" (UID: \"ffceea22-78f2-4032-a23a-71b4fb7ea83e\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-s8j95" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.446443 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d40e2672-6c82-45e6-a292-21615712032d-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-wp6kv\" (UID: \"d40e2672-6c82-45e6-a292-21615712032d\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-wp6kv" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.446496 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/a1d6c12e-ff5a-4687-80ca-5699dd1bb9fe-plugins-dir\") pod \"csi-hostpathplugin-nk67f\" (UID: \"a1d6c12e-ff5a-4687-80ca-5699dd1bb9fe\") " pod="hostpath-provisioner/csi-hostpathplugin-nk67f" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.446549 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/4d7b1b9b-33de-4fca-b629-6790c3a5cb7d-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-hqhlr\" (UID: \"4d7b1b9b-33de-4fca-b629-6790c3a5cb7d\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-hqhlr" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.446607 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/722b7a46-52a1-49a7-a053-803334dc5964-console-oauth-config\") pod \"console-f9d7485db-fpxws\" (UID: \"722b7a46-52a1-49a7-a053-803334dc5964\") " pod="openshift-console/console-f9d7485db-fpxws" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.446664 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-99zcx\" (UniqueName: \"kubernetes.io/projected/46e11b34-d0dd-4ea6-bed7-66eb707c7f92-kube-api-access-99zcx\") pod \"service-ca-operator-777779d784-5blgj\" (UID: \"46e11b34-d0dd-4ea6-bed7-66eb707c7f92\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-5blgj" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.446731 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b349b0af-0ba8-409f-a915-7fb06b2e3774-config\") pod \"machine-api-operator-5694c8668f-fjdwv\" (UID: \"b349b0af-0ba8-409f-a915-7fb06b2e3774\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-fjdwv" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.446780 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/d6e255a4-d3e7-4cef-bf40-f98aa11b3e3b-default-certificate\") pod \"router-default-5444994796-7qwsv\" (UID: \"d6e255a4-d3e7-4cef-bf40-f98aa11b3e3b\") " pod="openshift-ingress/router-default-5444994796-7qwsv" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.446825 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/323631a8-107c-4fd0-8b57-04550f0fc7d1-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-j2m8k\" (UID: \"323631a8-107c-4fd0-8b57-04550f0fc7d1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-j2m8k" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.446884 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w6ks8\" (UniqueName: \"kubernetes.io/projected/ffceea22-78f2-4032-a23a-71b4fb7ea83e-kube-api-access-w6ks8\") pod \"packageserver-d55dfcdfc-s8j95\" (UID: \"ffceea22-78f2-4032-a23a-71b4fb7ea83e\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-s8j95" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.446934 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xshbg\" (UniqueName: \"kubernetes.io/projected/92937768-de56-4cda-816c-9d1be4c07f0f-kube-api-access-xshbg\") pod \"migrator-59844c95c7-qpxrp\" (UID: \"92937768-de56-4cda-816c-9d1be4c07f0f\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-qpxrp" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.447021 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/2fa5ee3e-29f4-43b2-a154-3c988754897a-etcd-client\") pod \"apiserver-7bbb656c7d-zzrfg\" (UID: \"2fa5ee3e-29f4-43b2-a154-3c988754897a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zzrfg" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.447092 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l2n5j\" (UniqueName: \"kubernetes.io/projected/d6e255a4-d3e7-4cef-bf40-f98aa11b3e3b-kube-api-access-l2n5j\") pod \"router-default-5444994796-7qwsv\" (UID: \"d6e255a4-d3e7-4cef-bf40-f98aa11b3e3b\") " pod="openshift-ingress/router-default-5444994796-7qwsv" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.447155 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/744def19-0006-4ab4-9d45-8a2a39bd410a-node-bootstrap-token\") pod \"machine-config-server-fdbvz\" (UID: \"744def19-0006-4ab4-9d45-8a2a39bd410a\") " pod="openshift-machine-config-operator/machine-config-server-fdbvz" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.447213 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/46e11b34-d0dd-4ea6-bed7-66eb707c7f92-config\") pod \"service-ca-operator-777779d784-5blgj\" (UID: \"46e11b34-d0dd-4ea6-bed7-66eb707c7f92\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-5blgj" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.447279 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vz27c\" (UniqueName: \"kubernetes.io/projected/22545b78-29bb-45aa-a98c-fd7ce9703239-kube-api-access-vz27c\") pod \"console-operator-58897d9998-bjzzk\" (UID: \"22545b78-29bb-45aa-a98c-fd7ce9703239\") " pod="openshift-console-operator/console-operator-58897d9998-bjzzk" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.447338 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/d6e255a4-d3e7-4cef-bf40-f98aa11b3e3b-stats-auth\") pod \"router-default-5444994796-7qwsv\" (UID: \"d6e255a4-d3e7-4cef-bf40-f98aa11b3e3b\") " pod="openshift-ingress/router-default-5444994796-7qwsv" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.447394 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/2a76ccaa-9249-46ba-bbc5-dbbbc33b06bb-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-fn7dj\" (UID: \"2a76ccaa-9249-46ba-bbc5-dbbbc33b06bb\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-fn7dj" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.447460 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c8fc8c20-e06a-4039-80c0-82f84ea81391-config\") pod \"openshift-apiserver-operator-796bbdcf4f-sb9c8\" (UID: \"c8fc8c20-e06a-4039-80c0-82f84ea81391\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-sb9c8" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.447517 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/9da4a453-8e05-4360-b74c-ad1c4706b005-machine-approver-tls\") pod \"machine-approver-56656f9798-4hw5p\" (UID: \"9da4a453-8e05-4360-b74c-ad1c4706b005\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-4hw5p" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.447539 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b6f7d3cb-418e-4f73-95a0-2477e37976c3-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-nqmr6\" (UID: \"b6f7d3cb-418e-4f73-95a0-2477e37976c3\") " pod="openshift-controller-manager/controller-manager-879f6c89f-nqmr6" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.447574 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2h2fc\" (UniqueName: \"kubernetes.io/projected/b6f7d3cb-418e-4f73-95a0-2477e37976c3-kube-api-access-2h2fc\") pod \"controller-manager-879f6c89f-nqmr6\" (UID: \"b6f7d3cb-418e-4f73-95a0-2477e37976c3\") " pod="openshift-controller-manager/controller-manager-879f6c89f-nqmr6" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.447628 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jgp2m\" (UniqueName: \"kubernetes.io/projected/2a0286c5-9866-4a1c-9855-49cea5d04d69-kube-api-access-jgp2m\") pod \"machine-config-operator-74547568cd-vvtbh\" (UID: \"2a0286c5-9866-4a1c-9855-49cea5d04d69\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vvtbh" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.447679 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/d0546787-f9de-4b06-ab4f-0e1632d36230-srv-cert\") pod \"catalog-operator-68c6474976-2xznh\" (UID: \"d0546787-f9de-4b06-ab4f-0e1632d36230\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-2xznh" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.447733 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2qj4r\" (UniqueName: \"kubernetes.io/projected/dd50f0ac-ac97-413f-a350-7496681f7f0a-kube-api-access-2qj4r\") pod \"route-controller-manager-6576b87f9c-wk4bd\" (UID: \"dd50f0ac-ac97-413f-a350-7496681f7f0a\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wk4bd" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.447794 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ft4p7\" (UniqueName: \"kubernetes.io/projected/4d7b1b9b-33de-4fca-b629-6790c3a5cb7d-kube-api-access-ft4p7\") pod \"cluster-samples-operator-665b6dd947-hqhlr\" (UID: \"4d7b1b9b-33de-4fca-b629-6790c3a5cb7d\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-hqhlr" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.447854 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/7532dafc-fb61-4066-bb24-2470e4b2cfe6-metrics-tls\") pod \"dns-operator-744455d44c-6kn7b\" (UID: \"7532dafc-fb61-4066-bb24-2470e4b2cfe6\") " pod="openshift-dns-operator/dns-operator-744455d44c-6kn7b" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.447912 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/45682655-0e98-4a5e-b5e8-308f4d0f7d9a-config\") pod \"kube-apiserver-operator-766d6c64bb-prthm\" (UID: \"45682655-0e98-4a5e-b5e8-308f4d0f7d9a\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-prthm" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.448024 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8l8gd\" (UID: \"c2f6e84b-a28c-4ead-97f5-978b0639e5bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-8l8gd" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.448093 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4bceefc5-1905-48bd-b214-ab5824f0a302-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-mdv87\" (UID: \"4bceefc5-1905-48bd-b214-ab5824f0a302\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-mdv87" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.448146 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e81889bb-3225-458e-bce5-cfbafe6f4f92-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-xr4mg\" (UID: \"e81889bb-3225-458e-bce5-cfbafe6f4f92\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-xr4mg" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.448198 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/323631a8-107c-4fd0-8b57-04550f0fc7d1-service-ca-bundle\") pod \"authentication-operator-69f744f599-j2m8k\" (UID: \"323631a8-107c-4fd0-8b57-04550f0fc7d1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-j2m8k" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.448257 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4bceefc5-1905-48bd-b214-ab5824f0a302-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-mdv87\" (UID: \"4bceefc5-1905-48bd-b214-ab5824f0a302\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-mdv87" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.448313 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/a47e3600-79be-46c3-a1b8-6678d8da80c5-etcd-ca\") pod \"etcd-operator-b45778765-x4fdc\" (UID: \"a47e3600-79be-46c3-a1b8-6678d8da80c5\") " pod="openshift-etcd-operator/etcd-operator-b45778765-x4fdc" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.448359 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/722b7a46-52a1-49a7-a053-803334dc5964-service-ca\") pod \"console-f9d7485db-fpxws\" (UID: \"722b7a46-52a1-49a7-a053-803334dc5964\") " pod="openshift-console/console-f9d7485db-fpxws" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.448409 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l2lvw\" (UniqueName: \"kubernetes.io/projected/883a0642-0302-4fc5-b361-e25f1aba0878-kube-api-access-l2lvw\") pod \"ingress-canary-vpg8w\" (UID: \"883a0642-0302-4fc5-b361-e25f1aba0878\") " pod="openshift-ingress-canary/ingress-canary-vpg8w" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.448471 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/af8fbf5a-f90e-4d22-ae65-05d8d8253308-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-8z68k\" (UID: \"af8fbf5a-f90e-4d22-ae65-05d8d8253308\") " pod="openshift-authentication/oauth-openshift-558db77b4-8z68k" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.448531 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/323631a8-107c-4fd0-8b57-04550f0fc7d1-config\") pod \"authentication-operator-69f744f599-j2m8k\" (UID: \"323631a8-107c-4fd0-8b57-04550f0fc7d1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-j2m8k" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.448587 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/a47e3600-79be-46c3-a1b8-6678d8da80c5-etcd-client\") pod \"etcd-operator-b45778765-x4fdc\" (UID: \"a47e3600-79be-46c3-a1b8-6678d8da80c5\") " pod="openshift-etcd-operator/etcd-operator-b45778765-x4fdc" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.448645 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d6e255a4-d3e7-4cef-bf40-f98aa11b3e3b-metrics-certs\") pod \"router-default-5444994796-7qwsv\" (UID: \"d6e255a4-d3e7-4cef-bf40-f98aa11b3e3b\") " pod="openshift-ingress/router-default-5444994796-7qwsv" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.448701 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a47e3600-79be-46c3-a1b8-6678d8da80c5-config\") pod \"etcd-operator-b45778765-x4fdc\" (UID: \"a47e3600-79be-46c3-a1b8-6678d8da80c5\") " pod="openshift-etcd-operator/etcd-operator-b45778765-x4fdc" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.448749 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2fa5ee3e-29f4-43b2-a154-3c988754897a-audit-dir\") pod \"apiserver-7bbb656c7d-zzrfg\" (UID: \"2fa5ee3e-29f4-43b2-a154-3c988754897a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zzrfg" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.448800 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/a1d6c12e-ff5a-4687-80ca-5699dd1bb9fe-registration-dir\") pod \"csi-hostpathplugin-nk67f\" (UID: \"a1d6c12e-ff5a-4687-80ca-5699dd1bb9fe\") " pod="hostpath-provisioner/csi-hostpathplugin-nk67f" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.448858 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9b1f3faf-72a1-44e2-814b-964902e78d3e-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-wlbrp\" (UID: \"9b1f3faf-72a1-44e2-814b-964902e78d3e\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wlbrp" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.448919 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/26536a6b-3639-46ad-b024-5e21b7118fe7-config-volume\") pod \"collect-profiles-29329080-9kpdv\" (UID: \"26536a6b-3639-46ad-b024-5e21b7118fe7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329080-9kpdv" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.449016 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22545b78-29bb-45aa-a98c-fd7ce9703239-config\") pod \"console-operator-58897d9998-bjzzk\" (UID: \"22545b78-29bb-45aa-a98c-fd7ce9703239\") " pod="openshift-console-operator/console-operator-58897d9998-bjzzk" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.449079 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/911b7e6b-1df0-4ef3-822e-84925bd6bbdc-etcd-serving-ca\") pod \"apiserver-76f77b778f-q76nt\" (UID: \"911b7e6b-1df0-4ef3-822e-84925bd6bbdc\") " pod="openshift-apiserver/apiserver-76f77b778f-q76nt" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.449149 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dd50f0ac-ac97-413f-a350-7496681f7f0a-serving-cert\") pod \"route-controller-manager-6576b87f9c-wk4bd\" (UID: \"dd50f0ac-ac97-413f-a350-7496681f7f0a\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wk4bd" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.449210 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/3bdcba24-670d-427a-bfd6-4372c9f944a2-bound-sa-token\") pod \"ingress-operator-5b745b69d9-7p7wb\" (UID: \"3bdcba24-670d-427a-bfd6-4372c9f944a2\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7p7wb" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.449259 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/911b7e6b-1df0-4ef3-822e-84925bd6bbdc-trusted-ca-bundle\") pod \"apiserver-76f77b778f-q76nt\" (UID: \"911b7e6b-1df0-4ef3-822e-84925bd6bbdc\") " pod="openshift-apiserver/apiserver-76f77b778f-q76nt" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.449322 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/2fa5ee3e-29f4-43b2-a154-3c988754897a-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-zzrfg\" (UID: \"2fa5ee3e-29f4-43b2-a154-3c988754897a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zzrfg" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.449379 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6km6j\" (UniqueName: \"kubernetes.io/projected/2fa5ee3e-29f4-43b2-a154-3c988754897a-kube-api-access-6km6j\") pod \"apiserver-7bbb656c7d-zzrfg\" (UID: \"2fa5ee3e-29f4-43b2-a154-3c988754897a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zzrfg" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.449441 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/af8fbf5a-f90e-4d22-ae65-05d8d8253308-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-8z68k\" (UID: \"af8fbf5a-f90e-4d22-ae65-05d8d8253308\") " pod="openshift-authentication/oauth-openshift-558db77b4-8z68k" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.451029 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/2a0286c5-9866-4a1c-9855-49cea5d04d69-images\") pod \"machine-config-operator-74547568cd-vvtbh\" (UID: \"2a0286c5-9866-4a1c-9855-49cea5d04d69\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vvtbh" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.451105 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3bdcba24-670d-427a-bfd6-4372c9f944a2-trusted-ca\") pod \"ingress-operator-5b745b69d9-7p7wb\" (UID: \"3bdcba24-670d-427a-bfd6-4372c9f944a2\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7p7wb" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.451162 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/c2f6e84b-a28c-4ead-97f5-978b0639e5bf-installation-pull-secrets\") pod \"image-registry-697d97f7c8-8l8gd\" (UID: \"c2f6e84b-a28c-4ead-97f5-978b0639e5bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-8l8gd" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.451215 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nqdpp\" (UniqueName: \"kubernetes.io/projected/722b7a46-52a1-49a7-a053-803334dc5964-kube-api-access-nqdpp\") pod \"console-f9d7485db-fpxws\" (UID: \"722b7a46-52a1-49a7-a053-803334dc5964\") " pod="openshift-console/console-f9d7485db-fpxws" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.451394 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/911b7e6b-1df0-4ef3-822e-84925bd6bbdc-node-pullsecrets\") pod \"apiserver-76f77b778f-q76nt\" (UID: \"911b7e6b-1df0-4ef3-822e-84925bd6bbdc\") " pod="openshift-apiserver/apiserver-76f77b778f-q76nt" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.451467 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/911b7e6b-1df0-4ef3-822e-84925bd6bbdc-audit\") pod \"apiserver-76f77b778f-q76nt\" (UID: \"911b7e6b-1df0-4ef3-822e-84925bd6bbdc\") " pod="openshift-apiserver/apiserver-76f77b778f-q76nt" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.451526 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/26536a6b-3639-46ad-b024-5e21b7118fe7-secret-volume\") pod \"collect-profiles-29329080-9kpdv\" (UID: \"26536a6b-3639-46ad-b024-5e21b7118fe7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329080-9kpdv" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.451625 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/af8fbf5a-f90e-4d22-ae65-05d8d8253308-audit-policies\") pod \"oauth-openshift-558db77b4-8z68k\" (UID: \"af8fbf5a-f90e-4d22-ae65-05d8d8253308\") " pod="openshift-authentication/oauth-openshift-558db77b4-8z68k" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.451691 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/323631a8-107c-4fd0-8b57-04550f0fc7d1-serving-cert\") pod \"authentication-operator-69f744f599-j2m8k\" (UID: \"323631a8-107c-4fd0-8b57-04550f0fc7d1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-j2m8k" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.451753 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dd50f0ac-ac97-413f-a350-7496681f7f0a-config\") pod \"route-controller-manager-6576b87f9c-wk4bd\" (UID: \"dd50f0ac-ac97-413f-a350-7496681f7f0a\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wk4bd" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.451804 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d21f4737-c582-41e9-846c-dd75ec173afd-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-dklh2\" (UID: \"d21f4737-c582-41e9-846c-dd75ec173afd\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-dklh2" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.451866 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/883a0642-0302-4fc5-b361-e25f1aba0878-cert\") pod \"ingress-canary-vpg8w\" (UID: \"883a0642-0302-4fc5-b361-e25f1aba0878\") " pod="openshift-ingress-canary/ingress-canary-vpg8w" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.451923 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/a1d6c12e-ff5a-4687-80ca-5699dd1bb9fe-csi-data-dir\") pod \"csi-hostpathplugin-nk67f\" (UID: \"a1d6c12e-ff5a-4687-80ca-5699dd1bb9fe\") " pod="hostpath-provisioner/csi-hostpathplugin-nk67f" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.452009 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/ff05ab6e-248f-445c-9856-c1a2868f8840-signing-key\") pod \"service-ca-9c57cc56f-8rxsm\" (UID: \"ff05ab6e-248f-445c-9856-c1a2868f8840\") " pod="openshift-service-ca/service-ca-9c57cc56f-8rxsm" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.452074 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4rsfz\" (UniqueName: \"kubernetes.io/projected/af8fbf5a-f90e-4d22-ae65-05d8d8253308-kube-api-access-4rsfz\") pod \"oauth-openshift-558db77b4-8z68k\" (UID: \"af8fbf5a-f90e-4d22-ae65-05d8d8253308\") " pod="openshift-authentication/oauth-openshift-558db77b4-8z68k" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.452155 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/22545b78-29bb-45aa-a98c-fd7ce9703239-serving-cert\") pod \"console-operator-58897d9998-bjzzk\" (UID: \"22545b78-29bb-45aa-a98c-fd7ce9703239\") " pod="openshift-console-operator/console-operator-58897d9998-bjzzk" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.452216 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/af8fbf5a-f90e-4d22-ae65-05d8d8253308-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-8z68k\" (UID: \"af8fbf5a-f90e-4d22-ae65-05d8d8253308\") " pod="openshift-authentication/oauth-openshift-558db77b4-8z68k" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.454287 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/323631a8-107c-4fd0-8b57-04550f0fc7d1-config\") pod \"authentication-operator-69f744f599-j2m8k\" (UID: \"323631a8-107c-4fd0-8b57-04550f0fc7d1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-j2m8k" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.454540 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/911b7e6b-1df0-4ef3-822e-84925bd6bbdc-etcd-client\") pod \"apiserver-76f77b778f-q76nt\" (UID: \"911b7e6b-1df0-4ef3-822e-84925bd6bbdc\") " pod="openshift-apiserver/apiserver-76f77b778f-q76nt" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.456145 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b349b0af-0ba8-409f-a915-7fb06b2e3774-config\") pod \"machine-api-operator-5694c8668f-fjdwv\" (UID: \"b349b0af-0ba8-409f-a915-7fb06b2e3774\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-fjdwv" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.457497 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/911b7e6b-1df0-4ef3-822e-84925bd6bbdc-node-pullsecrets\") pod \"apiserver-76f77b778f-q76nt\" (UID: \"911b7e6b-1df0-4ef3-822e-84925bd6bbdc\") " pod="openshift-apiserver/apiserver-76f77b778f-q76nt" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.457581 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/4d7b1b9b-33de-4fca-b629-6790c3a5cb7d-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-hqhlr\" (UID: \"4d7b1b9b-33de-4fca-b629-6790c3a5cb7d\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-hqhlr" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.458047 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a47e3600-79be-46c3-a1b8-6678d8da80c5-serving-cert\") pod \"etcd-operator-b45778765-x4fdc\" (UID: \"a47e3600-79be-46c3-a1b8-6678d8da80c5\") " pod="openshift-etcd-operator/etcd-operator-b45778765-x4fdc" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.458159 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/744def19-0006-4ab4-9d45-8a2a39bd410a-certs\") pod \"machine-config-server-fdbvz\" (UID: \"744def19-0006-4ab4-9d45-8a2a39bd410a\") " pod="openshift-machine-config-operator/machine-config-server-fdbvz" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.458325 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e853926-f1e6-4571-863b-97918e660677-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-vn26d\" (UID: \"8e853926-f1e6-4571-863b-97918e660677\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-vn26d" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.458390 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8e853926-f1e6-4571-863b-97918e660677-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-vn26d\" (UID: \"8e853926-f1e6-4571-863b-97918e660677\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-vn26d" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.458495 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4bceefc5-1905-48bd-b214-ab5824f0a302-config\") pod \"kube-controller-manager-operator-78b949d7b-mdv87\" (UID: \"4bceefc5-1905-48bd-b214-ab5824f0a302\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-mdv87" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.458609 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9b1f3faf-72a1-44e2-814b-964902e78d3e-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-wlbrp\" (UID: \"9b1f3faf-72a1-44e2-814b-964902e78d3e\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wlbrp" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.458687 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7c497a24-d8cb-4560-993f-fc73b3af8c04-serving-cert\") pod \"openshift-config-operator-7777fb866f-v6gnw\" (UID: \"7c497a24-d8cb-4560-993f-fc73b3af8c04\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-v6gnw" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.458759 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xvhtm\" (UniqueName: \"kubernetes.io/projected/e81889bb-3225-458e-bce5-cfbafe6f4f92-kube-api-access-xvhtm\") pod \"machine-config-controller-84d6567774-xr4mg\" (UID: \"e81889bb-3225-458e-bce5-cfbafe6f4f92\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-xr4mg" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.458821 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c682e90b-3a5a-4b87-996b-8365b0936bda-config-volume\") pod \"dns-default-dr4mr\" (UID: \"c682e90b-3a5a-4b87-996b-8365b0936bda\") " pod="openshift-dns/dns-default-dr4mr" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.458871 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2fa5ee3e-29f4-43b2-a154-3c988754897a-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-zzrfg\" (UID: \"2fa5ee3e-29f4-43b2-a154-3c988754897a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zzrfg" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.458930 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/2fa5ee3e-29f4-43b2-a154-3c988754897a-encryption-config\") pod \"apiserver-7bbb656c7d-zzrfg\" (UID: \"2fa5ee3e-29f4-43b2-a154-3c988754897a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zzrfg" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.459134 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-94hbm\" (UniqueName: \"kubernetes.io/projected/7c497a24-d8cb-4560-993f-fc73b3af8c04-kube-api-access-94hbm\") pod \"openshift-config-operator-7777fb866f-v6gnw\" (UID: \"7c497a24-d8cb-4560-993f-fc73b3af8c04\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-v6gnw" Oct 06 10:00:42 crc kubenswrapper[4824]: E1006 10:00:42.459152 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-06 10:00:42.959126676 +0000 UTC m=+152.323549737 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8l8gd" (UID: "c2f6e84b-a28c-4ead-97f5-978b0639e5bf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.459238 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-26z5m\" (UniqueName: \"kubernetes.io/projected/678eeee3-60d8-4dbe-b815-10de5a46c4d1-kube-api-access-26z5m\") pod \"multus-admission-controller-857f4d67dd-56948\" (UID: \"678eeee3-60d8-4dbe-b815-10de5a46c4d1\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-56948" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.459318 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/b349b0af-0ba8-409f-a915-7fb06b2e3774-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-fjdwv\" (UID: \"b349b0af-0ba8-409f-a915-7fb06b2e3774\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-fjdwv" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.459381 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lzv6l\" (UniqueName: \"kubernetes.io/projected/2a76ccaa-9249-46ba-bbc5-dbbbc33b06bb-kube-api-access-lzv6l\") pod \"control-plane-machine-set-operator-78cbb6b69f-fn7dj\" (UID: \"2a76ccaa-9249-46ba-bbc5-dbbbc33b06bb\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-fn7dj" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.459449 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/a47e3600-79be-46c3-a1b8-6678d8da80c5-etcd-service-ca\") pod \"etcd-operator-b45778765-x4fdc\" (UID: \"a47e3600-79be-46c3-a1b8-6678d8da80c5\") " pod="openshift-etcd-operator/etcd-operator-b45778765-x4fdc" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.459518 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4lbvd\" (UniqueName: \"kubernetes.io/projected/3bdcba24-670d-427a-bfd6-4372c9f944a2-kube-api-access-4lbvd\") pod \"ingress-operator-5b745b69d9-7p7wb\" (UID: \"3bdcba24-670d-427a-bfd6-4372c9f944a2\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7p7wb" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.459596 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/722b7a46-52a1-49a7-a053-803334dc5964-oauth-serving-cert\") pod \"console-f9d7485db-fpxws\" (UID: \"722b7a46-52a1-49a7-a053-803334dc5964\") " pod="openshift-console/console-f9d7485db-fpxws" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.459639 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/911b7e6b-1df0-4ef3-822e-84925bd6bbdc-etcd-serving-ca\") pod \"apiserver-76f77b778f-q76nt\" (UID: \"911b7e6b-1df0-4ef3-822e-84925bd6bbdc\") " pod="openshift-apiserver/apiserver-76f77b778f-q76nt" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.459936 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/982129a4-11b4-4024-afb0-ac6024165c7a-srv-cert\") pod \"olm-operator-6b444d44fb-6tt2t\" (UID: \"982129a4-11b4-4024-afb0-ac6024165c7a\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6tt2t" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.460056 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wz4zb\" (UniqueName: \"kubernetes.io/projected/982129a4-11b4-4024-afb0-ac6024165c7a-kube-api-access-wz4zb\") pod \"olm-operator-6b444d44fb-6tt2t\" (UID: \"982129a4-11b4-4024-afb0-ac6024165c7a\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6tt2t" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.460111 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cjzws\" (UniqueName: \"kubernetes.io/projected/26536a6b-3639-46ad-b024-5e21b7118fe7-kube-api-access-cjzws\") pod \"collect-profiles-29329080-9kpdv\" (UID: \"26536a6b-3639-46ad-b024-5e21b7118fe7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329080-9kpdv" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.460175 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/9da4a453-8e05-4360-b74c-ad1c4706b005-auth-proxy-config\") pod \"machine-approver-56656f9798-4hw5p\" (UID: \"9da4a453-8e05-4360-b74c-ad1c4706b005\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-4hw5p" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.460240 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z44hr\" (UniqueName: \"kubernetes.io/projected/9da4a453-8e05-4360-b74c-ad1c4706b005-kube-api-access-z44hr\") pod \"machine-approver-56656f9798-4hw5p\" (UID: \"9da4a453-8e05-4360-b74c-ad1c4706b005\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-4hw5p" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.460302 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/911b7e6b-1df0-4ef3-822e-84925bd6bbdc-audit-dir\") pod \"apiserver-76f77b778f-q76nt\" (UID: \"911b7e6b-1df0-4ef3-822e-84925bd6bbdc\") " pod="openshift-apiserver/apiserver-76f77b778f-q76nt" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.460375 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/7c497a24-d8cb-4560-993f-fc73b3af8c04-available-featuregates\") pod \"openshift-config-operator-7777fb866f-v6gnw\" (UID: \"7c497a24-d8cb-4560-993f-fc73b3af8c04\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-v6gnw" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.460432 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2fa5ee3e-29f4-43b2-a154-3c988754897a-serving-cert\") pod \"apiserver-7bbb656c7d-zzrfg\" (UID: \"2fa5ee3e-29f4-43b2-a154-3c988754897a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zzrfg" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.460490 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/2fa5ee3e-29f4-43b2-a154-3c988754897a-audit-policies\") pod \"apiserver-7bbb656c7d-zzrfg\" (UID: \"2fa5ee3e-29f4-43b2-a154-3c988754897a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zzrfg" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.460550 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/45682655-0e98-4a5e-b5e8-308f4d0f7d9a-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-prthm\" (UID: \"45682655-0e98-4a5e-b5e8-308f4d0f7d9a\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-prthm" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.460608 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/323631a8-107c-4fd0-8b57-04550f0fc7d1-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-j2m8k\" (UID: \"323631a8-107c-4fd0-8b57-04550f0fc7d1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-j2m8k" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.460620 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/d21f4737-c582-41e9-846c-dd75ec173afd-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-dklh2\" (UID: \"d21f4737-c582-41e9-846c-dd75ec173afd\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-dklh2" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.460730 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d6e255a4-d3e7-4cef-bf40-f98aa11b3e3b-service-ca-bundle\") pod \"router-default-5444994796-7qwsv\" (UID: \"d6e255a4-d3e7-4cef-bf40-f98aa11b3e3b\") " pod="openshift-ingress/router-default-5444994796-7qwsv" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.460787 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rnnkq\" (UniqueName: \"kubernetes.io/projected/d2595750-aae8-4f17-8782-b7eccb0c2948-kube-api-access-rnnkq\") pod \"marketplace-operator-79b997595-qq4zz\" (UID: \"d2595750-aae8-4f17-8782-b7eccb0c2948\") " pod="openshift-marketplace/marketplace-operator-79b997595-qq4zz" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.460792 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22545b78-29bb-45aa-a98c-fd7ce9703239-config\") pod \"console-operator-58897d9998-bjzzk\" (UID: \"22545b78-29bb-45aa-a98c-fd7ce9703239\") " pod="openshift-console-operator/console-operator-58897d9998-bjzzk" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.460847 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tfjz5\" (UniqueName: \"kubernetes.io/projected/744def19-0006-4ab4-9d45-8a2a39bd410a-kube-api-access-tfjz5\") pod \"machine-config-server-fdbvz\" (UID: \"744def19-0006-4ab4-9d45-8a2a39bd410a\") " pod="openshift-machine-config-operator/machine-config-server-fdbvz" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.460906 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dpf26\" (UniqueName: \"kubernetes.io/projected/d40e2672-6c82-45e6-a292-21615712032d-kube-api-access-dpf26\") pod \"kube-storage-version-migrator-operator-b67b599dd-wp6kv\" (UID: \"d40e2672-6c82-45e6-a292-21615712032d\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-wp6kv" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.460975 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/b349b0af-0ba8-409f-a915-7fb06b2e3774-images\") pod \"machine-api-operator-5694c8668f-fjdwv\" (UID: \"b349b0af-0ba8-409f-a915-7fb06b2e3774\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-fjdwv" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.461074 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b6f7d3cb-418e-4f73-95a0-2477e37976c3-config\") pod \"controller-manager-879f6c89f-nqmr6\" (UID: \"b6f7d3cb-418e-4f73-95a0-2477e37976c3\") " pod="openshift-controller-manager/controller-manager-879f6c89f-nqmr6" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.461142 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gx2z5\" (UniqueName: \"kubernetes.io/projected/9613dce2-6a39-4480-8b7f-553288ebc45b-kube-api-access-gx2z5\") pod \"package-server-manager-789f6589d5-px5fp\" (UID: \"9613dce2-6a39-4480-8b7f-553288ebc45b\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-px5fp" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.461220 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c2f6e84b-a28c-4ead-97f5-978b0639e5bf-trusted-ca\") pod \"image-registry-697d97f7c8-8l8gd\" (UID: \"c2f6e84b-a28c-4ead-97f5-978b0639e5bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-8l8gd" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.461284 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c2f6e84b-a28c-4ead-97f5-978b0639e5bf-bound-sa-token\") pod \"image-registry-697d97f7c8-8l8gd\" (UID: \"c2f6e84b-a28c-4ead-97f5-978b0639e5bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-8l8gd" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.461353 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/45682655-0e98-4a5e-b5e8-308f4d0f7d9a-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-prthm\" (UID: \"45682655-0e98-4a5e-b5e8-308f4d0f7d9a\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-prthm" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.461420 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d2595750-aae8-4f17-8782-b7eccb0c2948-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-qq4zz\" (UID: \"d2595750-aae8-4f17-8782-b7eccb0c2948\") " pod="openshift-marketplace/marketplace-operator-79b997595-qq4zz" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.461487 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8e853926-f1e6-4571-863b-97918e660677-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-vn26d\" (UID: \"8e853926-f1e6-4571-863b-97918e660677\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-vn26d" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.461863 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/911b7e6b-1df0-4ef3-822e-84925bd6bbdc-audit\") pod \"apiserver-76f77b778f-q76nt\" (UID: \"911b7e6b-1df0-4ef3-822e-84925bd6bbdc\") " pod="openshift-apiserver/apiserver-76f77b778f-q76nt" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.462172 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/911b7e6b-1df0-4ef3-822e-84925bd6bbdc-encryption-config\") pod \"apiserver-76f77b778f-q76nt\" (UID: \"911b7e6b-1df0-4ef3-822e-84925bd6bbdc\") " pod="openshift-apiserver/apiserver-76f77b778f-q76nt" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.462259 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/af8fbf5a-f90e-4d22-ae65-05d8d8253308-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-8z68k\" (UID: \"af8fbf5a-f90e-4d22-ae65-05d8d8253308\") " pod="openshift-authentication/oauth-openshift-558db77b4-8z68k" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.462332 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/46e11b34-d0dd-4ea6-bed7-66eb707c7f92-serving-cert\") pod \"service-ca-operator-777779d784-5blgj\" (UID: \"46e11b34-d0dd-4ea6-bed7-66eb707c7f92\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-5blgj" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.462425 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-48kwg\" (UniqueName: \"kubernetes.io/projected/b349b0af-0ba8-409f-a915-7fb06b2e3774-kube-api-access-48kwg\") pod \"machine-api-operator-5694c8668f-fjdwv\" (UID: \"b349b0af-0ba8-409f-a915-7fb06b2e3774\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-fjdwv" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.462496 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ffceea22-78f2-4032-a23a-71b4fb7ea83e-webhook-cert\") pod \"packageserver-d55dfcdfc-s8j95\" (UID: \"ffceea22-78f2-4032-a23a-71b4fb7ea83e\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-s8j95" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.462561 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/af8fbf5a-f90e-4d22-ae65-05d8d8253308-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-8z68k\" (UID: \"af8fbf5a-f90e-4d22-ae65-05d8d8253308\") " pod="openshift-authentication/oauth-openshift-558db77b4-8z68k" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.462617 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k66kw\" (UniqueName: \"kubernetes.io/projected/a1d6c12e-ff5a-4687-80ca-5699dd1bb9fe-kube-api-access-k66kw\") pod \"csi-hostpathplugin-nk67f\" (UID: \"a1d6c12e-ff5a-4687-80ca-5699dd1bb9fe\") " pod="hostpath-provisioner/csi-hostpathplugin-nk67f" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.462686 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h7xwt\" (UniqueName: \"kubernetes.io/projected/c682e90b-3a5a-4b87-996b-8365b0936bda-kube-api-access-h7xwt\") pod \"dns-default-dr4mr\" (UID: \"c682e90b-3a5a-4b87-996b-8365b0936bda\") " pod="openshift-dns/dns-default-dr4mr" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.462761 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/c2f6e84b-a28c-4ead-97f5-978b0639e5bf-ca-trust-extracted\") pod \"image-registry-697d97f7c8-8l8gd\" (UID: \"c2f6e84b-a28c-4ead-97f5-978b0639e5bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-8l8gd" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.462827 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c8fc8c20-e06a-4039-80c0-82f84ea81391-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-sb9c8\" (UID: \"c8fc8c20-e06a-4039-80c0-82f84ea81391\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-sb9c8" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.463632 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/af8fbf5a-f90e-4d22-ae65-05d8d8253308-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-8z68k\" (UID: \"af8fbf5a-f90e-4d22-ae65-05d8d8253308\") " pod="openshift-authentication/oauth-openshift-558db77b4-8z68k" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.463750 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/dd50f0ac-ac97-413f-a350-7496681f7f0a-client-ca\") pod \"route-controller-manager-6576b87f9c-wk4bd\" (UID: \"dd50f0ac-ac97-413f-a350-7496681f7f0a\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wk4bd" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.463829 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/ff05ab6e-248f-445c-9856-c1a2868f8840-signing-cabundle\") pod \"service-ca-9c57cc56f-8rxsm\" (UID: \"ff05ab6e-248f-445c-9856-c1a2868f8840\") " pod="openshift-service-ca/service-ca-9c57cc56f-8rxsm" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.463908 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/678eeee3-60d8-4dbe-b815-10de5a46c4d1-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-56948\" (UID: \"678eeee3-60d8-4dbe-b815-10de5a46c4d1\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-56948" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.464226 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/911b7e6b-1df0-4ef3-822e-84925bd6bbdc-image-import-ca\") pod \"apiserver-76f77b778f-q76nt\" (UID: \"911b7e6b-1df0-4ef3-822e-84925bd6bbdc\") " pod="openshift-apiserver/apiserver-76f77b778f-q76nt" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.464307 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b6f7d3cb-418e-4f73-95a0-2477e37976c3-serving-cert\") pod \"controller-manager-879f6c89f-nqmr6\" (UID: \"b6f7d3cb-418e-4f73-95a0-2477e37976c3\") " pod="openshift-controller-manager/controller-manager-879f6c89f-nqmr6" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.464372 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z8t7q\" (UniqueName: \"kubernetes.io/projected/323631a8-107c-4fd0-8b57-04550f0fc7d1-kube-api-access-z8t7q\") pod \"authentication-operator-69f744f599-j2m8k\" (UID: \"323631a8-107c-4fd0-8b57-04550f0fc7d1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-j2m8k" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.464446 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/722b7a46-52a1-49a7-a053-803334dc5964-console-serving-cert\") pod \"console-f9d7485db-fpxws\" (UID: \"722b7a46-52a1-49a7-a053-803334dc5964\") " pod="openshift-console/console-f9d7485db-fpxws" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.464522 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9da4a453-8e05-4360-b74c-ad1c4706b005-config\") pod \"machine-approver-56656f9798-4hw5p\" (UID: \"9da4a453-8e05-4360-b74c-ad1c4706b005\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-4hw5p" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.464595 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2c9tb\" (UniqueName: \"kubernetes.io/projected/d21f4737-c582-41e9-846c-dd75ec173afd-kube-api-access-2c9tb\") pod \"cluster-image-registry-operator-dc59b4c8b-dklh2\" (UID: \"d21f4737-c582-41e9-846c-dd75ec173afd\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-dklh2" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.467541 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xsfd8\" (UniqueName: \"kubernetes.io/projected/ff05ab6e-248f-445c-9856-c1a2868f8840-kube-api-access-xsfd8\") pod \"service-ca-9c57cc56f-8rxsm\" (UID: \"ff05ab6e-248f-445c-9856-c1a2868f8840\") " pod="openshift-service-ca/service-ca-9c57cc56f-8rxsm" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.467651 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gvtfv\" (UniqueName: \"kubernetes.io/projected/7532dafc-fb61-4066-bb24-2470e4b2cfe6-kube-api-access-gvtfv\") pod \"dns-operator-744455d44c-6kn7b\" (UID: \"7532dafc-fb61-4066-bb24-2470e4b2cfe6\") " pod="openshift-dns-operator/dns-operator-744455d44c-6kn7b" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.467726 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/22545b78-29bb-45aa-a98c-fd7ce9703239-trusted-ca\") pod \"console-operator-58897d9998-bjzzk\" (UID: \"22545b78-29bb-45aa-a98c-fd7ce9703239\") " pod="openshift-console-operator/console-operator-58897d9998-bjzzk" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.467796 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b6f7d3cb-418e-4f73-95a0-2477e37976c3-client-ca\") pod \"controller-manager-879f6c89f-nqmr6\" (UID: \"b6f7d3cb-418e-4f73-95a0-2477e37976c3\") " pod="openshift-controller-manager/controller-manager-879f6c89f-nqmr6" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.467870 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/722b7a46-52a1-49a7-a053-803334dc5964-console-config\") pod \"console-f9d7485db-fpxws\" (UID: \"722b7a46-52a1-49a7-a053-803334dc5964\") " pod="openshift-console/console-f9d7485db-fpxws" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.467935 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6t6pf\" (UniqueName: \"kubernetes.io/projected/9b1f3faf-72a1-44e2-814b-964902e78d3e-kube-api-access-6t6pf\") pod \"openshift-controller-manager-operator-756b6f6bc6-wlbrp\" (UID: \"9b1f3faf-72a1-44e2-814b-964902e78d3e\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wlbrp" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.468032 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d21f4737-c582-41e9-846c-dd75ec173afd-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-dklh2\" (UID: \"d21f4737-c582-41e9-846c-dd75ec173afd\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-dklh2" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.468101 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/ffceea22-78f2-4032-a23a-71b4fb7ea83e-apiservice-cert\") pod \"packageserver-d55dfcdfc-s8j95\" (UID: \"ffceea22-78f2-4032-a23a-71b4fb7ea83e\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-s8j95" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.468152 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/af8fbf5a-f90e-4d22-ae65-05d8d8253308-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-8z68k\" (UID: \"af8fbf5a-f90e-4d22-ae65-05d8d8253308\") " pod="openshift-authentication/oauth-openshift-558db77b4-8z68k" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.471326 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b6f7d3cb-418e-4f73-95a0-2477e37976c3-config\") pod \"controller-manager-879f6c89f-nqmr6\" (UID: \"b6f7d3cb-418e-4f73-95a0-2477e37976c3\") " pod="openshift-controller-manager/controller-manager-879f6c89f-nqmr6" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.466836 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/b349b0af-0ba8-409f-a915-7fb06b2e3774-images\") pod \"machine-api-operator-5694c8668f-fjdwv\" (UID: \"b349b0af-0ba8-409f-a915-7fb06b2e3774\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-fjdwv" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.468619 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4bceefc5-1905-48bd-b214-ab5824f0a302-config\") pod \"kube-controller-manager-operator-78b949d7b-mdv87\" (UID: \"4bceefc5-1905-48bd-b214-ab5824f0a302\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-mdv87" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.467161 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/722b7a46-52a1-49a7-a053-803334dc5964-trusted-ca-bundle\") pod \"console-f9d7485db-fpxws\" (UID: \"722b7a46-52a1-49a7-a053-803334dc5964\") " pod="openshift-console/console-f9d7485db-fpxws" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.469414 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/45682655-0e98-4a5e-b5e8-308f4d0f7d9a-config\") pod \"kube-apiserver-operator-766d6c64bb-prthm\" (UID: \"45682655-0e98-4a5e-b5e8-308f4d0f7d9a\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-prthm" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.471491 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/911b7e6b-1df0-4ef3-822e-84925bd6bbdc-trusted-ca-bundle\") pod \"apiserver-76f77b778f-q76nt\" (UID: \"911b7e6b-1df0-4ef3-822e-84925bd6bbdc\") " pod="openshift-apiserver/apiserver-76f77b778f-q76nt" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.470288 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/b349b0af-0ba8-409f-a915-7fb06b2e3774-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-fjdwv\" (UID: \"b349b0af-0ba8-409f-a915-7fb06b2e3774\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-fjdwv" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.470535 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/c2f6e84b-a28c-4ead-97f5-978b0639e5bf-registry-tls\") pod \"image-registry-697d97f7c8-8l8gd\" (UID: \"c2f6e84b-a28c-4ead-97f5-978b0639e5bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-8l8gd" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.471226 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/323631a8-107c-4fd0-8b57-04550f0fc7d1-serving-cert\") pod \"authentication-operator-69f744f599-j2m8k\" (UID: \"323631a8-107c-4fd0-8b57-04550f0fc7d1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-j2m8k" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.468169 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/911b7e6b-1df0-4ef3-822e-84925bd6bbdc-audit-dir\") pod \"apiserver-76f77b778f-q76nt\" (UID: \"911b7e6b-1df0-4ef3-822e-84925bd6bbdc\") " pod="openshift-apiserver/apiserver-76f77b778f-q76nt" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.471491 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/911b7e6b-1df0-4ef3-822e-84925bd6bbdc-encryption-config\") pod \"apiserver-76f77b778f-q76nt\" (UID: \"911b7e6b-1df0-4ef3-822e-84925bd6bbdc\") " pod="openshift-apiserver/apiserver-76f77b778f-q76nt" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.471737 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dd50f0ac-ac97-413f-a350-7496681f7f0a-serving-cert\") pod \"route-controller-manager-6576b87f9c-wk4bd\" (UID: \"dd50f0ac-ac97-413f-a350-7496681f7f0a\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wk4bd" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.471333 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/9613dce2-6a39-4480-8b7f-553288ebc45b-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-px5fp\" (UID: \"9613dce2-6a39-4480-8b7f-553288ebc45b\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-px5fp" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.471830 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/af8fbf5a-f90e-4d22-ae65-05d8d8253308-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-8z68k\" (UID: \"af8fbf5a-f90e-4d22-ae65-05d8d8253308\") " pod="openshift-authentication/oauth-openshift-558db77b4-8z68k" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.471883 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/d6e255a4-d3e7-4cef-bf40-f98aa11b3e3b-stats-auth\") pod \"router-default-5444994796-7qwsv\" (UID: \"d6e255a4-d3e7-4cef-bf40-f98aa11b3e3b\") " pod="openshift-ingress/router-default-5444994796-7qwsv" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.471859 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/2a0286c5-9866-4a1c-9855-49cea5d04d69-auth-proxy-config\") pod \"machine-config-operator-74547568cd-vvtbh\" (UID: \"2a0286c5-9866-4a1c-9855-49cea5d04d69\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vvtbh" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.471954 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/9da4a453-8e05-4360-b74c-ad1c4706b005-machine-approver-tls\") pod \"machine-approver-56656f9798-4hw5p\" (UID: \"9da4a453-8e05-4360-b74c-ad1c4706b005\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-4hw5p" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.472010 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/e81889bb-3225-458e-bce5-cfbafe6f4f92-proxy-tls\") pod \"machine-config-controller-84d6567774-xr4mg\" (UID: \"e81889bb-3225-458e-bce5-cfbafe6f4f92\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-xr4mg" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.472089 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/c2f6e84b-a28c-4ead-97f5-978b0639e5bf-registry-certificates\") pod \"image-registry-697d97f7c8-8l8gd\" (UID: \"c2f6e84b-a28c-4ead-97f5-978b0639e5bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-8l8gd" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.472101 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c2f6e84b-a28c-4ead-97f5-978b0639e5bf-trusted-ca\") pod \"image-registry-697d97f7c8-8l8gd\" (UID: \"c2f6e84b-a28c-4ead-97f5-978b0639e5bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-8l8gd" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.472131 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/c682e90b-3a5a-4b87-996b-8365b0936bda-metrics-tls\") pod \"dns-default-dr4mr\" (UID: \"c682e90b-3a5a-4b87-996b-8365b0936bda\") " pod="openshift-dns/dns-default-dr4mr" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.472178 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/911b7e6b-1df0-4ef3-822e-84925bd6bbdc-serving-cert\") pod \"apiserver-76f77b778f-q76nt\" (UID: \"911b7e6b-1df0-4ef3-822e-84925bd6bbdc\") " pod="openshift-apiserver/apiserver-76f77b778f-q76nt" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.472218 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/d2595750-aae8-4f17-8782-b7eccb0c2948-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-qq4zz\" (UID: \"d2595750-aae8-4f17-8782-b7eccb0c2948\") " pod="openshift-marketplace/marketplace-operator-79b997595-qq4zz" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.472270 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/911b7e6b-1df0-4ef3-822e-84925bd6bbdc-config\") pod \"apiserver-76f77b778f-q76nt\" (UID: \"911b7e6b-1df0-4ef3-822e-84925bd6bbdc\") " pod="openshift-apiserver/apiserver-76f77b778f-q76nt" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.472312 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/af8fbf5a-f90e-4d22-ae65-05d8d8253308-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-8z68k\" (UID: \"af8fbf5a-f90e-4d22-ae65-05d8d8253308\") " pod="openshift-authentication/oauth-openshift-558db77b4-8z68k" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.472463 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/a1d6c12e-ff5a-4687-80ca-5699dd1bb9fe-socket-dir\") pod \"csi-hostpathplugin-nk67f\" (UID: \"a1d6c12e-ff5a-4687-80ca-5699dd1bb9fe\") " pod="hostpath-provisioner/csi-hostpathplugin-nk67f" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.472529 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jnc4f\" (UniqueName: \"kubernetes.io/projected/a47e3600-79be-46c3-a1b8-6678d8da80c5-kube-api-access-jnc4f\") pod \"etcd-operator-b45778765-x4fdc\" (UID: \"a47e3600-79be-46c3-a1b8-6678d8da80c5\") " pod="openshift-etcd-operator/etcd-operator-b45778765-x4fdc" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.472570 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/3bdcba24-670d-427a-bfd6-4372c9f944a2-metrics-tls\") pod \"ingress-operator-5b745b69d9-7p7wb\" (UID: \"3bdcba24-670d-427a-bfd6-4372c9f944a2\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7p7wb" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.472617 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/722b7a46-52a1-49a7-a053-803334dc5964-console-oauth-config\") pod \"console-f9d7485db-fpxws\" (UID: \"722b7a46-52a1-49a7-a053-803334dc5964\") " pod="openshift-console/console-f9d7485db-fpxws" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.472900 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9b1f3faf-72a1-44e2-814b-964902e78d3e-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-wlbrp\" (UID: \"9b1f3faf-72a1-44e2-814b-964902e78d3e\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wlbrp" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.475101 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c8fc8c20-e06a-4039-80c0-82f84ea81391-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-sb9c8\" (UID: \"c8fc8c20-e06a-4039-80c0-82f84ea81391\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-sb9c8" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.475373 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/323631a8-107c-4fd0-8b57-04550f0fc7d1-service-ca-bundle\") pod \"authentication-operator-69f744f599-j2m8k\" (UID: \"323631a8-107c-4fd0-8b57-04550f0fc7d1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-j2m8k" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.475394 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/a47e3600-79be-46c3-a1b8-6678d8da80c5-etcd-client\") pod \"etcd-operator-b45778765-x4fdc\" (UID: \"a47e3600-79be-46c3-a1b8-6678d8da80c5\") " pod="openshift-etcd-operator/etcd-operator-b45778765-x4fdc" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.475607 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d6e255a4-d3e7-4cef-bf40-f98aa11b3e3b-metrics-certs\") pod \"router-default-5444994796-7qwsv\" (UID: \"d6e255a4-d3e7-4cef-bf40-f98aa11b3e3b\") " pod="openshift-ingress/router-default-5444994796-7qwsv" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.475790 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/722b7a46-52a1-49a7-a053-803334dc5964-service-ca\") pod \"console-f9d7485db-fpxws\" (UID: \"722b7a46-52a1-49a7-a053-803334dc5964\") " pod="openshift-console/console-f9d7485db-fpxws" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.476194 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/22545b78-29bb-45aa-a98c-fd7ce9703239-serving-cert\") pod \"console-operator-58897d9998-bjzzk\" (UID: \"22545b78-29bb-45aa-a98c-fd7ce9703239\") " pod="openshift-console-operator/console-operator-58897d9998-bjzzk" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.476468 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9da4a453-8e05-4360-b74c-ad1c4706b005-config\") pod \"machine-approver-56656f9798-4hw5p\" (UID: \"9da4a453-8e05-4360-b74c-ad1c4706b005\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-4hw5p" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.477175 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/722b7a46-52a1-49a7-a053-803334dc5964-oauth-serving-cert\") pod \"console-f9d7485db-fpxws\" (UID: \"722b7a46-52a1-49a7-a053-803334dc5964\") " pod="openshift-console/console-f9d7485db-fpxws" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.477703 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/45682655-0e98-4a5e-b5e8-308f4d0f7d9a-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-prthm\" (UID: \"45682655-0e98-4a5e-b5e8-308f4d0f7d9a\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-prthm" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.478242 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/d6e255a4-d3e7-4cef-bf40-f98aa11b3e3b-default-certificate\") pod \"router-default-5444994796-7qwsv\" (UID: \"d6e255a4-d3e7-4cef-bf40-f98aa11b3e3b\") " pod="openshift-ingress/router-default-5444994796-7qwsv" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.478695 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/722b7a46-52a1-49a7-a053-803334dc5964-console-config\") pod \"console-f9d7485db-fpxws\" (UID: \"722b7a46-52a1-49a7-a053-803334dc5964\") " pod="openshift-console/console-f9d7485db-fpxws" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.479797 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/c2f6e84b-a28c-4ead-97f5-978b0639e5bf-installation-pull-secrets\") pod \"image-registry-697d97f7c8-8l8gd\" (UID: \"c2f6e84b-a28c-4ead-97f5-978b0639e5bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-8l8gd" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.480835 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/722b7a46-52a1-49a7-a053-803334dc5964-console-serving-cert\") pod \"console-f9d7485db-fpxws\" (UID: \"722b7a46-52a1-49a7-a053-803334dc5964\") " pod="openshift-console/console-f9d7485db-fpxws" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.481711 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dd50f0ac-ac97-413f-a350-7496681f7f0a-config\") pod \"route-controller-manager-6576b87f9c-wk4bd\" (UID: \"dd50f0ac-ac97-413f-a350-7496681f7f0a\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wk4bd" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.483147 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/911b7e6b-1df0-4ef3-822e-84925bd6bbdc-serving-cert\") pod \"apiserver-76f77b778f-q76nt\" (UID: \"911b7e6b-1df0-4ef3-822e-84925bd6bbdc\") " pod="openshift-apiserver/apiserver-76f77b778f-q76nt" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.486204 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d6e255a4-d3e7-4cef-bf40-f98aa11b3e3b-service-ca-bundle\") pod \"router-default-5444994796-7qwsv\" (UID: \"d6e255a4-d3e7-4cef-bf40-f98aa11b3e3b\") " pod="openshift-ingress/router-default-5444994796-7qwsv" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.486567 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c8fc8c20-e06a-4039-80c0-82f84ea81391-config\") pod \"openshift-apiserver-operator-796bbdcf4f-sb9c8\" (UID: \"c8fc8c20-e06a-4039-80c0-82f84ea81391\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-sb9c8" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.487021 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/c2f6e84b-a28c-4ead-97f5-978b0639e5bf-ca-trust-extracted\") pod \"image-registry-697d97f7c8-8l8gd\" (UID: \"c2f6e84b-a28c-4ead-97f5-978b0639e5bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-8l8gd" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.487092 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a47e3600-79be-46c3-a1b8-6678d8da80c5-config\") pod \"etcd-operator-b45778765-x4fdc\" (UID: \"a47e3600-79be-46c3-a1b8-6678d8da80c5\") " pod="openshift-etcd-operator/etcd-operator-b45778765-x4fdc" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.487636 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/911b7e6b-1df0-4ef3-822e-84925bd6bbdc-config\") pod \"apiserver-76f77b778f-q76nt\" (UID: \"911b7e6b-1df0-4ef3-822e-84925bd6bbdc\") " pod="openshift-apiserver/apiserver-76f77b778f-q76nt" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.487766 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/911b7e6b-1df0-4ef3-822e-84925bd6bbdc-image-import-ca\") pod \"apiserver-76f77b778f-q76nt\" (UID: \"911b7e6b-1df0-4ef3-822e-84925bd6bbdc\") " pod="openshift-apiserver/apiserver-76f77b778f-q76nt" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.493147 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-66bhb\" (UniqueName: \"kubernetes.io/projected/c2f6e84b-a28c-4ead-97f5-978b0639e5bf-kube-api-access-66bhb\") pod \"image-registry-697d97f7c8-8l8gd\" (UID: \"c2f6e84b-a28c-4ead-97f5-978b0639e5bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-8l8gd" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.496648 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4bceefc5-1905-48bd-b214-ab5824f0a302-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-mdv87\" (UID: \"4bceefc5-1905-48bd-b214-ab5824f0a302\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-mdv87" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.496845 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/9da4a453-8e05-4360-b74c-ad1c4706b005-auth-proxy-config\") pod \"machine-approver-56656f9798-4hw5p\" (UID: \"9da4a453-8e05-4360-b74c-ad1c4706b005\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-4hw5p" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.497340 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/d21f4737-c582-41e9-846c-dd75ec173afd-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-dklh2\" (UID: \"d21f4737-c582-41e9-846c-dd75ec173afd\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-dklh2" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.497346 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/7c497a24-d8cb-4560-993f-fc73b3af8c04-available-featuregates\") pod \"openshift-config-operator-7777fb866f-v6gnw\" (UID: \"7c497a24-d8cb-4560-993f-fc73b3af8c04\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-v6gnw" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.497938 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7c497a24-d8cb-4560-993f-fc73b3af8c04-serving-cert\") pod \"openshift-config-operator-7777fb866f-v6gnw\" (UID: \"7c497a24-d8cb-4560-993f-fc73b3af8c04\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-v6gnw" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.498805 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b6f7d3cb-418e-4f73-95a0-2477e37976c3-client-ca\") pod \"controller-manager-879f6c89f-nqmr6\" (UID: \"b6f7d3cb-418e-4f73-95a0-2477e37976c3\") " pod="openshift-controller-manager/controller-manager-879f6c89f-nqmr6" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.498599 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9b1f3faf-72a1-44e2-814b-964902e78d3e-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-wlbrp\" (UID: \"9b1f3faf-72a1-44e2-814b-964902e78d3e\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wlbrp" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.498569 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/22545b78-29bb-45aa-a98c-fd7ce9703239-trusted-ca\") pod \"console-operator-58897d9998-bjzzk\" (UID: \"22545b78-29bb-45aa-a98c-fd7ce9703239\") " pod="openshift-console-operator/console-operator-58897d9998-bjzzk" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.499229 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/c2f6e84b-a28c-4ead-97f5-978b0639e5bf-registry-certificates\") pod \"image-registry-697d97f7c8-8l8gd\" (UID: \"c2f6e84b-a28c-4ead-97f5-978b0639e5bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-8l8gd" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.499672 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/a47e3600-79be-46c3-a1b8-6678d8da80c5-etcd-service-ca\") pod \"etcd-operator-b45778765-x4fdc\" (UID: \"a47e3600-79be-46c3-a1b8-6678d8da80c5\") " pod="openshift-etcd-operator/etcd-operator-b45778765-x4fdc" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.499802 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/a47e3600-79be-46c3-a1b8-6678d8da80c5-etcd-ca\") pod \"etcd-operator-b45778765-x4fdc\" (UID: \"a47e3600-79be-46c3-a1b8-6678d8da80c5\") " pod="openshift-etcd-operator/etcd-operator-b45778765-x4fdc" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.500291 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/dd50f0ac-ac97-413f-a350-7496681f7f0a-client-ca\") pod \"route-controller-manager-6576b87f9c-wk4bd\" (UID: \"dd50f0ac-ac97-413f-a350-7496681f7f0a\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wk4bd" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.501196 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3bdcba24-670d-427a-bfd6-4372c9f944a2-trusted-ca\") pod \"ingress-operator-5b745b69d9-7p7wb\" (UID: \"3bdcba24-670d-427a-bfd6-4372c9f944a2\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7p7wb" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.501737 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d21f4737-c582-41e9-846c-dd75ec173afd-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-dklh2\" (UID: \"d21f4737-c582-41e9-846c-dd75ec173afd\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-dklh2" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.501790 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/3bdcba24-670d-427a-bfd6-4372c9f944a2-metrics-tls\") pod \"ingress-operator-5b745b69d9-7p7wb\" (UID: \"3bdcba24-670d-427a-bfd6-4372c9f944a2\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7p7wb" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.502891 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b6f7d3cb-418e-4f73-95a0-2477e37976c3-serving-cert\") pod \"controller-manager-879f6c89f-nqmr6\" (UID: \"b6f7d3cb-418e-4f73-95a0-2477e37976c3\") " pod="openshift-controller-manager/controller-manager-879f6c89f-nqmr6" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.506186 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/7532dafc-fb61-4066-bb24-2470e4b2cfe6-metrics-tls\") pod \"dns-operator-744455d44c-6kn7b\" (UID: \"7532dafc-fb61-4066-bb24-2470e4b2cfe6\") " pod="openshift-dns-operator/dns-operator-744455d44c-6kn7b" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.512142 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fmfcl\" (UniqueName: \"kubernetes.io/projected/c8fc8c20-e06a-4039-80c0-82f84ea81391-kube-api-access-fmfcl\") pod \"openshift-apiserver-operator-796bbdcf4f-sb9c8\" (UID: \"c8fc8c20-e06a-4039-80c0-82f84ea81391\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-sb9c8" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.537118 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2h2fc\" (UniqueName: \"kubernetes.io/projected/b6f7d3cb-418e-4f73-95a0-2477e37976c3-kube-api-access-2h2fc\") pod \"controller-manager-879f6c89f-nqmr6\" (UID: \"b6f7d3cb-418e-4f73-95a0-2477e37976c3\") " pod="openshift-controller-manager/controller-manager-879f6c89f-nqmr6" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.565051 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-57vkp\" (UniqueName: \"kubernetes.io/projected/911b7e6b-1df0-4ef3-822e-84925bd6bbdc-kube-api-access-57vkp\") pod \"apiserver-76f77b778f-q76nt\" (UID: \"911b7e6b-1df0-4ef3-822e-84925bd6bbdc\") " pod="openshift-apiserver/apiserver-76f77b778f-q76nt" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.573694 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 10:00:42 crc kubenswrapper[4824]: E1006 10:00:42.573937 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-06 10:00:43.073887653 +0000 UTC m=+152.438310514 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.574057 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8e853926-f1e6-4571-863b-97918e660677-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-vn26d\" (UID: \"8e853926-f1e6-4571-863b-97918e660677\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-vn26d" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.574092 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d2595750-aae8-4f17-8782-b7eccb0c2948-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-qq4zz\" (UID: \"d2595750-aae8-4f17-8782-b7eccb0c2948\") " pod="openshift-marketplace/marketplace-operator-79b997595-qq4zz" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.574121 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/af8fbf5a-f90e-4d22-ae65-05d8d8253308-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-8z68k\" (UID: \"af8fbf5a-f90e-4d22-ae65-05d8d8253308\") " pod="openshift-authentication/oauth-openshift-558db77b4-8z68k" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.574153 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/46e11b34-d0dd-4ea6-bed7-66eb707c7f92-serving-cert\") pod \"service-ca-operator-777779d784-5blgj\" (UID: \"46e11b34-d0dd-4ea6-bed7-66eb707c7f92\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-5blgj" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.574183 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ffceea22-78f2-4032-a23a-71b4fb7ea83e-webhook-cert\") pod \"packageserver-d55dfcdfc-s8j95\" (UID: \"ffceea22-78f2-4032-a23a-71b4fb7ea83e\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-s8j95" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.574204 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/af8fbf5a-f90e-4d22-ae65-05d8d8253308-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-8z68k\" (UID: \"af8fbf5a-f90e-4d22-ae65-05d8d8253308\") " pod="openshift-authentication/oauth-openshift-558db77b4-8z68k" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.574227 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k66kw\" (UniqueName: \"kubernetes.io/projected/a1d6c12e-ff5a-4687-80ca-5699dd1bb9fe-kube-api-access-k66kw\") pod \"csi-hostpathplugin-nk67f\" (UID: \"a1d6c12e-ff5a-4687-80ca-5699dd1bb9fe\") " pod="hostpath-provisioner/csi-hostpathplugin-nk67f" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.574251 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/af8fbf5a-f90e-4d22-ae65-05d8d8253308-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-8z68k\" (UID: \"af8fbf5a-f90e-4d22-ae65-05d8d8253308\") " pod="openshift-authentication/oauth-openshift-558db77b4-8z68k" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.574277 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h7xwt\" (UniqueName: \"kubernetes.io/projected/c682e90b-3a5a-4b87-996b-8365b0936bda-kube-api-access-h7xwt\") pod \"dns-default-dr4mr\" (UID: \"c682e90b-3a5a-4b87-996b-8365b0936bda\") " pod="openshift-dns/dns-default-dr4mr" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.574298 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/ff05ab6e-248f-445c-9856-c1a2868f8840-signing-cabundle\") pod \"service-ca-9c57cc56f-8rxsm\" (UID: \"ff05ab6e-248f-445c-9856-c1a2868f8840\") " pod="openshift-service-ca/service-ca-9c57cc56f-8rxsm" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.574320 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/678eeee3-60d8-4dbe-b815-10de5a46c4d1-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-56948\" (UID: \"678eeee3-60d8-4dbe-b815-10de5a46c4d1\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-56948" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.574361 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xsfd8\" (UniqueName: \"kubernetes.io/projected/ff05ab6e-248f-445c-9856-c1a2868f8840-kube-api-access-xsfd8\") pod \"service-ca-9c57cc56f-8rxsm\" (UID: \"ff05ab6e-248f-445c-9856-c1a2868f8840\") " pod="openshift-service-ca/service-ca-9c57cc56f-8rxsm" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.574458 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/ffceea22-78f2-4032-a23a-71b4fb7ea83e-apiservice-cert\") pod \"packageserver-d55dfcdfc-s8j95\" (UID: \"ffceea22-78f2-4032-a23a-71b4fb7ea83e\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-s8j95" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.574484 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/9613dce2-6a39-4480-8b7f-553288ebc45b-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-px5fp\" (UID: \"9613dce2-6a39-4480-8b7f-553288ebc45b\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-px5fp" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.574507 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/af8fbf5a-f90e-4d22-ae65-05d8d8253308-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-8z68k\" (UID: \"af8fbf5a-f90e-4d22-ae65-05d8d8253308\") " pod="openshift-authentication/oauth-openshift-558db77b4-8z68k" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.574528 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/2a0286c5-9866-4a1c-9855-49cea5d04d69-auth-proxy-config\") pod \"machine-config-operator-74547568cd-vvtbh\" (UID: \"2a0286c5-9866-4a1c-9855-49cea5d04d69\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vvtbh" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.574551 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/e81889bb-3225-458e-bce5-cfbafe6f4f92-proxy-tls\") pod \"machine-config-controller-84d6567774-xr4mg\" (UID: \"e81889bb-3225-458e-bce5-cfbafe6f4f92\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-xr4mg" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.574594 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/af8fbf5a-f90e-4d22-ae65-05d8d8253308-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-8z68k\" (UID: \"af8fbf5a-f90e-4d22-ae65-05d8d8253308\") " pod="openshift-authentication/oauth-openshift-558db77b4-8z68k" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.574615 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/c682e90b-3a5a-4b87-996b-8365b0936bda-metrics-tls\") pod \"dns-default-dr4mr\" (UID: \"c682e90b-3a5a-4b87-996b-8365b0936bda\") " pod="openshift-dns/dns-default-dr4mr" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.574638 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/d2595750-aae8-4f17-8782-b7eccb0c2948-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-qq4zz\" (UID: \"d2595750-aae8-4f17-8782-b7eccb0c2948\") " pod="openshift-marketplace/marketplace-operator-79b997595-qq4zz" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.574661 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/af8fbf5a-f90e-4d22-ae65-05d8d8253308-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-8z68k\" (UID: \"af8fbf5a-f90e-4d22-ae65-05d8d8253308\") " pod="openshift-authentication/oauth-openshift-558db77b4-8z68k" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.574682 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/a1d6c12e-ff5a-4687-80ca-5699dd1bb9fe-socket-dir\") pod \"csi-hostpathplugin-nk67f\" (UID: \"a1d6c12e-ff5a-4687-80ca-5699dd1bb9fe\") " pod="hostpath-provisioner/csi-hostpathplugin-nk67f" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.574714 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t2gcn\" (UniqueName: \"kubernetes.io/projected/d0546787-f9de-4b06-ab4f-0e1632d36230-kube-api-access-t2gcn\") pod \"catalog-operator-68c6474976-2xznh\" (UID: \"d0546787-f9de-4b06-ab4f-0e1632d36230\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-2xznh" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.574738 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/982129a4-11b4-4024-afb0-ac6024165c7a-profile-collector-cert\") pod \"olm-operator-6b444d44fb-6tt2t\" (UID: \"982129a4-11b4-4024-afb0-ac6024165c7a\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6tt2t" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.574762 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/2a0286c5-9866-4a1c-9855-49cea5d04d69-proxy-tls\") pod \"machine-config-operator-74547568cd-vvtbh\" (UID: \"2a0286c5-9866-4a1c-9855-49cea5d04d69\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vvtbh" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.574785 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/af8fbf5a-f90e-4d22-ae65-05d8d8253308-audit-dir\") pod \"oauth-openshift-558db77b4-8z68k\" (UID: \"af8fbf5a-f90e-4d22-ae65-05d8d8253308\") " pod="openshift-authentication/oauth-openshift-558db77b4-8z68k" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.574807 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/d0546787-f9de-4b06-ab4f-0e1632d36230-profile-collector-cert\") pod \"catalog-operator-68c6474976-2xznh\" (UID: \"d0546787-f9de-4b06-ab4f-0e1632d36230\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-2xznh" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.574828 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/af8fbf5a-f90e-4d22-ae65-05d8d8253308-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-8z68k\" (UID: \"af8fbf5a-f90e-4d22-ae65-05d8d8253308\") " pod="openshift-authentication/oauth-openshift-558db77b4-8z68k" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.574853 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/a1d6c12e-ff5a-4687-80ca-5699dd1bb9fe-mountpoint-dir\") pod \"csi-hostpathplugin-nk67f\" (UID: \"a1d6c12e-ff5a-4687-80ca-5699dd1bb9fe\") " pod="hostpath-provisioner/csi-hostpathplugin-nk67f" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.574875 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/af8fbf5a-f90e-4d22-ae65-05d8d8253308-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-8z68k\" (UID: \"af8fbf5a-f90e-4d22-ae65-05d8d8253308\") " pod="openshift-authentication/oauth-openshift-558db77b4-8z68k" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.574898 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d40e2672-6c82-45e6-a292-21615712032d-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-wp6kv\" (UID: \"d40e2672-6c82-45e6-a292-21615712032d\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-wp6kv" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.574920 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/ffceea22-78f2-4032-a23a-71b4fb7ea83e-tmpfs\") pod \"packageserver-d55dfcdfc-s8j95\" (UID: \"ffceea22-78f2-4032-a23a-71b4fb7ea83e\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-s8j95" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.574945 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d40e2672-6c82-45e6-a292-21615712032d-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-wp6kv\" (UID: \"d40e2672-6c82-45e6-a292-21615712032d\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-wp6kv" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.574967 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/a1d6c12e-ff5a-4687-80ca-5699dd1bb9fe-plugins-dir\") pod \"csi-hostpathplugin-nk67f\" (UID: \"a1d6c12e-ff5a-4687-80ca-5699dd1bb9fe\") " pod="hostpath-provisioner/csi-hostpathplugin-nk67f" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.575026 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-99zcx\" (UniqueName: \"kubernetes.io/projected/46e11b34-d0dd-4ea6-bed7-66eb707c7f92-kube-api-access-99zcx\") pod \"service-ca-operator-777779d784-5blgj\" (UID: \"46e11b34-d0dd-4ea6-bed7-66eb707c7f92\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-5blgj" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.575059 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w6ks8\" (UniqueName: \"kubernetes.io/projected/ffceea22-78f2-4032-a23a-71b4fb7ea83e-kube-api-access-w6ks8\") pod \"packageserver-d55dfcdfc-s8j95\" (UID: \"ffceea22-78f2-4032-a23a-71b4fb7ea83e\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-s8j95" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.575083 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xshbg\" (UniqueName: \"kubernetes.io/projected/92937768-de56-4cda-816c-9d1be4c07f0f-kube-api-access-xshbg\") pod \"migrator-59844c95c7-qpxrp\" (UID: \"92937768-de56-4cda-816c-9d1be4c07f0f\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-qpxrp" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.575108 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/2fa5ee3e-29f4-43b2-a154-3c988754897a-etcd-client\") pod \"apiserver-7bbb656c7d-zzrfg\" (UID: \"2fa5ee3e-29f4-43b2-a154-3c988754897a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zzrfg" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.575140 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/744def19-0006-4ab4-9d45-8a2a39bd410a-node-bootstrap-token\") pod \"machine-config-server-fdbvz\" (UID: \"744def19-0006-4ab4-9d45-8a2a39bd410a\") " pod="openshift-machine-config-operator/machine-config-server-fdbvz" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.575164 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/46e11b34-d0dd-4ea6-bed7-66eb707c7f92-config\") pod \"service-ca-operator-777779d784-5blgj\" (UID: \"46e11b34-d0dd-4ea6-bed7-66eb707c7f92\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-5blgj" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.575201 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/2a76ccaa-9249-46ba-bbc5-dbbbc33b06bb-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-fn7dj\" (UID: \"2a76ccaa-9249-46ba-bbc5-dbbbc33b06bb\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-fn7dj" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.575231 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jgp2m\" (UniqueName: \"kubernetes.io/projected/2a0286c5-9866-4a1c-9855-49cea5d04d69-kube-api-access-jgp2m\") pod \"machine-config-operator-74547568cd-vvtbh\" (UID: \"2a0286c5-9866-4a1c-9855-49cea5d04d69\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vvtbh" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.575256 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/d0546787-f9de-4b06-ab4f-0e1632d36230-srv-cert\") pod \"catalog-operator-68c6474976-2xznh\" (UID: \"d0546787-f9de-4b06-ab4f-0e1632d36230\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-2xznh" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.575309 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8l8gd\" (UID: \"c2f6e84b-a28c-4ead-97f5-978b0639e5bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-8l8gd" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.575335 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e81889bb-3225-458e-bce5-cfbafe6f4f92-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-xr4mg\" (UID: \"e81889bb-3225-458e-bce5-cfbafe6f4f92\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-xr4mg" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.575372 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l2lvw\" (UniqueName: \"kubernetes.io/projected/883a0642-0302-4fc5-b361-e25f1aba0878-kube-api-access-l2lvw\") pod \"ingress-canary-vpg8w\" (UID: \"883a0642-0302-4fc5-b361-e25f1aba0878\") " pod="openshift-ingress-canary/ingress-canary-vpg8w" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.575396 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/af8fbf5a-f90e-4d22-ae65-05d8d8253308-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-8z68k\" (UID: \"af8fbf5a-f90e-4d22-ae65-05d8d8253308\") " pod="openshift-authentication/oauth-openshift-558db77b4-8z68k" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.575424 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2fa5ee3e-29f4-43b2-a154-3c988754897a-audit-dir\") pod \"apiserver-7bbb656c7d-zzrfg\" (UID: \"2fa5ee3e-29f4-43b2-a154-3c988754897a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zzrfg" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.575680 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/ff05ab6e-248f-445c-9856-c1a2868f8840-signing-cabundle\") pod \"service-ca-9c57cc56f-8rxsm\" (UID: \"ff05ab6e-248f-445c-9856-c1a2868f8840\") " pod="openshift-service-ca/service-ca-9c57cc56f-8rxsm" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.575755 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/a1d6c12e-ff5a-4687-80ca-5699dd1bb9fe-registration-dir\") pod \"csi-hostpathplugin-nk67f\" (UID: \"a1d6c12e-ff5a-4687-80ca-5699dd1bb9fe\") " pod="hostpath-provisioner/csi-hostpathplugin-nk67f" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.575802 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/26536a6b-3639-46ad-b024-5e21b7118fe7-config-volume\") pod \"collect-profiles-29329080-9kpdv\" (UID: \"26536a6b-3639-46ad-b024-5e21b7118fe7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329080-9kpdv" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.575886 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/2fa5ee3e-29f4-43b2-a154-3c988754897a-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-zzrfg\" (UID: \"2fa5ee3e-29f4-43b2-a154-3c988754897a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zzrfg" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.575910 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6km6j\" (UniqueName: \"kubernetes.io/projected/2fa5ee3e-29f4-43b2-a154-3c988754897a-kube-api-access-6km6j\") pod \"apiserver-7bbb656c7d-zzrfg\" (UID: \"2fa5ee3e-29f4-43b2-a154-3c988754897a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zzrfg" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.575942 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/af8fbf5a-f90e-4d22-ae65-05d8d8253308-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-8z68k\" (UID: \"af8fbf5a-f90e-4d22-ae65-05d8d8253308\") " pod="openshift-authentication/oauth-openshift-558db77b4-8z68k" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.575968 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/2a0286c5-9866-4a1c-9855-49cea5d04d69-images\") pod \"machine-config-operator-74547568cd-vvtbh\" (UID: \"2a0286c5-9866-4a1c-9855-49cea5d04d69\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vvtbh" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.576959 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/26536a6b-3639-46ad-b024-5e21b7118fe7-secret-volume\") pod \"collect-profiles-29329080-9kpdv\" (UID: \"26536a6b-3639-46ad-b024-5e21b7118fe7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329080-9kpdv" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.577015 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/af8fbf5a-f90e-4d22-ae65-05d8d8253308-audit-policies\") pod \"oauth-openshift-558db77b4-8z68k\" (UID: \"af8fbf5a-f90e-4d22-ae65-05d8d8253308\") " pod="openshift-authentication/oauth-openshift-558db77b4-8z68k" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.577051 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/883a0642-0302-4fc5-b361-e25f1aba0878-cert\") pod \"ingress-canary-vpg8w\" (UID: \"883a0642-0302-4fc5-b361-e25f1aba0878\") " pod="openshift-ingress-canary/ingress-canary-vpg8w" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.577076 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/a1d6c12e-ff5a-4687-80ca-5699dd1bb9fe-csi-data-dir\") pod \"csi-hostpathplugin-nk67f\" (UID: \"a1d6c12e-ff5a-4687-80ca-5699dd1bb9fe\") " pod="hostpath-provisioner/csi-hostpathplugin-nk67f" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.577106 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/ff05ab6e-248f-445c-9856-c1a2868f8840-signing-key\") pod \"service-ca-9c57cc56f-8rxsm\" (UID: \"ff05ab6e-248f-445c-9856-c1a2868f8840\") " pod="openshift-service-ca/service-ca-9c57cc56f-8rxsm" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.577204 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4rsfz\" (UniqueName: \"kubernetes.io/projected/af8fbf5a-f90e-4d22-ae65-05d8d8253308-kube-api-access-4rsfz\") pod \"oauth-openshift-558db77b4-8z68k\" (UID: \"af8fbf5a-f90e-4d22-ae65-05d8d8253308\") " pod="openshift-authentication/oauth-openshift-558db77b4-8z68k" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.577235 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/af8fbf5a-f90e-4d22-ae65-05d8d8253308-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-8z68k\" (UID: \"af8fbf5a-f90e-4d22-ae65-05d8d8253308\") " pod="openshift-authentication/oauth-openshift-558db77b4-8z68k" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.577260 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/744def19-0006-4ab4-9d45-8a2a39bd410a-certs\") pod \"machine-config-server-fdbvz\" (UID: \"744def19-0006-4ab4-9d45-8a2a39bd410a\") " pod="openshift-machine-config-operator/machine-config-server-fdbvz" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.577287 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e853926-f1e6-4571-863b-97918e660677-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-vn26d\" (UID: \"8e853926-f1e6-4571-863b-97918e660677\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-vn26d" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.577312 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8e853926-f1e6-4571-863b-97918e660677-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-vn26d\" (UID: \"8e853926-f1e6-4571-863b-97918e660677\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-vn26d" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.577357 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xvhtm\" (UniqueName: \"kubernetes.io/projected/e81889bb-3225-458e-bce5-cfbafe6f4f92-kube-api-access-xvhtm\") pod \"machine-config-controller-84d6567774-xr4mg\" (UID: \"e81889bb-3225-458e-bce5-cfbafe6f4f92\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-xr4mg" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.577384 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2fa5ee3e-29f4-43b2-a154-3c988754897a-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-zzrfg\" (UID: \"2fa5ee3e-29f4-43b2-a154-3c988754897a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zzrfg" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.577408 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/2fa5ee3e-29f4-43b2-a154-3c988754897a-encryption-config\") pod \"apiserver-7bbb656c7d-zzrfg\" (UID: \"2fa5ee3e-29f4-43b2-a154-3c988754897a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zzrfg" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.577433 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c682e90b-3a5a-4b87-996b-8365b0936bda-config-volume\") pod \"dns-default-dr4mr\" (UID: \"c682e90b-3a5a-4b87-996b-8365b0936bda\") " pod="openshift-dns/dns-default-dr4mr" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.577499 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-26z5m\" (UniqueName: \"kubernetes.io/projected/678eeee3-60d8-4dbe-b815-10de5a46c4d1-kube-api-access-26z5m\") pod \"multus-admission-controller-857f4d67dd-56948\" (UID: \"678eeee3-60d8-4dbe-b815-10de5a46c4d1\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-56948" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.577553 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lzv6l\" (UniqueName: \"kubernetes.io/projected/2a76ccaa-9249-46ba-bbc5-dbbbc33b06bb-kube-api-access-lzv6l\") pod \"control-plane-machine-set-operator-78cbb6b69f-fn7dj\" (UID: \"2a76ccaa-9249-46ba-bbc5-dbbbc33b06bb\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-fn7dj" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.577579 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/982129a4-11b4-4024-afb0-ac6024165c7a-srv-cert\") pod \"olm-operator-6b444d44fb-6tt2t\" (UID: \"982129a4-11b4-4024-afb0-ac6024165c7a\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6tt2t" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.577606 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wz4zb\" (UniqueName: \"kubernetes.io/projected/982129a4-11b4-4024-afb0-ac6024165c7a-kube-api-access-wz4zb\") pod \"olm-operator-6b444d44fb-6tt2t\" (UID: \"982129a4-11b4-4024-afb0-ac6024165c7a\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6tt2t" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.577644 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2fa5ee3e-29f4-43b2-a154-3c988754897a-serving-cert\") pod \"apiserver-7bbb656c7d-zzrfg\" (UID: \"2fa5ee3e-29f4-43b2-a154-3c988754897a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zzrfg" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.577670 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cjzws\" (UniqueName: \"kubernetes.io/projected/26536a6b-3639-46ad-b024-5e21b7118fe7-kube-api-access-cjzws\") pod \"collect-profiles-29329080-9kpdv\" (UID: \"26536a6b-3639-46ad-b024-5e21b7118fe7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329080-9kpdv" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.577702 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rnnkq\" (UniqueName: \"kubernetes.io/projected/d2595750-aae8-4f17-8782-b7eccb0c2948-kube-api-access-rnnkq\") pod \"marketplace-operator-79b997595-qq4zz\" (UID: \"d2595750-aae8-4f17-8782-b7eccb0c2948\") " pod="openshift-marketplace/marketplace-operator-79b997595-qq4zz" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.577730 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tfjz5\" (UniqueName: \"kubernetes.io/projected/744def19-0006-4ab4-9d45-8a2a39bd410a-kube-api-access-tfjz5\") pod \"machine-config-server-fdbvz\" (UID: \"744def19-0006-4ab4-9d45-8a2a39bd410a\") " pod="openshift-machine-config-operator/machine-config-server-fdbvz" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.577762 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dpf26\" (UniqueName: \"kubernetes.io/projected/d40e2672-6c82-45e6-a292-21615712032d-kube-api-access-dpf26\") pod \"kube-storage-version-migrator-operator-b67b599dd-wp6kv\" (UID: \"d40e2672-6c82-45e6-a292-21615712032d\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-wp6kv" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.577793 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/2fa5ee3e-29f4-43b2-a154-3c988754897a-audit-policies\") pod \"apiserver-7bbb656c7d-zzrfg\" (UID: \"2fa5ee3e-29f4-43b2-a154-3c988754897a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zzrfg" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.577828 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gx2z5\" (UniqueName: \"kubernetes.io/projected/9613dce2-6a39-4480-8b7f-553288ebc45b-kube-api-access-gx2z5\") pod \"package-server-manager-789f6589d5-px5fp\" (UID: \"9613dce2-6a39-4480-8b7f-553288ebc45b\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-px5fp" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.581624 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/af8fbf5a-f90e-4d22-ae65-05d8d8253308-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-8z68k\" (UID: \"af8fbf5a-f90e-4d22-ae65-05d8d8253308\") " pod="openshift-authentication/oauth-openshift-558db77b4-8z68k" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.581725 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8e853926-f1e6-4571-863b-97918e660677-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-vn26d\" (UID: \"8e853926-f1e6-4571-863b-97918e660677\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-vn26d" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.582720 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e81889bb-3225-458e-bce5-cfbafe6f4f92-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-xr4mg\" (UID: \"e81889bb-3225-458e-bce5-cfbafe6f4f92\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-xr4mg" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.584851 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/46e11b34-d0dd-4ea6-bed7-66eb707c7f92-serving-cert\") pod \"service-ca-operator-777779d784-5blgj\" (UID: \"46e11b34-d0dd-4ea6-bed7-66eb707c7f92\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-5blgj" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.586120 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/a1d6c12e-ff5a-4687-80ca-5699dd1bb9fe-mountpoint-dir\") pod \"csi-hostpathplugin-nk67f\" (UID: \"a1d6c12e-ff5a-4687-80ca-5699dd1bb9fe\") " pod="hostpath-provisioner/csi-hostpathplugin-nk67f" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.586202 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/ffceea22-78f2-4032-a23a-71b4fb7ea83e-apiservice-cert\") pod \"packageserver-d55dfcdfc-s8j95\" (UID: \"ffceea22-78f2-4032-a23a-71b4fb7ea83e\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-s8j95" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.586747 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/a1d6c12e-ff5a-4687-80ca-5699dd1bb9fe-socket-dir\") pod \"csi-hostpathplugin-nk67f\" (UID: \"a1d6c12e-ff5a-4687-80ca-5699dd1bb9fe\") " pod="hostpath-provisioner/csi-hostpathplugin-nk67f" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.587549 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ffceea22-78f2-4032-a23a-71b4fb7ea83e-webhook-cert\") pod \"packageserver-d55dfcdfc-s8j95\" (UID: \"ffceea22-78f2-4032-a23a-71b4fb7ea83e\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-s8j95" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.587713 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/678eeee3-60d8-4dbe-b815-10de5a46c4d1-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-56948\" (UID: \"678eeee3-60d8-4dbe-b815-10de5a46c4d1\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-56948" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.587788 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/af8fbf5a-f90e-4d22-ae65-05d8d8253308-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-8z68k\" (UID: \"af8fbf5a-f90e-4d22-ae65-05d8d8253308\") " pod="openshift-authentication/oauth-openshift-558db77b4-8z68k" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.588007 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/af8fbf5a-f90e-4d22-ae65-05d8d8253308-audit-dir\") pod \"oauth-openshift-558db77b4-8z68k\" (UID: \"af8fbf5a-f90e-4d22-ae65-05d8d8253308\") " pod="openshift-authentication/oauth-openshift-558db77b4-8z68k" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.590176 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e853926-f1e6-4571-863b-97918e660677-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-vn26d\" (UID: \"8e853926-f1e6-4571-863b-97918e660677\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-vn26d" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.590479 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/af8fbf5a-f90e-4d22-ae65-05d8d8253308-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-8z68k\" (UID: \"af8fbf5a-f90e-4d22-ae65-05d8d8253308\") " pod="openshift-authentication/oauth-openshift-558db77b4-8z68k" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.590647 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2fa5ee3e-29f4-43b2-a154-3c988754897a-audit-dir\") pod \"apiserver-7bbb656c7d-zzrfg\" (UID: \"2fa5ee3e-29f4-43b2-a154-3c988754897a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zzrfg" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.590754 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/a1d6c12e-ff5a-4687-80ca-5699dd1bb9fe-registration-dir\") pod \"csi-hostpathplugin-nk67f\" (UID: \"a1d6c12e-ff5a-4687-80ca-5699dd1bb9fe\") " pod="hostpath-provisioner/csi-hostpathplugin-nk67f" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.590971 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2fa5ee3e-29f4-43b2-a154-3c988754897a-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-zzrfg\" (UID: \"2fa5ee3e-29f4-43b2-a154-3c988754897a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zzrfg" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.591241 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/af8fbf5a-f90e-4d22-ae65-05d8d8253308-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-8z68k\" (UID: \"af8fbf5a-f90e-4d22-ae65-05d8d8253308\") " pod="openshift-authentication/oauth-openshift-558db77b4-8z68k" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.591949 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2qj4r\" (UniqueName: \"kubernetes.io/projected/dd50f0ac-ac97-413f-a350-7496681f7f0a-kube-api-access-2qj4r\") pod \"route-controller-manager-6576b87f9c-wk4bd\" (UID: \"dd50f0ac-ac97-413f-a350-7496681f7f0a\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wk4bd" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.592655 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/ffceea22-78f2-4032-a23a-71b4fb7ea83e-tmpfs\") pod \"packageserver-d55dfcdfc-s8j95\" (UID: \"ffceea22-78f2-4032-a23a-71b4fb7ea83e\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-s8j95" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.592666 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/26536a6b-3639-46ad-b024-5e21b7118fe7-config-volume\") pod \"collect-profiles-29329080-9kpdv\" (UID: \"26536a6b-3639-46ad-b024-5e21b7118fe7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329080-9kpdv" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.593829 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/2fa5ee3e-29f4-43b2-a154-3c988754897a-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-zzrfg\" (UID: \"2fa5ee3e-29f4-43b2-a154-3c988754897a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zzrfg" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.594181 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d40e2672-6c82-45e6-a292-21615712032d-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-wp6kv\" (UID: \"d40e2672-6c82-45e6-a292-21615712032d\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-wp6kv" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.594256 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/a1d6c12e-ff5a-4687-80ca-5699dd1bb9fe-plugins-dir\") pod \"csi-hostpathplugin-nk67f\" (UID: \"a1d6c12e-ff5a-4687-80ca-5699dd1bb9fe\") " pod="hostpath-provisioner/csi-hostpathplugin-nk67f" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.594284 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c682e90b-3a5a-4b87-996b-8365b0936bda-config-volume\") pod \"dns-default-dr4mr\" (UID: \"c682e90b-3a5a-4b87-996b-8365b0936bda\") " pod="openshift-dns/dns-default-dr4mr" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.594853 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/af8fbf5a-f90e-4d22-ae65-05d8d8253308-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-8z68k\" (UID: \"af8fbf5a-f90e-4d22-ae65-05d8d8253308\") " pod="openshift-authentication/oauth-openshift-558db77b4-8z68k" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.595225 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/af8fbf5a-f90e-4d22-ae65-05d8d8253308-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-8z68k\" (UID: \"af8fbf5a-f90e-4d22-ae65-05d8d8253308\") " pod="openshift-authentication/oauth-openshift-558db77b4-8z68k" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.595711 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/2fa5ee3e-29f4-43b2-a154-3c988754897a-encryption-config\") pod \"apiserver-7bbb656c7d-zzrfg\" (UID: \"2fa5ee3e-29f4-43b2-a154-3c988754897a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zzrfg" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.596730 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/2a0286c5-9866-4a1c-9855-49cea5d04d69-auth-proxy-config\") pod \"machine-config-operator-74547568cd-vvtbh\" (UID: \"2a0286c5-9866-4a1c-9855-49cea5d04d69\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vvtbh" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.596778 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/46e11b34-d0dd-4ea6-bed7-66eb707c7f92-config\") pod \"service-ca-operator-777779d784-5blgj\" (UID: \"46e11b34-d0dd-4ea6-bed7-66eb707c7f92\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-5blgj" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.597133 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/2fa5ee3e-29f4-43b2-a154-3c988754897a-etcd-client\") pod \"apiserver-7bbb656c7d-zzrfg\" (UID: \"2fa5ee3e-29f4-43b2-a154-3c988754897a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zzrfg" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.576801 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d2595750-aae8-4f17-8782-b7eccb0c2948-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-qq4zz\" (UID: \"d2595750-aae8-4f17-8782-b7eccb0c2948\") " pod="openshift-marketplace/marketplace-operator-79b997595-qq4zz" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.598003 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/2a0286c5-9866-4a1c-9855-49cea5d04d69-images\") pod \"machine-config-operator-74547568cd-vvtbh\" (UID: \"2a0286c5-9866-4a1c-9855-49cea5d04d69\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vvtbh" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.598215 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/2fa5ee3e-29f4-43b2-a154-3c988754897a-audit-policies\") pod \"apiserver-7bbb656c7d-zzrfg\" (UID: \"2fa5ee3e-29f4-43b2-a154-3c988754897a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zzrfg" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.598218 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d40e2672-6c82-45e6-a292-21615712032d-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-wp6kv\" (UID: \"d40e2672-6c82-45e6-a292-21615712032d\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-wp6kv" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.598701 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/744def19-0006-4ab4-9d45-8a2a39bd410a-certs\") pod \"machine-config-server-fdbvz\" (UID: \"744def19-0006-4ab4-9d45-8a2a39bd410a\") " pod="openshift-machine-config-operator/machine-config-server-fdbvz" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.598861 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/a1d6c12e-ff5a-4687-80ca-5699dd1bb9fe-csi-data-dir\") pod \"csi-hostpathplugin-nk67f\" (UID: \"a1d6c12e-ff5a-4687-80ca-5699dd1bb9fe\") " pod="hostpath-provisioner/csi-hostpathplugin-nk67f" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.599290 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/af8fbf5a-f90e-4d22-ae65-05d8d8253308-audit-policies\") pod \"oauth-openshift-558db77b4-8z68k\" (UID: \"af8fbf5a-f90e-4d22-ae65-05d8d8253308\") " pod="openshift-authentication/oauth-openshift-558db77b4-8z68k" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.599413 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/af8fbf5a-f90e-4d22-ae65-05d8d8253308-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-8z68k\" (UID: \"af8fbf5a-f90e-4d22-ae65-05d8d8253308\") " pod="openshift-authentication/oauth-openshift-558db77b4-8z68k" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.599725 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/d2595750-aae8-4f17-8782-b7eccb0c2948-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-qq4zz\" (UID: \"d2595750-aae8-4f17-8782-b7eccb0c2948\") " pod="openshift-marketplace/marketplace-operator-79b997595-qq4zz" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.600071 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/883a0642-0302-4fc5-b361-e25f1aba0878-cert\") pod \"ingress-canary-vpg8w\" (UID: \"883a0642-0302-4fc5-b361-e25f1aba0878\") " pod="openshift-ingress-canary/ingress-canary-vpg8w" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.601039 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/af8fbf5a-f90e-4d22-ae65-05d8d8253308-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-8z68k\" (UID: \"af8fbf5a-f90e-4d22-ae65-05d8d8253308\") " pod="openshift-authentication/oauth-openshift-558db77b4-8z68k" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.601360 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/af8fbf5a-f90e-4d22-ae65-05d8d8253308-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-8z68k\" (UID: \"af8fbf5a-f90e-4d22-ae65-05d8d8253308\") " pod="openshift-authentication/oauth-openshift-558db77b4-8z68k" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.602273 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/26536a6b-3639-46ad-b024-5e21b7118fe7-secret-volume\") pod \"collect-profiles-29329080-9kpdv\" (UID: \"26536a6b-3639-46ad-b024-5e21b7118fe7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329080-9kpdv" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.602446 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/d0546787-f9de-4b06-ab4f-0e1632d36230-srv-cert\") pod \"catalog-operator-68c6474976-2xznh\" (UID: \"d0546787-f9de-4b06-ab4f-0e1632d36230\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-2xznh" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.604891 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/2a76ccaa-9249-46ba-bbc5-dbbbc33b06bb-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-fn7dj\" (UID: \"2a76ccaa-9249-46ba-bbc5-dbbbc33b06bb\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-fn7dj" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.605821 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/e81889bb-3225-458e-bce5-cfbafe6f4f92-proxy-tls\") pod \"machine-config-controller-84d6567774-xr4mg\" (UID: \"e81889bb-3225-458e-bce5-cfbafe6f4f92\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-xr4mg" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.606161 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/af8fbf5a-f90e-4d22-ae65-05d8d8253308-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-8z68k\" (UID: \"af8fbf5a-f90e-4d22-ae65-05d8d8253308\") " pod="openshift-authentication/oauth-openshift-558db77b4-8z68k" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.606281 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/982129a4-11b4-4024-afb0-ac6024165c7a-srv-cert\") pod \"olm-operator-6b444d44fb-6tt2t\" (UID: \"982129a4-11b4-4024-afb0-ac6024165c7a\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6tt2t" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.606842 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/d0546787-f9de-4b06-ab4f-0e1632d36230-profile-collector-cert\") pod \"catalog-operator-68c6474976-2xznh\" (UID: \"d0546787-f9de-4b06-ab4f-0e1632d36230\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-2xznh" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.607041 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/ff05ab6e-248f-445c-9856-c1a2868f8840-signing-key\") pod \"service-ca-9c57cc56f-8rxsm\" (UID: \"ff05ab6e-248f-445c-9856-c1a2868f8840\") " pod="openshift-service-ca/service-ca-9c57cc56f-8rxsm" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.607105 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2fa5ee3e-29f4-43b2-a154-3c988754897a-serving-cert\") pod \"apiserver-7bbb656c7d-zzrfg\" (UID: \"2fa5ee3e-29f4-43b2-a154-3c988754897a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zzrfg" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.607167 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/2a0286c5-9866-4a1c-9855-49cea5d04d69-proxy-tls\") pod \"machine-config-operator-74547568cd-vvtbh\" (UID: \"2a0286c5-9866-4a1c-9855-49cea5d04d69\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vvtbh" Oct 06 10:00:42 crc kubenswrapper[4824]: E1006 10:00:42.607293 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-06 10:00:43.107266766 +0000 UTC m=+152.471689837 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8l8gd" (UID: "c2f6e84b-a28c-4ead-97f5-978b0639e5bf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.607606 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/982129a4-11b4-4024-afb0-ac6024165c7a-profile-collector-cert\") pod \"olm-operator-6b444d44fb-6tt2t\" (UID: \"982129a4-11b4-4024-afb0-ac6024165c7a\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6tt2t" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.607625 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/9613dce2-6a39-4480-8b7f-553288ebc45b-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-px5fp\" (UID: \"9613dce2-6a39-4480-8b7f-553288ebc45b\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-px5fp" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.610226 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/744def19-0006-4ab4-9d45-8a2a39bd410a-node-bootstrap-token\") pod \"machine-config-server-fdbvz\" (UID: \"744def19-0006-4ab4-9d45-8a2a39bd410a\") " pod="openshift-machine-config-operator/machine-config-server-fdbvz" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.610716 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/af8fbf5a-f90e-4d22-ae65-05d8d8253308-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-8z68k\" (UID: \"af8fbf5a-f90e-4d22-ae65-05d8d8253308\") " pod="openshift-authentication/oauth-openshift-558db77b4-8z68k" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.614506 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/c682e90b-3a5a-4b87-996b-8365b0936bda-metrics-tls\") pod \"dns-default-dr4mr\" (UID: \"c682e90b-3a5a-4b87-996b-8365b0936bda\") " pod="openshift-dns/dns-default-dr4mr" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.615413 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ft4p7\" (UniqueName: \"kubernetes.io/projected/4d7b1b9b-33de-4fca-b629-6790c3a5cb7d-kube-api-access-ft4p7\") pod \"cluster-samples-operator-665b6dd947-hqhlr\" (UID: \"4d7b1b9b-33de-4fca-b629-6790c3a5cb7d\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-hqhlr" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.627865 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l2n5j\" (UniqueName: \"kubernetes.io/projected/d6e255a4-d3e7-4cef-bf40-f98aa11b3e3b-kube-api-access-l2n5j\") pod \"router-default-5444994796-7qwsv\" (UID: \"d6e255a4-d3e7-4cef-bf40-f98aa11b3e3b\") " pod="openshift-ingress/router-default-5444994796-7qwsv" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.644934 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4lbvd\" (UniqueName: \"kubernetes.io/projected/3bdcba24-670d-427a-bfd6-4372c9f944a2-kube-api-access-4lbvd\") pod \"ingress-operator-5b745b69d9-7p7wb\" (UID: \"3bdcba24-670d-427a-bfd6-4372c9f944a2\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7p7wb" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.653303 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-nqmr6" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.660124 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-hqhlr" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.663934 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4bceefc5-1905-48bd-b214-ab5824f0a302-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-mdv87\" (UID: \"4bceefc5-1905-48bd-b214-ab5824f0a302\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-mdv87" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.672737 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-s4ht9"] Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.673664 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-sb9c8" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.679408 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 10:00:42 crc kubenswrapper[4824]: E1006 10:00:42.679547 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-06 10:00:43.179512806 +0000 UTC m=+152.543935667 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.681226 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8l8gd\" (UID: \"c2f6e84b-a28c-4ead-97f5-978b0639e5bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-8l8gd" Oct 06 10:00:42 crc kubenswrapper[4824]: E1006 10:00:42.681750 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-06 10:00:43.18172769 +0000 UTC m=+152.546150551 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8l8gd" (UID: "c2f6e84b-a28c-4ead-97f5-978b0639e5bf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 10:00:42 crc kubenswrapper[4824]: W1006 10:00:42.683843 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod171abefb_ea19_40fa_8a22_4b38956d5e4b.slice/crio-9965d3b9e59d06c71f910614d088567eccc92e02213049542dcac79a41e018cd WatchSource:0}: Error finding container 9965d3b9e59d06c71f910614d088567eccc92e02213049542dcac79a41e018cd: Status 404 returned error can't find the container with id 9965d3b9e59d06c71f910614d088567eccc92e02213049542dcac79a41e018cd Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.704430 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-94hbm\" (UniqueName: \"kubernetes.io/projected/7c497a24-d8cb-4560-993f-fc73b3af8c04-kube-api-access-94hbm\") pod \"openshift-config-operator-7777fb866f-v6gnw\" (UID: \"7c497a24-d8cb-4560-993f-fc73b3af8c04\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-v6gnw" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.716580 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c2f6e84b-a28c-4ead-97f5-978b0639e5bf-bound-sa-token\") pod \"image-registry-697d97f7c8-8l8gd\" (UID: \"c2f6e84b-a28c-4ead-97f5-978b0639e5bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-8l8gd" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.723643 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/3bdcba24-670d-427a-bfd6-4372c9f944a2-bound-sa-token\") pod \"ingress-operator-5b745b69d9-7p7wb\" (UID: \"3bdcba24-670d-427a-bfd6-4372c9f944a2\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7p7wb" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.749258 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/45682655-0e98-4a5e-b5e8-308f4d0f7d9a-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-prthm\" (UID: \"45682655-0e98-4a5e-b5e8-308f4d0f7d9a\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-prthm" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.772712 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2c9tb\" (UniqueName: \"kubernetes.io/projected/d21f4737-c582-41e9-846c-dd75ec173afd-kube-api-access-2c9tb\") pod \"cluster-image-registry-operator-dc59b4c8b-dklh2\" (UID: \"d21f4737-c582-41e9-846c-dd75ec173afd\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-dklh2" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.779672 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-q76nt" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.782294 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 10:00:42 crc kubenswrapper[4824]: E1006 10:00:42.782813 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-06 10:00:43.282793241 +0000 UTC m=+152.647216102 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.797838 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nqdpp\" (UniqueName: \"kubernetes.io/projected/722b7a46-52a1-49a7-a053-803334dc5964-kube-api-access-nqdpp\") pod \"console-f9d7485db-fpxws\" (UID: \"722b7a46-52a1-49a7-a053-803334dc5964\") " pod="openshift-console/console-f9d7485db-fpxws" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.806112 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-48kwg\" (UniqueName: \"kubernetes.io/projected/b349b0af-0ba8-409f-a915-7fb06b2e3774-kube-api-access-48kwg\") pod \"machine-api-operator-5694c8668f-fjdwv\" (UID: \"b349b0af-0ba8-409f-a915-7fb06b2e3774\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-fjdwv" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.810508 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-v6gnw" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.824286 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-mdv87" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.825896 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z44hr\" (UniqueName: \"kubernetes.io/projected/9da4a453-8e05-4360-b74c-ad1c4706b005-kube-api-access-z44hr\") pod \"machine-approver-56656f9798-4hw5p\" (UID: \"9da4a453-8e05-4360-b74c-ad1c4706b005\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-4hw5p" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.835142 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-fjdwv" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.841451 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wk4bd" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.844559 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d21f4737-c582-41e9-846c-dd75ec173afd-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-dklh2\" (UID: \"d21f4737-c582-41e9-846c-dd75ec173afd\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-dklh2" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.849566 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-7qwsv" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.867002 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gvtfv\" (UniqueName: \"kubernetes.io/projected/7532dafc-fb61-4066-bb24-2470e4b2cfe6-kube-api-access-gvtfv\") pod \"dns-operator-744455d44c-6kn7b\" (UID: \"7532dafc-fb61-4066-bb24-2470e4b2cfe6\") " pod="openshift-dns-operator/dns-operator-744455d44c-6kn7b" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.888867 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vz27c\" (UniqueName: \"kubernetes.io/projected/22545b78-29bb-45aa-a98c-fd7ce9703239-kube-api-access-vz27c\") pod \"console-operator-58897d9998-bjzzk\" (UID: \"22545b78-29bb-45aa-a98c-fd7ce9703239\") " pod="openshift-console-operator/console-operator-58897d9998-bjzzk" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.905384 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7p7wb" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.906498 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8l8gd\" (UID: \"c2f6e84b-a28c-4ead-97f5-978b0639e5bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-8l8gd" Oct 06 10:00:42 crc kubenswrapper[4824]: E1006 10:00:42.906880 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-06 10:00:43.406861027 +0000 UTC m=+152.771283888 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8l8gd" (UID: "c2f6e84b-a28c-4ead-97f5-978b0639e5bf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.907320 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-nqmr6"] Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.907385 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-fpxws" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.910749 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6t6pf\" (UniqueName: \"kubernetes.io/projected/9b1f3faf-72a1-44e2-814b-964902e78d3e-kube-api-access-6t6pf\") pod \"openshift-controller-manager-operator-756b6f6bc6-wlbrp\" (UID: \"9b1f3faf-72a1-44e2-814b-964902e78d3e\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wlbrp" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.920219 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-dklh2" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.927485 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-sb9c8"] Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.929850 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-prthm" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.939679 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-hqhlr"] Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.942173 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z8t7q\" (UniqueName: \"kubernetes.io/projected/323631a8-107c-4fd0-8b57-04550f0fc7d1-kube-api-access-z8t7q\") pod \"authentication-operator-69f744f599-j2m8k\" (UID: \"323631a8-107c-4fd0-8b57-04550f0fc7d1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-j2m8k" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.958453 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jnc4f\" (UniqueName: \"kubernetes.io/projected/a47e3600-79be-46c3-a1b8-6678d8da80c5-kube-api-access-jnc4f\") pod \"etcd-operator-b45778765-x4fdc\" (UID: \"a47e3600-79be-46c3-a1b8-6678d8da80c5\") " pod="openshift-etcd-operator/etcd-operator-b45778765-x4fdc" Oct 06 10:00:42 crc kubenswrapper[4824]: W1006 10:00:42.960078 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb6f7d3cb_418e_4f73_95a0_2477e37976c3.slice/crio-243447b3c3ad6710011f6daf375f053746d337edca1d1ec2b5c40e13fb029dc6 WatchSource:0}: Error finding container 243447b3c3ad6710011f6daf375f053746d337edca1d1ec2b5c40e13fb029dc6: Status 404 returned error can't find the container with id 243447b3c3ad6710011f6daf375f053746d337edca1d1ec2b5c40e13fb029dc6 Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.965964 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-j2m8k" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.968959 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-bjzzk" Oct 06 10:00:42 crc kubenswrapper[4824]: I1006 10:00:42.984539 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gx2z5\" (UniqueName: \"kubernetes.io/projected/9613dce2-6a39-4480-8b7f-553288ebc45b-kube-api-access-gx2z5\") pod \"package-server-manager-789f6589d5-px5fp\" (UID: \"9613dce2-6a39-4480-8b7f-553288ebc45b\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-px5fp" Oct 06 10:00:43 crc kubenswrapper[4824]: I1006 10:00:42.998125 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-6kn7b" Oct 06 10:00:43 crc kubenswrapper[4824]: W1006 10:00:43.000772 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc8fc8c20_e06a_4039_80c0_82f84ea81391.slice/crio-d073584d2516c38b683792a5d19bbaa7da5053a1e97d3e73491319a15593f2bc WatchSource:0}: Error finding container d073584d2516c38b683792a5d19bbaa7da5053a1e97d3e73491319a15593f2bc: Status 404 returned error can't find the container with id d073584d2516c38b683792a5d19bbaa7da5053a1e97d3e73491319a15593f2bc Oct 06 10:00:43 crc kubenswrapper[4824]: I1006 10:00:43.006994 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xsfd8\" (UniqueName: \"kubernetes.io/projected/ff05ab6e-248f-445c-9856-c1a2868f8840-kube-api-access-xsfd8\") pod \"service-ca-9c57cc56f-8rxsm\" (UID: \"ff05ab6e-248f-445c-9856-c1a2868f8840\") " pod="openshift-service-ca/service-ca-9c57cc56f-8rxsm" Oct 06 10:00:43 crc kubenswrapper[4824]: I1006 10:00:43.007475 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 10:00:43 crc kubenswrapper[4824]: E1006 10:00:43.008175 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-06 10:00:43.508153823 +0000 UTC m=+152.872576694 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 10:00:43 crc kubenswrapper[4824]: I1006 10:00:43.035247 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t2gcn\" (UniqueName: \"kubernetes.io/projected/d0546787-f9de-4b06-ab4f-0e1632d36230-kube-api-access-t2gcn\") pod \"catalog-operator-68c6474976-2xznh\" (UID: \"d0546787-f9de-4b06-ab4f-0e1632d36230\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-2xznh" Oct 06 10:00:43 crc kubenswrapper[4824]: I1006 10:00:43.045937 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-2xznh" Oct 06 10:00:43 crc kubenswrapper[4824]: I1006 10:00:43.049784 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l2lvw\" (UniqueName: \"kubernetes.io/projected/883a0642-0302-4fc5-b361-e25f1aba0878-kube-api-access-l2lvw\") pod \"ingress-canary-vpg8w\" (UID: \"883a0642-0302-4fc5-b361-e25f1aba0878\") " pod="openshift-ingress-canary/ingress-canary-vpg8w" Oct 06 10:00:43 crc kubenswrapper[4824]: I1006 10:00:43.072012 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lzv6l\" (UniqueName: \"kubernetes.io/projected/2a76ccaa-9249-46ba-bbc5-dbbbc33b06bb-kube-api-access-lzv6l\") pod \"control-plane-machine-set-operator-78cbb6b69f-fn7dj\" (UID: \"2a76ccaa-9249-46ba-bbc5-dbbbc33b06bb\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-fn7dj" Oct 06 10:00:43 crc kubenswrapper[4824]: I1006 10:00:43.077071 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-4hw5p" Oct 06 10:00:43 crc kubenswrapper[4824]: I1006 10:00:43.078058 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-px5fp" Oct 06 10:00:43 crc kubenswrapper[4824]: I1006 10:00:43.085057 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-8rxsm" Oct 06 10:00:43 crc kubenswrapper[4824]: I1006 10:00:43.095868 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wlbrp" Oct 06 10:00:43 crc kubenswrapper[4824]: I1006 10:00:43.097532 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8e853926-f1e6-4571-863b-97918e660677-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-vn26d\" (UID: \"8e853926-f1e6-4571-863b-97918e660677\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-vn26d" Oct 06 10:00:43 crc kubenswrapper[4824]: I1006 10:00:43.106129 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xvhtm\" (UniqueName: \"kubernetes.io/projected/e81889bb-3225-458e-bce5-cfbafe6f4f92-kube-api-access-xvhtm\") pod \"machine-config-controller-84d6567774-xr4mg\" (UID: \"e81889bb-3225-458e-bce5-cfbafe6f4f92\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-xr4mg" Oct 06 10:00:43 crc kubenswrapper[4824]: I1006 10:00:43.109637 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8l8gd\" (UID: \"c2f6e84b-a28c-4ead-97f5-978b0639e5bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-8l8gd" Oct 06 10:00:43 crc kubenswrapper[4824]: E1006 10:00:43.110132 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-06 10:00:43.610116495 +0000 UTC m=+152.974539356 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8l8gd" (UID: "c2f6e84b-a28c-4ead-97f5-978b0639e5bf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 10:00:43 crc kubenswrapper[4824]: I1006 10:00:43.126890 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k66kw\" (UniqueName: \"kubernetes.io/projected/a1d6c12e-ff5a-4687-80ca-5699dd1bb9fe-kube-api-access-k66kw\") pod \"csi-hostpathplugin-nk67f\" (UID: \"a1d6c12e-ff5a-4687-80ca-5699dd1bb9fe\") " pod="hostpath-provisioner/csi-hostpathplugin-nk67f" Oct 06 10:00:43 crc kubenswrapper[4824]: I1006 10:00:43.142236 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-q76nt"] Oct 06 10:00:43 crc kubenswrapper[4824]: I1006 10:00:43.146966 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h7xwt\" (UniqueName: \"kubernetes.io/projected/c682e90b-3a5a-4b87-996b-8365b0936bda-kube-api-access-h7xwt\") pod \"dns-default-dr4mr\" (UID: \"c682e90b-3a5a-4b87-996b-8365b0936bda\") " pod="openshift-dns/dns-default-dr4mr" Oct 06 10:00:43 crc kubenswrapper[4824]: I1006 10:00:43.154896 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-v6gnw"] Oct 06 10:00:43 crc kubenswrapper[4824]: I1006 10:00:43.155272 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-nk67f" Oct 06 10:00:43 crc kubenswrapper[4824]: I1006 10:00:43.159508 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-dr4mr" Oct 06 10:00:43 crc kubenswrapper[4824]: I1006 10:00:43.165406 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-vpg8w" Oct 06 10:00:43 crc kubenswrapper[4824]: I1006 10:00:43.172091 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wz4zb\" (UniqueName: \"kubernetes.io/projected/982129a4-11b4-4024-afb0-ac6024165c7a-kube-api-access-wz4zb\") pod \"olm-operator-6b444d44fb-6tt2t\" (UID: \"982129a4-11b4-4024-afb0-ac6024165c7a\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6tt2t" Oct 06 10:00:43 crc kubenswrapper[4824]: I1006 10:00:43.189009 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6km6j\" (UniqueName: \"kubernetes.io/projected/2fa5ee3e-29f4-43b2-a154-3c988754897a-kube-api-access-6km6j\") pod \"apiserver-7bbb656c7d-zzrfg\" (UID: \"2fa5ee3e-29f4-43b2-a154-3c988754897a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zzrfg" Oct 06 10:00:43 crc kubenswrapper[4824]: I1006 10:00:43.195077 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-x4fdc" Oct 06 10:00:43 crc kubenswrapper[4824]: I1006 10:00:43.205378 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-99zcx\" (UniqueName: \"kubernetes.io/projected/46e11b34-d0dd-4ea6-bed7-66eb707c7f92-kube-api-access-99zcx\") pod \"service-ca-operator-777779d784-5blgj\" (UID: \"46e11b34-d0dd-4ea6-bed7-66eb707c7f92\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-5blgj" Oct 06 10:00:43 crc kubenswrapper[4824]: I1006 10:00:43.210489 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 10:00:43 crc kubenswrapper[4824]: E1006 10:00:43.211131 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-06 10:00:43.711106564 +0000 UTC m=+153.075529435 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 10:00:43 crc kubenswrapper[4824]: I1006 10:00:43.211412 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8l8gd\" (UID: \"c2f6e84b-a28c-4ead-97f5-978b0639e5bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-8l8gd" Oct 06 10:00:43 crc kubenswrapper[4824]: E1006 10:00:43.211751 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-06 10:00:43.711741069 +0000 UTC m=+153.076163930 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8l8gd" (UID: "c2f6e84b-a28c-4ead-97f5-978b0639e5bf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 10:00:43 crc kubenswrapper[4824]: I1006 10:00:43.231119 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-sb9c8" event={"ID":"c8fc8c20-e06a-4039-80c0-82f84ea81391","Type":"ContainerStarted","Data":"d073584d2516c38b683792a5d19bbaa7da5053a1e97d3e73491319a15593f2bc"} Oct 06 10:00:43 crc kubenswrapper[4824]: I1006 10:00:43.233411 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-s4ht9" event={"ID":"171abefb-ea19-40fa-8a22-4b38956d5e4b","Type":"ContainerStarted","Data":"3ba32c6b9988516b69352c9229d53e5613d4a17ca0a4d45a60d6288e5e895941"} Oct 06 10:00:43 crc kubenswrapper[4824]: I1006 10:00:43.233470 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-s4ht9" event={"ID":"171abefb-ea19-40fa-8a22-4b38956d5e4b","Type":"ContainerStarted","Data":"9965d3b9e59d06c71f910614d088567eccc92e02213049542dcac79a41e018cd"} Oct 06 10:00:43 crc kubenswrapper[4824]: I1006 10:00:43.233460 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-26z5m\" (UniqueName: \"kubernetes.io/projected/678eeee3-60d8-4dbe-b815-10de5a46c4d1-kube-api-access-26z5m\") pod \"multus-admission-controller-857f4d67dd-56948\" (UID: \"678eeee3-60d8-4dbe-b815-10de5a46c4d1\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-56948" Oct 06 10:00:43 crc kubenswrapper[4824]: I1006 10:00:43.233640 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-s4ht9" Oct 06 10:00:43 crc kubenswrapper[4824]: I1006 10:00:43.236801 4824 patch_prober.go:28] interesting pod/downloads-7954f5f757-s4ht9 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.7:8080/\": dial tcp 10.217.0.7:8080: connect: connection refused" start-of-body= Oct 06 10:00:43 crc kubenswrapper[4824]: I1006 10:00:43.236879 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-s4ht9" podUID="171abefb-ea19-40fa-8a22-4b38956d5e4b" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.7:8080/\": dial tcp 10.217.0.7:8080: connect: connection refused" Oct 06 10:00:43 crc kubenswrapper[4824]: I1006 10:00:43.240707 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-nqmr6" event={"ID":"b6f7d3cb-418e-4f73-95a0-2477e37976c3","Type":"ContainerStarted","Data":"610b272f4cb7b19cc0d283e6e5d3a4489f50cd7ffafc76e70072bfa5b0da0396"} Oct 06 10:00:43 crc kubenswrapper[4824]: I1006 10:00:43.240753 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-nqmr6" event={"ID":"b6f7d3cb-418e-4f73-95a0-2477e37976c3","Type":"ContainerStarted","Data":"243447b3c3ad6710011f6daf375f053746d337edca1d1ec2b5c40e13fb029dc6"} Oct 06 10:00:43 crc kubenswrapper[4824]: I1006 10:00:43.247487 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-hqhlr" event={"ID":"4d7b1b9b-33de-4fca-b629-6790c3a5cb7d","Type":"ContainerStarted","Data":"36c21cf110c2800d9f5222157b518dc212a832cf1c2c8a640b194101acc8bf05"} Oct 06 10:00:43 crc kubenswrapper[4824]: I1006 10:00:43.250522 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w6ks8\" (UniqueName: \"kubernetes.io/projected/ffceea22-78f2-4032-a23a-71b4fb7ea83e-kube-api-access-w6ks8\") pod \"packageserver-d55dfcdfc-s8j95\" (UID: \"ffceea22-78f2-4032-a23a-71b4fb7ea83e\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-s8j95" Oct 06 10:00:43 crc kubenswrapper[4824]: I1006 10:00:43.253883 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-7qwsv" event={"ID":"d6e255a4-d3e7-4cef-bf40-f98aa11b3e3b","Type":"ContainerStarted","Data":"e86484bb75a4ab5a79c7515d7bff165602c1ec5c4bf494b11c28b11200785764"} Oct 06 10:00:43 crc kubenswrapper[4824]: I1006 10:00:43.253916 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-7qwsv" event={"ID":"d6e255a4-d3e7-4cef-bf40-f98aa11b3e3b","Type":"ContainerStarted","Data":"5a92e6892020d3dbf96d5c9c6fde1190e308094de484268daf0f82d6be092dbe"} Oct 06 10:00:43 crc kubenswrapper[4824]: I1006 10:00:43.295893 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cjzws\" (UniqueName: \"kubernetes.io/projected/26536a6b-3639-46ad-b024-5e21b7118fe7-kube-api-access-cjzws\") pod \"collect-profiles-29329080-9kpdv\" (UID: \"26536a6b-3639-46ad-b024-5e21b7118fe7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329080-9kpdv" Oct 06 10:00:43 crc kubenswrapper[4824]: I1006 10:00:43.299352 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xshbg\" (UniqueName: \"kubernetes.io/projected/92937768-de56-4cda-816c-9d1be4c07f0f-kube-api-access-xshbg\") pod \"migrator-59844c95c7-qpxrp\" (UID: \"92937768-de56-4cda-816c-9d1be4c07f0f\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-qpxrp" Oct 06 10:00:43 crc kubenswrapper[4824]: I1006 10:00:43.305525 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-vn26d" Oct 06 10:00:43 crc kubenswrapper[4824]: I1006 10:00:43.310943 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-xr4mg" Oct 06 10:00:43 crc kubenswrapper[4824]: I1006 10:00:43.312729 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 10:00:43 crc kubenswrapper[4824]: E1006 10:00:43.314361 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-06 10:00:43.814339446 +0000 UTC m=+153.178762297 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 10:00:43 crc kubenswrapper[4824]: I1006 10:00:43.319026 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-56948" Oct 06 10:00:43 crc kubenswrapper[4824]: I1006 10:00:43.321486 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rnnkq\" (UniqueName: \"kubernetes.io/projected/d2595750-aae8-4f17-8782-b7eccb0c2948-kube-api-access-rnnkq\") pod \"marketplace-operator-79b997595-qq4zz\" (UID: \"d2595750-aae8-4f17-8782-b7eccb0c2948\") " pod="openshift-marketplace/marketplace-operator-79b997595-qq4zz" Oct 06 10:00:43 crc kubenswrapper[4824]: I1006 10:00:43.328601 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tfjz5\" (UniqueName: \"kubernetes.io/projected/744def19-0006-4ab4-9d45-8a2a39bd410a-kube-api-access-tfjz5\") pod \"machine-config-server-fdbvz\" (UID: \"744def19-0006-4ab4-9d45-8a2a39bd410a\") " pod="openshift-machine-config-operator/machine-config-server-fdbvz" Oct 06 10:00:43 crc kubenswrapper[4824]: W1006 10:00:43.329547 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod911b7e6b_1df0_4ef3_822e_84925bd6bbdc.slice/crio-4739c300bb6be27278f16614bcf0708680eb1dc01bbc7ad41f79e0e2c2596607 WatchSource:0}: Error finding container 4739c300bb6be27278f16614bcf0708680eb1dc01bbc7ad41f79e0e2c2596607: Status 404 returned error can't find the container with id 4739c300bb6be27278f16614bcf0708680eb1dc01bbc7ad41f79e0e2c2596607 Oct 06 10:00:43 crc kubenswrapper[4824]: I1006 10:00:43.333059 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-fn7dj" Oct 06 10:00:43 crc kubenswrapper[4824]: I1006 10:00:43.346734 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zzrfg" Oct 06 10:00:43 crc kubenswrapper[4824]: I1006 10:00:43.352478 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6tt2t" Oct 06 10:00:43 crc kubenswrapper[4824]: I1006 10:00:43.362524 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-s8j95" Oct 06 10:00:43 crc kubenswrapper[4824]: I1006 10:00:43.369800 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dpf26\" (UniqueName: \"kubernetes.io/projected/d40e2672-6c82-45e6-a292-21615712032d-kube-api-access-dpf26\") pod \"kube-storage-version-migrator-operator-b67b599dd-wp6kv\" (UID: \"d40e2672-6c82-45e6-a292-21615712032d\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-wp6kv" Oct 06 10:00:43 crc kubenswrapper[4824]: I1006 10:00:43.376665 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-qpxrp" Oct 06 10:00:43 crc kubenswrapper[4824]: I1006 10:00:43.391637 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4rsfz\" (UniqueName: \"kubernetes.io/projected/af8fbf5a-f90e-4d22-ae65-05d8d8253308-kube-api-access-4rsfz\") pod \"oauth-openshift-558db77b4-8z68k\" (UID: \"af8fbf5a-f90e-4d22-ae65-05d8d8253308\") " pod="openshift-authentication/oauth-openshift-558db77b4-8z68k" Oct 06 10:00:43 crc kubenswrapper[4824]: I1006 10:00:43.392739 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jgp2m\" (UniqueName: \"kubernetes.io/projected/2a0286c5-9866-4a1c-9855-49cea5d04d69-kube-api-access-jgp2m\") pod \"machine-config-operator-74547568cd-vvtbh\" (UID: \"2a0286c5-9866-4a1c-9855-49cea5d04d69\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vvtbh" Oct 06 10:00:43 crc kubenswrapper[4824]: I1006 10:00:43.393871 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29329080-9kpdv" Oct 06 10:00:43 crc kubenswrapper[4824]: I1006 10:00:43.405161 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-qq4zz" Oct 06 10:00:43 crc kubenswrapper[4824]: I1006 10:00:43.415355 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8l8gd\" (UID: \"c2f6e84b-a28c-4ead-97f5-978b0639e5bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-8l8gd" Oct 06 10:00:43 crc kubenswrapper[4824]: E1006 10:00:43.417112 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-06 10:00:43.917088049 +0000 UTC m=+153.281510910 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8l8gd" (UID: "c2f6e84b-a28c-4ead-97f5-978b0639e5bf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 10:00:43 crc kubenswrapper[4824]: I1006 10:00:43.424029 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-5blgj" Oct 06 10:00:43 crc kubenswrapper[4824]: I1006 10:00:43.432053 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-fdbvz" Oct 06 10:00:43 crc kubenswrapper[4824]: I1006 10:00:43.518853 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 10:00:43 crc kubenswrapper[4824]: E1006 10:00:43.519617 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-06 10:00:44.019597024 +0000 UTC m=+153.384019885 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 10:00:43 crc kubenswrapper[4824]: I1006 10:00:43.582239 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-8z68k" Oct 06 10:00:43 crc kubenswrapper[4824]: I1006 10:00:43.596474 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vvtbh" Oct 06 10:00:43 crc kubenswrapper[4824]: I1006 10:00:43.620372 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8l8gd\" (UID: \"c2f6e84b-a28c-4ead-97f5-978b0639e5bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-8l8gd" Oct 06 10:00:43 crc kubenswrapper[4824]: E1006 10:00:43.620825 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-06 10:00:44.120808797 +0000 UTC m=+153.485231668 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8l8gd" (UID: "c2f6e84b-a28c-4ead-97f5-978b0639e5bf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 10:00:43 crc kubenswrapper[4824]: I1006 10:00:43.624409 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-wp6kv" Oct 06 10:00:43 crc kubenswrapper[4824]: W1006 10:00:43.678585 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod744def19_0006_4ab4_9d45_8a2a39bd410a.slice/crio-52e57d77b297ecf151b814f73c38c4e672196f7fa77bf331d5ea79f443000075 WatchSource:0}: Error finding container 52e57d77b297ecf151b814f73c38c4e672196f7fa77bf331d5ea79f443000075: Status 404 returned error can't find the container with id 52e57d77b297ecf151b814f73c38c4e672196f7fa77bf331d5ea79f443000075 Oct 06 10:00:43 crc kubenswrapper[4824]: I1006 10:00:43.721419 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 10:00:43 crc kubenswrapper[4824]: E1006 10:00:43.721722 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-06 10:00:44.221701573 +0000 UTC m=+153.586124424 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 10:00:43 crc kubenswrapper[4824]: I1006 10:00:43.775818 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-wk4bd"] Oct 06 10:00:43 crc kubenswrapper[4824]: I1006 10:00:43.786010 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-dklh2"] Oct 06 10:00:43 crc kubenswrapper[4824]: I1006 10:00:43.799672 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-mdv87"] Oct 06 10:00:43 crc kubenswrapper[4824]: I1006 10:00:43.821052 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-j2m8k"] Oct 06 10:00:43 crc kubenswrapper[4824]: I1006 10:00:43.822385 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-fjdwv"] Oct 06 10:00:43 crc kubenswrapper[4824]: I1006 10:00:43.823293 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8l8gd\" (UID: \"c2f6e84b-a28c-4ead-97f5-978b0639e5bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-8l8gd" Oct 06 10:00:43 crc kubenswrapper[4824]: E1006 10:00:43.823676 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-06 10:00:44.323661205 +0000 UTC m=+153.688084066 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8l8gd" (UID: "c2f6e84b-a28c-4ead-97f5-978b0639e5bf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 10:00:43 crc kubenswrapper[4824]: I1006 10:00:43.851134 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-7qwsv" Oct 06 10:00:43 crc kubenswrapper[4824]: I1006 10:00:43.864795 4824 patch_prober.go:28] interesting pod/router-default-5444994796-7qwsv container/router namespace/openshift-ingress: Startup probe status=failure output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" start-of-body= Oct 06 10:00:43 crc kubenswrapper[4824]: I1006 10:00:43.864890 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-7qwsv" podUID="d6e255a4-d3e7-4cef-bf40-f98aa11b3e3b" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" Oct 06 10:00:43 crc kubenswrapper[4824]: I1006 10:00:43.873503 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-2xznh"] Oct 06 10:00:43 crc kubenswrapper[4824]: I1006 10:00:43.914949 4824 patch_prober.go:28] interesting pod/machine-config-daemon-khgzw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 10:00:43 crc kubenswrapper[4824]: I1006 10:00:43.915025 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 10:00:43 crc kubenswrapper[4824]: I1006 10:00:43.930125 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 10:00:43 crc kubenswrapper[4824]: E1006 10:00:43.930440 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-06 10:00:44.430419925 +0000 UTC m=+153.794842786 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 10:00:44 crc kubenswrapper[4824]: I1006 10:00:44.032269 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8l8gd\" (UID: \"c2f6e84b-a28c-4ead-97f5-978b0639e5bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-8l8gd" Oct 06 10:00:44 crc kubenswrapper[4824]: E1006 10:00:44.032579 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-06 10:00:44.532567012 +0000 UTC m=+153.896989873 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8l8gd" (UID: "c2f6e84b-a28c-4ead-97f5-978b0639e5bf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 10:00:44 crc kubenswrapper[4824]: I1006 10:00:44.062401 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-7p7wb"] Oct 06 10:00:44 crc kubenswrapper[4824]: I1006 10:00:44.094142 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-bjzzk"] Oct 06 10:00:44 crc kubenswrapper[4824]: I1006 10:00:44.105311 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-6kn7b"] Oct 06 10:00:44 crc kubenswrapper[4824]: I1006 10:00:44.113112 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-prthm"] Oct 06 10:00:44 crc kubenswrapper[4824]: I1006 10:00:44.116547 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-fpxws"] Oct 06 10:00:44 crc kubenswrapper[4824]: I1006 10:00:44.133811 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 10:00:44 crc kubenswrapper[4824]: E1006 10:00:44.134026 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-06 10:00:44.634003992 +0000 UTC m=+153.998426843 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 10:00:44 crc kubenswrapper[4824]: I1006 10:00:44.134115 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8l8gd\" (UID: \"c2f6e84b-a28c-4ead-97f5-978b0639e5bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-8l8gd" Oct 06 10:00:44 crc kubenswrapper[4824]: E1006 10:00:44.134650 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-06 10:00:44.634641887 +0000 UTC m=+153.999064748 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8l8gd" (UID: "c2f6e84b-a28c-4ead-97f5-978b0639e5bf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 10:00:44 crc kubenswrapper[4824]: I1006 10:00:44.236388 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 10:00:44 crc kubenswrapper[4824]: E1006 10:00:44.238402 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-06 10:00:44.738380383 +0000 UTC m=+154.102803244 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 10:00:44 crc kubenswrapper[4824]: I1006 10:00:44.265337 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-fjdwv" event={"ID":"b349b0af-0ba8-409f-a915-7fb06b2e3774","Type":"ContainerStarted","Data":"b96e97a4762878e697d4817c01c3463788d0de20d3a1b8c8f431ec281c711884"} Oct 06 10:00:44 crc kubenswrapper[4824]: I1006 10:00:44.266817 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-dklh2" event={"ID":"d21f4737-c582-41e9-846c-dd75ec173afd","Type":"ContainerStarted","Data":"7db74fcdfad3c3e1099199ef6946ea78be133f41487a598d367fef9fd947694f"} Oct 06 10:00:44 crc kubenswrapper[4824]: I1006 10:00:44.267546 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-fdbvz" event={"ID":"744def19-0006-4ab4-9d45-8a2a39bd410a","Type":"ContainerStarted","Data":"52e57d77b297ecf151b814f73c38c4e672196f7fa77bf331d5ea79f443000075"} Oct 06 10:00:44 crc kubenswrapper[4824]: I1006 10:00:44.287836 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-hqhlr" event={"ID":"4d7b1b9b-33de-4fca-b629-6790c3a5cb7d","Type":"ContainerStarted","Data":"399985f310dc4bac06bd49e0b080fa9b1881c633a104bd2d6c48ed8796519234"} Oct 06 10:00:44 crc kubenswrapper[4824]: I1006 10:00:44.287902 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-hqhlr" event={"ID":"4d7b1b9b-33de-4fca-b629-6790c3a5cb7d","Type":"ContainerStarted","Data":"1f9c73a4595c75a16f9b4e8354d0184e38f288f0ca89bd2df86038a6ab77f013"} Oct 06 10:00:44 crc kubenswrapper[4824]: I1006 10:00:44.293231 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-sb9c8" event={"ID":"c8fc8c20-e06a-4039-80c0-82f84ea81391","Type":"ContainerStarted","Data":"c4d24a243d07fb725fd3d1769da8950a2aef4aad98a726e58e1c87fa31f606da"} Oct 06 10:00:44 crc kubenswrapper[4824]: I1006 10:00:44.294858 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wk4bd" event={"ID":"dd50f0ac-ac97-413f-a350-7496681f7f0a","Type":"ContainerStarted","Data":"caa6556ff759f287324f7bb3ad5df851767a64c5aebfb75b072ba6fdd55b4ef8"} Oct 06 10:00:44 crc kubenswrapper[4824]: I1006 10:00:44.296187 4824 generic.go:334] "Generic (PLEG): container finished" podID="7c497a24-d8cb-4560-993f-fc73b3af8c04" containerID="831f8b2e901fa23b74f1e0cd4b24b38ad82576217e8fb877bbc634881ee019a8" exitCode=0 Oct 06 10:00:44 crc kubenswrapper[4824]: I1006 10:00:44.296238 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-v6gnw" event={"ID":"7c497a24-d8cb-4560-993f-fc73b3af8c04","Type":"ContainerDied","Data":"831f8b2e901fa23b74f1e0cd4b24b38ad82576217e8fb877bbc634881ee019a8"} Oct 06 10:00:44 crc kubenswrapper[4824]: I1006 10:00:44.296257 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-v6gnw" event={"ID":"7c497a24-d8cb-4560-993f-fc73b3af8c04","Type":"ContainerStarted","Data":"2fe79614745ef2d6d39db31529f8346fe8af9391bfbfe0b7231ba09a9f700f8a"} Oct 06 10:00:44 crc kubenswrapper[4824]: I1006 10:00:44.299690 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-mdv87" event={"ID":"4bceefc5-1905-48bd-b214-ab5824f0a302","Type":"ContainerStarted","Data":"d6bbffa84b956a4fef934b586e5cb7e375de29efe0e2ade071e1ca774b458d2b"} Oct 06 10:00:44 crc kubenswrapper[4824]: I1006 10:00:44.304185 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-j2m8k" event={"ID":"323631a8-107c-4fd0-8b57-04550f0fc7d1","Type":"ContainerStarted","Data":"48f490bcbb8a5d29a4ca48b0543cdf88f078ea24bb99b08e151ad548985bc5f2"} Oct 06 10:00:44 crc kubenswrapper[4824]: I1006 10:00:44.326330 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-4hw5p" event={"ID":"9da4a453-8e05-4360-b74c-ad1c4706b005","Type":"ContainerStarted","Data":"0fff266b322ef3ccdf18a2b8b6e7dfd4d7bcb84521aa5dd980f1a893544400fb"} Oct 06 10:00:44 crc kubenswrapper[4824]: I1006 10:00:44.335792 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-2xznh" event={"ID":"d0546787-f9de-4b06-ab4f-0e1632d36230","Type":"ContainerStarted","Data":"5bae69febef831f48ade3723fcc96d12230dd4aa9575bb54fbe58e0c8ff2daab"} Oct 06 10:00:44 crc kubenswrapper[4824]: I1006 10:00:44.339186 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8l8gd\" (UID: \"c2f6e84b-a28c-4ead-97f5-978b0639e5bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-8l8gd" Oct 06 10:00:44 crc kubenswrapper[4824]: E1006 10:00:44.339601 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-06 10:00:44.839582107 +0000 UTC m=+154.204004958 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8l8gd" (UID: "c2f6e84b-a28c-4ead-97f5-978b0639e5bf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 10:00:44 crc kubenswrapper[4824]: I1006 10:00:44.350171 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-q76nt" event={"ID":"911b7e6b-1df0-4ef3-822e-84925bd6bbdc","Type":"ContainerStarted","Data":"4739c300bb6be27278f16614bcf0708680eb1dc01bbc7ad41f79e0e2c2596607"} Oct 06 10:00:44 crc kubenswrapper[4824]: I1006 10:00:44.350241 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-nqmr6" Oct 06 10:00:44 crc kubenswrapper[4824]: I1006 10:00:44.351501 4824 patch_prober.go:28] interesting pod/downloads-7954f5f757-s4ht9 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.7:8080/\": dial tcp 10.217.0.7:8080: connect: connection refused" start-of-body= Oct 06 10:00:44 crc kubenswrapper[4824]: I1006 10:00:44.351582 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-s4ht9" podUID="171abefb-ea19-40fa-8a22-4b38956d5e4b" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.7:8080/\": dial tcp 10.217.0.7:8080: connect: connection refused" Oct 06 10:00:44 crc kubenswrapper[4824]: I1006 10:00:44.361911 4824 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-nqmr6 container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.17:8443/healthz\": dial tcp 10.217.0.17:8443: connect: connection refused" start-of-body= Oct 06 10:00:44 crc kubenswrapper[4824]: I1006 10:00:44.362025 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-nqmr6" podUID="b6f7d3cb-418e-4f73-95a0-2477e37976c3" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.17:8443/healthz\": dial tcp 10.217.0.17:8443: connect: connection refused" Oct 06 10:00:44 crc kubenswrapper[4824]: I1006 10:00:44.440241 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 10:00:44 crc kubenswrapper[4824]: E1006 10:00:44.443286 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-06 10:00:44.943260121 +0000 UTC m=+154.307682982 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 10:00:44 crc kubenswrapper[4824]: I1006 10:00:44.544903 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8l8gd\" (UID: \"c2f6e84b-a28c-4ead-97f5-978b0639e5bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-8l8gd" Oct 06 10:00:44 crc kubenswrapper[4824]: E1006 10:00:44.545556 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-06 10:00:45.045519821 +0000 UTC m=+154.409942682 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8l8gd" (UID: "c2f6e84b-a28c-4ead-97f5-978b0639e5bf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 10:00:44 crc kubenswrapper[4824]: I1006 10:00:44.635132 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-dr4mr"] Oct 06 10:00:44 crc kubenswrapper[4824]: I1006 10:00:44.649343 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 10:00:44 crc kubenswrapper[4824]: E1006 10:00:44.649594 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-06 10:00:45.149556593 +0000 UTC m=+154.513979454 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 10:00:44 crc kubenswrapper[4824]: I1006 10:00:44.649667 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8l8gd\" (UID: \"c2f6e84b-a28c-4ead-97f5-978b0639e5bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-8l8gd" Oct 06 10:00:44 crc kubenswrapper[4824]: E1006 10:00:44.650298 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-06 10:00:45.150279121 +0000 UTC m=+154.514701972 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8l8gd" (UID: "c2f6e84b-a28c-4ead-97f5-978b0639e5bf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 10:00:44 crc kubenswrapper[4824]: I1006 10:00:44.653877 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-nk67f"] Oct 06 10:00:44 crc kubenswrapper[4824]: I1006 10:00:44.660554 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-px5fp"] Oct 06 10:00:44 crc kubenswrapper[4824]: I1006 10:00:44.680362 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-xr4mg"] Oct 06 10:00:44 crc kubenswrapper[4824]: I1006 10:00:44.750576 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 10:00:44 crc kubenswrapper[4824]: E1006 10:00:44.751238 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-06 10:00:45.251218629 +0000 UTC m=+154.615641490 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 10:00:44 crc kubenswrapper[4824]: I1006 10:00:44.771675 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-sb9c8" podStartSLOduration=132.771653481 podStartE2EDuration="2m12.771653481s" podCreationTimestamp="2025-10-06 09:58:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:00:44.768887963 +0000 UTC m=+154.133310824" watchObservedRunningTime="2025-10-06 10:00:44.771653481 +0000 UTC m=+154.136076342" Oct 06 10:00:44 crc kubenswrapper[4824]: I1006 10:00:44.814057 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-56948"] Oct 06 10:00:44 crc kubenswrapper[4824]: I1006 10:00:44.818258 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-x4fdc"] Oct 06 10:00:44 crc kubenswrapper[4824]: I1006 10:00:44.826243 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-qq4zz"] Oct 06 10:00:44 crc kubenswrapper[4824]: I1006 10:00:44.835020 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-vpg8w"] Oct 06 10:00:44 crc kubenswrapper[4824]: I1006 10:00:44.839334 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-fn7dj"] Oct 06 10:00:44 crc kubenswrapper[4824]: I1006 10:00:44.852245 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8l8gd\" (UID: \"c2f6e84b-a28c-4ead-97f5-978b0639e5bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-8l8gd" Oct 06 10:00:44 crc kubenswrapper[4824]: E1006 10:00:44.852736 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-06 10:00:45.352700628 +0000 UTC m=+154.717123489 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8l8gd" (UID: "c2f6e84b-a28c-4ead-97f5-978b0639e5bf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 10:00:44 crc kubenswrapper[4824]: I1006 10:00:44.858923 4824 patch_prober.go:28] interesting pod/router-default-5444994796-7qwsv container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 06 10:00:44 crc kubenswrapper[4824]: [-]has-synced failed: reason withheld Oct 06 10:00:44 crc kubenswrapper[4824]: [+]process-running ok Oct 06 10:00:44 crc kubenswrapper[4824]: healthz check failed Oct 06 10:00:44 crc kubenswrapper[4824]: I1006 10:00:44.858995 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-7qwsv" podUID="d6e255a4-d3e7-4cef-bf40-f98aa11b3e3b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 06 10:00:44 crc kubenswrapper[4824]: I1006 10:00:44.892176 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-8rxsm"] Oct 06 10:00:44 crc kubenswrapper[4824]: I1006 10:00:44.892230 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-vn26d"] Oct 06 10:00:44 crc kubenswrapper[4824]: I1006 10:00:44.912861 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wlbrp"] Oct 06 10:00:44 crc kubenswrapper[4824]: W1006 10:00:44.930847 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd2595750_aae8_4f17_8782_b7eccb0c2948.slice/crio-fde24d5c6f8274bd8d527ae7747c3941c580a57dc5640c813a74782ac1f9d4b5 WatchSource:0}: Error finding container fde24d5c6f8274bd8d527ae7747c3941c580a57dc5640c813a74782ac1f9d4b5: Status 404 returned error can't find the container with id fde24d5c6f8274bd8d527ae7747c3941c580a57dc5640c813a74782ac1f9d4b5 Oct 06 10:00:44 crc kubenswrapper[4824]: I1006 10:00:44.932904 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-8z68k"] Oct 06 10:00:44 crc kubenswrapper[4824]: I1006 10:00:44.940351 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-5blgj"] Oct 06 10:00:44 crc kubenswrapper[4824]: I1006 10:00:44.942485 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-wp6kv"] Oct 06 10:00:44 crc kubenswrapper[4824]: I1006 10:00:44.946252 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-qpxrp"] Oct 06 10:00:44 crc kubenswrapper[4824]: I1006 10:00:44.949339 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29329080-9kpdv"] Oct 06 10:00:44 crc kubenswrapper[4824]: I1006 10:00:44.953168 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 10:00:44 crc kubenswrapper[4824]: E1006 10:00:44.953425 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-06 10:00:45.453390249 +0000 UTC m=+154.817813110 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 10:00:44 crc kubenswrapper[4824]: I1006 10:00:44.953469 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8l8gd\" (UID: \"c2f6e84b-a28c-4ead-97f5-978b0639e5bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-8l8gd" Oct 06 10:00:44 crc kubenswrapper[4824]: E1006 10:00:44.954219 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-06 10:00:45.454183729 +0000 UTC m=+154.818606590 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8l8gd" (UID: "c2f6e84b-a28c-4ead-97f5-978b0639e5bf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 10:00:44 crc kubenswrapper[4824]: I1006 10:00:44.954513 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-s8j95"] Oct 06 10:00:44 crc kubenswrapper[4824]: W1006 10:00:44.974034 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2a76ccaa_9249_46ba_bbc5_dbbbc33b06bb.slice/crio-1e0e447f77f99bf899d42b429c5d3857aa4989557a1099fa4db6442b35ef6f7e WatchSource:0}: Error finding container 1e0e447f77f99bf899d42b429c5d3857aa4989557a1099fa4db6442b35ef6f7e: Status 404 returned error can't find the container with id 1e0e447f77f99bf899d42b429c5d3857aa4989557a1099fa4db6442b35ef6f7e Oct 06 10:00:44 crc kubenswrapper[4824]: I1006 10:00:44.988087 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-vvtbh"] Oct 06 10:00:45 crc kubenswrapper[4824]: I1006 10:00:45.053595 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-s4ht9" podStartSLOduration=134.053573677 podStartE2EDuration="2m14.053573677s" podCreationTimestamp="2025-10-06 09:58:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:00:45.052734247 +0000 UTC m=+154.417157108" watchObservedRunningTime="2025-10-06 10:00:45.053573677 +0000 UTC m=+154.417996538" Oct 06 10:00:45 crc kubenswrapper[4824]: E1006 10:00:45.058567 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-06 10:00:45.558532109 +0000 UTC m=+154.922954970 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 10:00:45 crc kubenswrapper[4824]: I1006 10:00:45.066022 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 10:00:45 crc kubenswrapper[4824]: I1006 10:00:45.071251 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8l8gd\" (UID: \"c2f6e84b-a28c-4ead-97f5-978b0639e5bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-8l8gd" Oct 06 10:00:45 crc kubenswrapper[4824]: I1006 10:00:45.071925 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-zzrfg"] Oct 06 10:00:45 crc kubenswrapper[4824]: E1006 10:00:45.073647 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-06 10:00:45.573358685 +0000 UTC m=+154.937781546 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8l8gd" (UID: "c2f6e84b-a28c-4ead-97f5-978b0639e5bf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 10:00:45 crc kubenswrapper[4824]: I1006 10:00:45.099149 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6tt2t"] Oct 06 10:00:45 crc kubenswrapper[4824]: I1006 10:00:45.176689 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 10:00:45 crc kubenswrapper[4824]: E1006 10:00:45.177181 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-06 10:00:45.677159332 +0000 UTC m=+155.041582193 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 10:00:45 crc kubenswrapper[4824]: I1006 10:00:45.208078 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-nqmr6" podStartSLOduration=134.208058624 podStartE2EDuration="2m14.208058624s" podCreationTimestamp="2025-10-06 09:58:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:00:45.199838021 +0000 UTC m=+154.564260882" watchObservedRunningTime="2025-10-06 10:00:45.208058624 +0000 UTC m=+154.572481475" Oct 06 10:00:45 crc kubenswrapper[4824]: I1006 10:00:45.278408 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8l8gd\" (UID: \"c2f6e84b-a28c-4ead-97f5-978b0639e5bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-8l8gd" Oct 06 10:00:45 crc kubenswrapper[4824]: E1006 10:00:45.278897 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-06 10:00:45.778881488 +0000 UTC m=+155.143304349 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8l8gd" (UID: "c2f6e84b-a28c-4ead-97f5-978b0639e5bf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 10:00:45 crc kubenswrapper[4824]: I1006 10:00:45.318208 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-7qwsv" podStartSLOduration=133.318182696 podStartE2EDuration="2m13.318182696s" podCreationTimestamp="2025-10-06 09:58:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:00:45.316907436 +0000 UTC m=+154.681330297" watchObservedRunningTime="2025-10-06 10:00:45.318182696 +0000 UTC m=+154.682605557" Oct 06 10:00:45 crc kubenswrapper[4824]: I1006 10:00:45.362262 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-fpxws" event={"ID":"722b7a46-52a1-49a7-a053-803334dc5964","Type":"ContainerStarted","Data":"5a683afc848a67c0e8bf3bf69fd9a0c4ede0a48fc5f7c8ef2264f4232013651b"} Oct 06 10:00:45 crc kubenswrapper[4824]: I1006 10:00:45.362319 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-fpxws" event={"ID":"722b7a46-52a1-49a7-a053-803334dc5964","Type":"ContainerStarted","Data":"e7f551e87c80776ae6e7ead338ebe026b4ce3179396bfa6a61c62c451c93592c"} Oct 06 10:00:45 crc kubenswrapper[4824]: I1006 10:00:45.380472 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 10:00:45 crc kubenswrapper[4824]: E1006 10:00:45.383682 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-06 10:00:45.883649889 +0000 UTC m=+155.248072750 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 10:00:45 crc kubenswrapper[4824]: I1006 10:00:45.401286 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-wp6kv" event={"ID":"d40e2672-6c82-45e6-a292-21615712032d","Type":"ContainerStarted","Data":"ffe8e8ff93a66af2cc09c627ba49443e9de28ff797e9d2dd4a5a536003b3b5e6"} Oct 06 10:00:45 crc kubenswrapper[4824]: I1006 10:00:45.402390 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-hqhlr" podStartSLOduration=134.402365261 podStartE2EDuration="2m14.402365261s" podCreationTimestamp="2025-10-06 09:58:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:00:45.40193602 +0000 UTC m=+154.766358881" watchObservedRunningTime="2025-10-06 10:00:45.402365261 +0000 UTC m=+154.766788122" Oct 06 10:00:45 crc kubenswrapper[4824]: I1006 10:00:45.415857 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-dklh2" event={"ID":"d21f4737-c582-41e9-846c-dd75ec173afd","Type":"ContainerStarted","Data":"2a53943a32801712a14a229b6eb092906d7d1a9055d200c9613092db734eb3fa"} Oct 06 10:00:45 crc kubenswrapper[4824]: I1006 10:00:45.417847 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-8rxsm" event={"ID":"ff05ab6e-248f-445c-9856-c1a2868f8840","Type":"ContainerStarted","Data":"219d9bd34b6171f81cf635ac1db2cd754b7ea966d45d464888788a65383c05de"} Oct 06 10:00:45 crc kubenswrapper[4824]: I1006 10:00:45.420036 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-4hw5p" event={"ID":"9da4a453-8e05-4360-b74c-ad1c4706b005","Type":"ContainerStarted","Data":"7e229dca2832f511d0074fb36e08faa9df1af837b54b2e3e19428ad9a458922a"} Oct 06 10:00:45 crc kubenswrapper[4824]: I1006 10:00:45.421352 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-8z68k" event={"ID":"af8fbf5a-f90e-4d22-ae65-05d8d8253308","Type":"ContainerStarted","Data":"039afdf33c495c5167b8f389b7d84c3a275a2fd6aa47bc9c99e0195359cafe51"} Oct 06 10:00:45 crc kubenswrapper[4824]: I1006 10:00:45.424759 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-prthm" event={"ID":"45682655-0e98-4a5e-b5e8-308f4d0f7d9a","Type":"ContainerStarted","Data":"bf863d0c2626fcadbbc32b35885d45e378f8a153a906dfdd745c7f10ef2b3898"} Oct 06 10:00:45 crc kubenswrapper[4824]: I1006 10:00:45.424787 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-prthm" event={"ID":"45682655-0e98-4a5e-b5e8-308f4d0f7d9a","Type":"ContainerStarted","Data":"91bc3a43eb595e3266b1e6c1f7fa1b7d2dee1fc877481396ffb6464e765e9d98"} Oct 06 10:00:45 crc kubenswrapper[4824]: I1006 10:00:45.426620 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wk4bd" event={"ID":"dd50f0ac-ac97-413f-a350-7496681f7f0a","Type":"ContainerStarted","Data":"37781d4ace294c944018c677becc05dc9d3fcc964a5419c612190d8c821c4cb7"} Oct 06 10:00:45 crc kubenswrapper[4824]: I1006 10:00:45.427106 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wk4bd" Oct 06 10:00:45 crc kubenswrapper[4824]: I1006 10:00:45.428462 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-56948" event={"ID":"678eeee3-60d8-4dbe-b815-10de5a46c4d1","Type":"ContainerStarted","Data":"a68b33ad70bbe16e1a9687ecc35eff9e7384633eef1598d61036a01fc4d0f022"} Oct 06 10:00:45 crc kubenswrapper[4824]: I1006 10:00:45.429256 4824 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-wk4bd container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.11:8443/healthz\": dial tcp 10.217.0.11:8443: connect: connection refused" start-of-body= Oct 06 10:00:45 crc kubenswrapper[4824]: I1006 10:00:45.429304 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wk4bd" podUID="dd50f0ac-ac97-413f-a350-7496681f7f0a" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.11:8443/healthz\": dial tcp 10.217.0.11:8443: connect: connection refused" Oct 06 10:00:45 crc kubenswrapper[4824]: I1006 10:00:45.444548 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-fpxws" podStartSLOduration=134.44452859 podStartE2EDuration="2m14.44452859s" podCreationTimestamp="2025-10-06 09:58:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:00:45.444423937 +0000 UTC m=+154.808846798" watchObservedRunningTime="2025-10-06 10:00:45.44452859 +0000 UTC m=+154.808951451" Oct 06 10:00:45 crc kubenswrapper[4824]: I1006 10:00:45.453568 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-xr4mg" event={"ID":"e81889bb-3225-458e-bce5-cfbafe6f4f92","Type":"ContainerStarted","Data":"f78c6fed745d66d8f3302b292cbb1809b1e345dd3f0917fdd4a8180422dcb804"} Oct 06 10:00:45 crc kubenswrapper[4824]: I1006 10:00:45.453820 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-xr4mg" event={"ID":"e81889bb-3225-458e-bce5-cfbafe6f4f92","Type":"ContainerStarted","Data":"19ced0e2f0f21164854e679199eb40f618f1e8a96817dfbcd1bc59b2aea3bf68"} Oct 06 10:00:45 crc kubenswrapper[4824]: I1006 10:00:45.458492 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-fjdwv" event={"ID":"b349b0af-0ba8-409f-a915-7fb06b2e3774","Type":"ContainerStarted","Data":"613408a87ce5ce811c0322fac132b018310cae991287de39d9ac175e7fdf3ea9"} Oct 06 10:00:45 crc kubenswrapper[4824]: I1006 10:00:45.460019 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zzrfg" event={"ID":"2fa5ee3e-29f4-43b2-a154-3c988754897a","Type":"ContainerStarted","Data":"c6a71faa7fe842ad7ed4269a8dd75b7654d0d246779427c75564d9046b030832"} Oct 06 10:00:45 crc kubenswrapper[4824]: I1006 10:00:45.462557 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-mdv87" event={"ID":"4bceefc5-1905-48bd-b214-ab5824f0a302","Type":"ContainerStarted","Data":"a23a268b50c932f2e8edf9cbf2d29287e37dd6c93ac2bc737281f6a8f8ea1dc8"} Oct 06 10:00:45 crc kubenswrapper[4824]: I1006 10:00:45.478066 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-j2m8k" event={"ID":"323631a8-107c-4fd0-8b57-04550f0fc7d1","Type":"ContainerStarted","Data":"3076bd0cdb53eb26dc5b5050647bb0a5074d26d0b0a3b1ea5a82558398086ed2"} Oct 06 10:00:45 crc kubenswrapper[4824]: I1006 10:00:45.486213 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8l8gd\" (UID: \"c2f6e84b-a28c-4ead-97f5-978b0639e5bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-8l8gd" Oct 06 10:00:45 crc kubenswrapper[4824]: I1006 10:00:45.486636 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-vn26d" event={"ID":"8e853926-f1e6-4571-863b-97918e660677","Type":"ContainerStarted","Data":"7398ec0b13caf60001a7f8f560ae6c47ff4b7b891f2e370e52106059c1c1bc5a"} Oct 06 10:00:45 crc kubenswrapper[4824]: E1006 10:00:45.488382 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-06 10:00:45.988362809 +0000 UTC m=+155.352785670 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8l8gd" (UID: "c2f6e84b-a28c-4ead-97f5-978b0639e5bf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 10:00:45 crc kubenswrapper[4824]: I1006 10:00:45.507925 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-2xznh" event={"ID":"d0546787-f9de-4b06-ab4f-0e1632d36230","Type":"ContainerStarted","Data":"e0f37aa5f03aa8d5d0948d80bcb68b9eda247316f898f89fdc6e953ff8684584"} Oct 06 10:00:45 crc kubenswrapper[4824]: I1006 10:00:45.508583 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-2xznh" Oct 06 10:00:45 crc kubenswrapper[4824]: I1006 10:00:45.511334 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vvtbh" event={"ID":"2a0286c5-9866-4a1c-9855-49cea5d04d69","Type":"ContainerStarted","Data":"92318a8264a18314fbcd3f1d03f0710e7613826634c76eda933c2c7315e7b600"} Oct 06 10:00:45 crc kubenswrapper[4824]: I1006 10:00:45.514366 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-bjzzk" event={"ID":"22545b78-29bb-45aa-a98c-fd7ce9703239","Type":"ContainerStarted","Data":"d9b85fac3db7eef560d91a27356cc20f8982f931ca7c4e0c48b491b0a91aeb55"} Oct 06 10:00:45 crc kubenswrapper[4824]: I1006 10:00:45.514394 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-bjzzk" event={"ID":"22545b78-29bb-45aa-a98c-fd7ce9703239","Type":"ContainerStarted","Data":"a8ec3fbc953d4ab3729185117219fe65ad55d0b3ce68d786ea383c60995ce2d6"} Oct 06 10:00:45 crc kubenswrapper[4824]: I1006 10:00:45.515167 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-bjzzk" Oct 06 10:00:45 crc kubenswrapper[4824]: I1006 10:00:45.515283 4824 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-2xznh container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.36:8443/healthz\": dial tcp 10.217.0.36:8443: connect: connection refused" start-of-body= Oct 06 10:00:45 crc kubenswrapper[4824]: I1006 10:00:45.515346 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-2xznh" podUID="d0546787-f9de-4b06-ab4f-0e1632d36230" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.36:8443/healthz\": dial tcp 10.217.0.36:8443: connect: connection refused" Oct 06 10:00:45 crc kubenswrapper[4824]: I1006 10:00:45.516651 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6tt2t" event={"ID":"982129a4-11b4-4024-afb0-ac6024165c7a","Type":"ContainerStarted","Data":"5e31620b33fecbcefb4aa98ea57f4b615c9a4ac0ec775fc0800d9081ed63a2ad"} Oct 06 10:00:45 crc kubenswrapper[4824]: I1006 10:00:45.518553 4824 patch_prober.go:28] interesting pod/console-operator-58897d9998-bjzzk container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.5:8443/readyz\": dial tcp 10.217.0.5:8443: connect: connection refused" start-of-body= Oct 06 10:00:45 crc kubenswrapper[4824]: I1006 10:00:45.518632 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-bjzzk" podUID="22545b78-29bb-45aa-a98c-fd7ce9703239" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.5:8443/readyz\": dial tcp 10.217.0.5:8443: connect: connection refused" Oct 06 10:00:45 crc kubenswrapper[4824]: I1006 10:00:45.543139 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wk4bd" podStartSLOduration=133.543117038 podStartE2EDuration="2m13.543117038s" podCreationTimestamp="2025-10-06 09:58:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:00:45.488040062 +0000 UTC m=+154.852462923" watchObservedRunningTime="2025-10-06 10:00:45.543117038 +0000 UTC m=+154.907539899" Oct 06 10:00:45 crc kubenswrapper[4824]: I1006 10:00:45.544672 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-dklh2" podStartSLOduration=133.544666036 podStartE2EDuration="2m13.544666036s" podCreationTimestamp="2025-10-06 09:58:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:00:45.542622037 +0000 UTC m=+154.907044898" watchObservedRunningTime="2025-10-06 10:00:45.544666036 +0000 UTC m=+154.909088898" Oct 06 10:00:45 crc kubenswrapper[4824]: I1006 10:00:45.552002 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-nk67f" event={"ID":"a1d6c12e-ff5a-4687-80ca-5699dd1bb9fe","Type":"ContainerStarted","Data":"56fdbc287e5f12d0ae7d3865ddbc862ecd95386119628bcd66576d19cb923446"} Oct 06 10:00:45 crc kubenswrapper[4824]: I1006 10:00:45.553023 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wlbrp" event={"ID":"9b1f3faf-72a1-44e2-814b-964902e78d3e","Type":"ContainerStarted","Data":"721ad0b1763f314c8ae46b7bab9400ac1d239b021513dc14b00e95bb2b035387"} Oct 06 10:00:45 crc kubenswrapper[4824]: I1006 10:00:45.555441 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-dr4mr" event={"ID":"c682e90b-3a5a-4b87-996b-8365b0936bda","Type":"ContainerStarted","Data":"42be8956149dd08b611cb5426099fce3b6d959441d021b633b0f03c707f81b51"} Oct 06 10:00:45 crc kubenswrapper[4824]: I1006 10:00:45.555461 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-dr4mr" event={"ID":"c682e90b-3a5a-4b87-996b-8365b0936bda","Type":"ContainerStarted","Data":"4421b9897e96faa7299d25b02888e58c103be0a471c57cb5f40bb03a542ba84d"} Oct 06 10:00:45 crc kubenswrapper[4824]: I1006 10:00:45.560279 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-5blgj" event={"ID":"46e11b34-d0dd-4ea6-bed7-66eb707c7f92","Type":"ContainerStarted","Data":"33cd86c61afe8b361bc3c591a815ba856cf2a0b2e34f5ae2e577405901034ac2"} Oct 06 10:00:45 crc kubenswrapper[4824]: I1006 10:00:45.564573 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-x4fdc" event={"ID":"a47e3600-79be-46c3-a1b8-6678d8da80c5","Type":"ContainerStarted","Data":"eba636315ade19faa9d3e1bc344fdc79f766120d616af298c85da8eca4b5eb8e"} Oct 06 10:00:45 crc kubenswrapper[4824]: I1006 10:00:45.577912 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-prthm" podStartSLOduration=133.577881525 podStartE2EDuration="2m13.577881525s" podCreationTimestamp="2025-10-06 09:58:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:00:45.573441916 +0000 UTC m=+154.937864777" watchObservedRunningTime="2025-10-06 10:00:45.577881525 +0000 UTC m=+154.942304386" Oct 06 10:00:45 crc kubenswrapper[4824]: I1006 10:00:45.587013 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 10:00:45 crc kubenswrapper[4824]: E1006 10:00:45.589170 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-06 10:00:46.089141492 +0000 UTC m=+155.453564353 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 10:00:45 crc kubenswrapper[4824]: I1006 10:00:45.592200 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-fn7dj" event={"ID":"2a76ccaa-9249-46ba-bbc5-dbbbc33b06bb","Type":"ContainerStarted","Data":"1e0e447f77f99bf899d42b429c5d3857aa4989557a1099fa4db6442b35ef6f7e"} Oct 06 10:00:45 crc kubenswrapper[4824]: I1006 10:00:45.620736 4824 generic.go:334] "Generic (PLEG): container finished" podID="911b7e6b-1df0-4ef3-822e-84925bd6bbdc" containerID="39a1d063e31923adb572e8ef0cacbbc9e36acc4ff8fffe9e17b8305afaad27b1" exitCode=0 Oct 06 10:00:45 crc kubenswrapper[4824]: I1006 10:00:45.620885 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-q76nt" event={"ID":"911b7e6b-1df0-4ef3-822e-84925bd6bbdc","Type":"ContainerDied","Data":"39a1d063e31923adb572e8ef0cacbbc9e36acc4ff8fffe9e17b8305afaad27b1"} Oct 06 10:00:45 crc kubenswrapper[4824]: I1006 10:00:45.628560 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-v6gnw" event={"ID":"7c497a24-d8cb-4560-993f-fc73b3af8c04","Type":"ContainerStarted","Data":"d8280a66f7c6a89b80956d254101fcac9b7d999d165f47429be884a1b4363f0f"} Oct 06 10:00:45 crc kubenswrapper[4824]: I1006 10:00:45.628652 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-v6gnw" Oct 06 10:00:45 crc kubenswrapper[4824]: I1006 10:00:45.633836 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29329080-9kpdv" event={"ID":"26536a6b-3639-46ad-b024-5e21b7118fe7","Type":"ContainerStarted","Data":"b38020ad9a04fa2d400919ee281ddf257bfb04804b7a93f037bded057e20f8aa"} Oct 06 10:00:45 crc kubenswrapper[4824]: I1006 10:00:45.665445 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-6kn7b" event={"ID":"7532dafc-fb61-4066-bb24-2470e4b2cfe6","Type":"ContainerStarted","Data":"5659c02ba34e4b8750c40c9be4837a6d656c51a4b28a178884bcb14e57ee6426"} Oct 06 10:00:45 crc kubenswrapper[4824]: I1006 10:00:45.666002 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-6kn7b" event={"ID":"7532dafc-fb61-4066-bb24-2470e4b2cfe6","Type":"ContainerStarted","Data":"4ccad7884ffe2096d126337d61a57ba086826998470a757a7795457f38aed232"} Oct 06 10:00:45 crc kubenswrapper[4824]: I1006 10:00:45.668918 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-2xznh" podStartSLOduration=133.668894327 podStartE2EDuration="2m13.668894327s" podCreationTimestamp="2025-10-06 09:58:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:00:45.65439311 +0000 UTC m=+155.018815972" watchObservedRunningTime="2025-10-06 10:00:45.668894327 +0000 UTC m=+155.033317188" Oct 06 10:00:45 crc kubenswrapper[4824]: I1006 10:00:45.671738 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-mdv87" podStartSLOduration=133.671730987 podStartE2EDuration="2m13.671730987s" podCreationTimestamp="2025-10-06 09:58:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:00:45.60852256 +0000 UTC m=+154.972945421" watchObservedRunningTime="2025-10-06 10:00:45.671730987 +0000 UTC m=+155.036153848" Oct 06 10:00:45 crc kubenswrapper[4824]: I1006 10:00:45.678577 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-px5fp" event={"ID":"9613dce2-6a39-4480-8b7f-553288ebc45b","Type":"ContainerStarted","Data":"433949ba0fd5cdb600fce45c3f729cff1ed004bc9fd43c84b3b3b692f17d0c21"} Oct 06 10:00:45 crc kubenswrapper[4824]: I1006 10:00:45.678648 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-px5fp" event={"ID":"9613dce2-6a39-4480-8b7f-553288ebc45b","Type":"ContainerStarted","Data":"d9f3f10ff3d8b51b000c86cbf7d13d6583ffc8d5c64cf29a31439819eda20147"} Oct 06 10:00:45 crc kubenswrapper[4824]: I1006 10:00:45.688651 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-j2m8k" podStartSLOduration=133.688630384 podStartE2EDuration="2m13.688630384s" podCreationTimestamp="2025-10-06 09:58:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:00:45.687626029 +0000 UTC m=+155.052048890" watchObservedRunningTime="2025-10-06 10:00:45.688630384 +0000 UTC m=+155.053053245" Oct 06 10:00:45 crc kubenswrapper[4824]: I1006 10:00:45.709048 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8l8gd\" (UID: \"c2f6e84b-a28c-4ead-97f5-978b0639e5bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-8l8gd" Oct 06 10:00:45 crc kubenswrapper[4824]: E1006 10:00:45.710239 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-06 10:00:46.210219695 +0000 UTC m=+155.574642556 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8l8gd" (UID: "c2f6e84b-a28c-4ead-97f5-978b0639e5bf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 10:00:45 crc kubenswrapper[4824]: I1006 10:00:45.718715 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7p7wb" event={"ID":"3bdcba24-670d-427a-bfd6-4372c9f944a2","Type":"ContainerStarted","Data":"2fbf638787e302e7f3c904c654c0b42e68d0b7f1c24c70209ba78431db2f2d86"} Oct 06 10:00:45 crc kubenswrapper[4824]: I1006 10:00:45.718774 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7p7wb" event={"ID":"3bdcba24-670d-427a-bfd6-4372c9f944a2","Type":"ContainerStarted","Data":"c854330000c79236aa87835f8e8b524b1e18def17bcc61893a2fe62395a856a2"} Oct 06 10:00:45 crc kubenswrapper[4824]: I1006 10:00:45.764347 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-s8j95" event={"ID":"ffceea22-78f2-4032-a23a-71b4fb7ea83e","Type":"ContainerStarted","Data":"ed78f76ee5ccbd5cad793a7af60805ecc372722b804623aa2105a8bdfb8fe7a2"} Oct 06 10:00:45 crc kubenswrapper[4824]: I1006 10:00:45.791774 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-vpg8w" event={"ID":"883a0642-0302-4fc5-b361-e25f1aba0878","Type":"ContainerStarted","Data":"7747475c7bb21bd47a6f652cce3ec98314abea403308dcb0ac792fb63a020c00"} Oct 06 10:00:45 crc kubenswrapper[4824]: I1006 10:00:45.793260 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-bjzzk" podStartSLOduration=134.793236221 podStartE2EDuration="2m14.793236221s" podCreationTimestamp="2025-10-06 09:58:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:00:45.758822613 +0000 UTC m=+155.123245474" watchObservedRunningTime="2025-10-06 10:00:45.793236221 +0000 UTC m=+155.157659082" Oct 06 10:00:45 crc kubenswrapper[4824]: I1006 10:00:45.810860 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 10:00:45 crc kubenswrapper[4824]: E1006 10:00:45.819173 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-06 10:00:46.319145499 +0000 UTC m=+155.683568360 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 10:00:45 crc kubenswrapper[4824]: I1006 10:00:45.820188 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-qpxrp" event={"ID":"92937768-de56-4cda-816c-9d1be4c07f0f","Type":"ContainerStarted","Data":"571c4f1aa159da79b45933f17a9416339cdc189017626aa1115101d49740a3c8"} Oct 06 10:00:45 crc kubenswrapper[4824]: I1006 10:00:45.864794 4824 patch_prober.go:28] interesting pod/router-default-5444994796-7qwsv container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 06 10:00:45 crc kubenswrapper[4824]: [-]has-synced failed: reason withheld Oct 06 10:00:45 crc kubenswrapper[4824]: [+]process-running ok Oct 06 10:00:45 crc kubenswrapper[4824]: healthz check failed Oct 06 10:00:45 crc kubenswrapper[4824]: I1006 10:00:45.864850 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-7qwsv" podUID="d6e255a4-d3e7-4cef-bf40-f98aa11b3e3b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 06 10:00:45 crc kubenswrapper[4824]: I1006 10:00:45.876386 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-qq4zz" event={"ID":"d2595750-aae8-4f17-8782-b7eccb0c2948","Type":"ContainerStarted","Data":"fde24d5c6f8274bd8d527ae7747c3941c580a57dc5640c813a74782ac1f9d4b5"} Oct 06 10:00:45 crc kubenswrapper[4824]: I1006 10:00:45.878231 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-v6gnw" podStartSLOduration=134.878217324 podStartE2EDuration="2m14.878217324s" podCreationTimestamp="2025-10-06 09:58:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:00:45.833130443 +0000 UTC m=+155.197553304" watchObservedRunningTime="2025-10-06 10:00:45.878217324 +0000 UTC m=+155.242640185" Oct 06 10:00:45 crc kubenswrapper[4824]: I1006 10:00:45.901518 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-fdbvz" event={"ID":"744def19-0006-4ab4-9d45-8a2a39bd410a","Type":"ContainerStarted","Data":"3da7a02497fa9d13f215a841fc1b576a592c0670b7a96188ffac6056b92e887a"} Oct 06 10:00:45 crc kubenswrapper[4824]: I1006 10:00:45.912425 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8l8gd\" (UID: \"c2f6e84b-a28c-4ead-97f5-978b0639e5bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-8l8gd" Oct 06 10:00:45 crc kubenswrapper[4824]: E1006 10:00:45.913054 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-06 10:00:46.413037912 +0000 UTC m=+155.777460773 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8l8gd" (UID: "c2f6e84b-a28c-4ead-97f5-978b0639e5bf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 10:00:45 crc kubenswrapper[4824]: I1006 10:00:45.928057 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7p7wb" podStartSLOduration=133.928033032 podStartE2EDuration="2m13.928033032s" podCreationTimestamp="2025-10-06 09:58:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:00:45.879715822 +0000 UTC m=+155.244138683" watchObservedRunningTime="2025-10-06 10:00:45.928033032 +0000 UTC m=+155.292455893" Oct 06 10:00:45 crc kubenswrapper[4824]: I1006 10:00:45.947608 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-nqmr6" Oct 06 10:00:46 crc kubenswrapper[4824]: I1006 10:00:46.013767 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 10:00:46 crc kubenswrapper[4824]: E1006 10:00:46.015090 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-06 10:00:46.515070546 +0000 UTC m=+155.879493407 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 10:00:46 crc kubenswrapper[4824]: I1006 10:00:46.032703 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-fdbvz" podStartSLOduration=6.03267986 podStartE2EDuration="6.03267986s" podCreationTimestamp="2025-10-06 10:00:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:00:45.929009476 +0000 UTC m=+155.293432357" watchObservedRunningTime="2025-10-06 10:00:46.03267986 +0000 UTC m=+155.397102721" Oct 06 10:00:46 crc kubenswrapper[4824]: I1006 10:00:46.116952 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8l8gd\" (UID: \"c2f6e84b-a28c-4ead-97f5-978b0639e5bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-8l8gd" Oct 06 10:00:46 crc kubenswrapper[4824]: E1006 10:00:46.117366 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-06 10:00:46.617351806 +0000 UTC m=+155.981774667 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8l8gd" (UID: "c2f6e84b-a28c-4ead-97f5-978b0639e5bf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 10:00:46 crc kubenswrapper[4824]: I1006 10:00:46.217998 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 10:00:46 crc kubenswrapper[4824]: E1006 10:00:46.218484 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-06 10:00:46.718465437 +0000 UTC m=+156.082888298 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 10:00:46 crc kubenswrapper[4824]: I1006 10:00:46.320092 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8l8gd\" (UID: \"c2f6e84b-a28c-4ead-97f5-978b0639e5bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-8l8gd" Oct 06 10:00:46 crc kubenswrapper[4824]: E1006 10:00:46.320450 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-06 10:00:46.820435429 +0000 UTC m=+156.184858290 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8l8gd" (UID: "c2f6e84b-a28c-4ead-97f5-978b0639e5bf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 10:00:46 crc kubenswrapper[4824]: I1006 10:00:46.422099 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 10:00:46 crc kubenswrapper[4824]: E1006 10:00:46.422311 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-06 10:00:46.922280199 +0000 UTC m=+156.286703060 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 10:00:46 crc kubenswrapper[4824]: I1006 10:00:46.422722 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8l8gd\" (UID: \"c2f6e84b-a28c-4ead-97f5-978b0639e5bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-8l8gd" Oct 06 10:00:46 crc kubenswrapper[4824]: E1006 10:00:46.423186 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-06 10:00:46.923169831 +0000 UTC m=+156.287592682 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8l8gd" (UID: "c2f6e84b-a28c-4ead-97f5-978b0639e5bf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 10:00:46 crc kubenswrapper[4824]: I1006 10:00:46.524622 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 10:00:46 crc kubenswrapper[4824]: E1006 10:00:46.524961 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-06 10:00:47.024941478 +0000 UTC m=+156.389364339 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 10:00:46 crc kubenswrapper[4824]: I1006 10:00:46.627433 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8l8gd\" (UID: \"c2f6e84b-a28c-4ead-97f5-978b0639e5bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-8l8gd" Oct 06 10:00:46 crc kubenswrapper[4824]: E1006 10:00:46.628390 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-06 10:00:47.128370296 +0000 UTC m=+156.492793157 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8l8gd" (UID: "c2f6e84b-a28c-4ead-97f5-978b0639e5bf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 10:00:46 crc kubenswrapper[4824]: I1006 10:00:46.729415 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 10:00:46 crc kubenswrapper[4824]: E1006 10:00:46.729888 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-06 10:00:47.229868467 +0000 UTC m=+156.594291328 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 10:00:46 crc kubenswrapper[4824]: I1006 10:00:46.833586 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8l8gd\" (UID: \"c2f6e84b-a28c-4ead-97f5-978b0639e5bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-8l8gd" Oct 06 10:00:46 crc kubenswrapper[4824]: E1006 10:00:46.834107 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-06 10:00:47.334084165 +0000 UTC m=+156.698507026 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8l8gd" (UID: "c2f6e84b-a28c-4ead-97f5-978b0639e5bf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 10:00:46 crc kubenswrapper[4824]: I1006 10:00:46.854691 4824 patch_prober.go:28] interesting pod/router-default-5444994796-7qwsv container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 06 10:00:46 crc kubenswrapper[4824]: [-]has-synced failed: reason withheld Oct 06 10:00:46 crc kubenswrapper[4824]: [+]process-running ok Oct 06 10:00:46 crc kubenswrapper[4824]: healthz check failed Oct 06 10:00:46 crc kubenswrapper[4824]: I1006 10:00:46.854769 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-7qwsv" podUID="d6e255a4-d3e7-4cef-bf40-f98aa11b3e3b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 06 10:00:46 crc kubenswrapper[4824]: I1006 10:00:46.936043 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 10:00:46 crc kubenswrapper[4824]: E1006 10:00:46.936404 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-06 10:00:47.436384025 +0000 UTC m=+156.800806886 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 10:00:46 crc kubenswrapper[4824]: I1006 10:00:46.976428 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7p7wb" event={"ID":"3bdcba24-670d-427a-bfd6-4372c9f944a2","Type":"ContainerStarted","Data":"e1b98fdcab65540684feb4aa664f30c885765ea584f0253574b66ae671972626"} Oct 06 10:00:47 crc kubenswrapper[4824]: I1006 10:00:47.016896 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-fjdwv" event={"ID":"b349b0af-0ba8-409f-a915-7fb06b2e3774","Type":"ContainerStarted","Data":"5dd1fb024f12e1cb87b9b2bd4944d80dd54e8ab962d18a8e4b5ebed0a023dfcc"} Oct 06 10:00:47 crc kubenswrapper[4824]: I1006 10:00:47.040848 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8l8gd\" (UID: \"c2f6e84b-a28c-4ead-97f5-978b0639e5bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-8l8gd" Oct 06 10:00:47 crc kubenswrapper[4824]: E1006 10:00:47.041311 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-06 10:00:47.54129373 +0000 UTC m=+156.905716591 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8l8gd" (UID: "c2f6e84b-a28c-4ead-97f5-978b0639e5bf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 10:00:47 crc kubenswrapper[4824]: I1006 10:00:47.083631 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" Oct 06 10:00:47 crc kubenswrapper[4824]: I1006 10:00:47.145615 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 10:00:47 crc kubenswrapper[4824]: E1006 10:00:47.146807 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-06 10:00:47.646785438 +0000 UTC m=+157.011208299 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 10:00:47 crc kubenswrapper[4824]: I1006 10:00:47.208113 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-fjdwv" podStartSLOduration=135.208063428 podStartE2EDuration="2m15.208063428s" podCreationTimestamp="2025-10-06 09:58:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:00:47.063258501 +0000 UTC m=+156.427681362" watchObservedRunningTime="2025-10-06 10:00:47.208063428 +0000 UTC m=+156.572486289" Oct 06 10:00:47 crc kubenswrapper[4824]: I1006 10:00:47.243736 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29329080-9kpdv" event={"ID":"26536a6b-3639-46ad-b024-5e21b7118fe7","Type":"ContainerStarted","Data":"04591b6d2b8e78446287d500ea05b4d620e1226e9a5b376c57340ea2b03971f9"} Oct 06 10:00:47 crc kubenswrapper[4824]: I1006 10:00:47.250885 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8l8gd\" (UID: \"c2f6e84b-a28c-4ead-97f5-978b0639e5bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-8l8gd" Oct 06 10:00:47 crc kubenswrapper[4824]: E1006 10:00:47.256907 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-06 10:00:47.756868801 +0000 UTC m=+157.121291662 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8l8gd" (UID: "c2f6e84b-a28c-4ead-97f5-978b0639e5bf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 10:00:47 crc kubenswrapper[4824]: I1006 10:00:47.302609 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-5blgj" podStartSLOduration=135.302580298 podStartE2EDuration="2m15.302580298s" podCreationTimestamp="2025-10-06 09:58:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:00:47.196611146 +0000 UTC m=+156.561034007" watchObservedRunningTime="2025-10-06 10:00:47.302580298 +0000 UTC m=+156.667003159" Oct 06 10:00:47 crc kubenswrapper[4824]: I1006 10:00:47.304907 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29329080-9kpdv" podStartSLOduration=47.304899514 podStartE2EDuration="47.304899514s" podCreationTimestamp="2025-10-06 10:00:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:00:47.297467431 +0000 UTC m=+156.661890292" watchObservedRunningTime="2025-10-06 10:00:47.304899514 +0000 UTC m=+156.669322375" Oct 06 10:00:47 crc kubenswrapper[4824]: I1006 10:00:47.318024 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-qq4zz" event={"ID":"d2595750-aae8-4f17-8782-b7eccb0c2948","Type":"ContainerStarted","Data":"d27b1d97570bdd6569bbf1dc7b40419c3ce223f7aac865ed7a1b5c65b7acfcaa"} Oct 06 10:00:47 crc kubenswrapper[4824]: I1006 10:00:47.318165 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-qq4zz" Oct 06 10:00:47 crc kubenswrapper[4824]: I1006 10:00:47.322613 4824 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-qq4zz container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.34:8080/healthz\": dial tcp 10.217.0.34:8080: connect: connection refused" start-of-body= Oct 06 10:00:47 crc kubenswrapper[4824]: I1006 10:00:47.322667 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-qq4zz" podUID="d2595750-aae8-4f17-8782-b7eccb0c2948" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.34:8080/healthz\": dial tcp 10.217.0.34:8080: connect: connection refused" Oct 06 10:00:47 crc kubenswrapper[4824]: I1006 10:00:47.353259 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 10:00:47 crc kubenswrapper[4824]: E1006 10:00:47.354457 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-06 10:00:47.854437225 +0000 UTC m=+157.218860086 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 10:00:47 crc kubenswrapper[4824]: I1006 10:00:47.366786 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-qq4zz" podStartSLOduration=135.366760109 podStartE2EDuration="2m15.366760109s" podCreationTimestamp="2025-10-06 09:58:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:00:47.366264076 +0000 UTC m=+156.730686937" watchObservedRunningTime="2025-10-06 10:00:47.366760109 +0000 UTC m=+156.731182970" Oct 06 10:00:47 crc kubenswrapper[4824]: I1006 10:00:47.367584 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-fn7dj" event={"ID":"2a76ccaa-9249-46ba-bbc5-dbbbc33b06bb","Type":"ContainerStarted","Data":"7ea8874c3ce1ff57886948645065d8434b961db27868957da2db7ddb4e594f35"} Oct 06 10:00:47 crc kubenswrapper[4824]: I1006 10:00:47.416199 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-nk67f" event={"ID":"a1d6c12e-ff5a-4687-80ca-5699dd1bb9fe","Type":"ContainerStarted","Data":"047ce80340fd9f16f1800e9c91b8bd076a8f46880ac7e06d9a079f275906c876"} Oct 06 10:00:47 crc kubenswrapper[4824]: I1006 10:00:47.436220 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-wp6kv" event={"ID":"d40e2672-6c82-45e6-a292-21615712032d","Type":"ContainerStarted","Data":"1e8651fef8c4b2cc7c4dac96c0cf206e1e3754d45d0433103d2bf0eeaffcfbae"} Oct 06 10:00:47 crc kubenswrapper[4824]: I1006 10:00:47.454879 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8l8gd\" (UID: \"c2f6e84b-a28c-4ead-97f5-978b0639e5bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-8l8gd" Oct 06 10:00:47 crc kubenswrapper[4824]: E1006 10:00:47.457149 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-06 10:00:47.957127575 +0000 UTC m=+157.321550436 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8l8gd" (UID: "c2f6e84b-a28c-4ead-97f5-978b0639e5bf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 10:00:47 crc kubenswrapper[4824]: I1006 10:00:47.501364 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-56948" event={"ID":"678eeee3-60d8-4dbe-b815-10de5a46c4d1","Type":"ContainerStarted","Data":"2c440f0ca9886b35c4c01ba888a75e61104809d0e4286f44cd38156650cf918e"} Oct 06 10:00:47 crc kubenswrapper[4824]: I1006 10:00:47.520193 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-fn7dj" podStartSLOduration=135.520171758 podStartE2EDuration="2m15.520171758s" podCreationTimestamp="2025-10-06 09:58:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:00:47.439892571 +0000 UTC m=+156.804315432" watchObservedRunningTime="2025-10-06 10:00:47.520171758 +0000 UTC m=+156.884594619" Oct 06 10:00:47 crc kubenswrapper[4824]: I1006 10:00:47.520895 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-wp6kv" podStartSLOduration=135.520889336 podStartE2EDuration="2m15.520889336s" podCreationTimestamp="2025-10-06 09:58:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:00:47.518560979 +0000 UTC m=+156.882983840" watchObservedRunningTime="2025-10-06 10:00:47.520889336 +0000 UTC m=+156.885312197" Oct 06 10:00:47 crc kubenswrapper[4824]: I1006 10:00:47.556720 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wlbrp" event={"ID":"9b1f3faf-72a1-44e2-814b-964902e78d3e","Type":"ContainerStarted","Data":"e4d472234362bef0aa2d940ebe5aa607aecf3e7071c6ac28fdb41b8fd77bc5c8"} Oct 06 10:00:47 crc kubenswrapper[4824]: I1006 10:00:47.557551 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 10:00:47 crc kubenswrapper[4824]: E1006 10:00:47.558915 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-06 10:00:48.058893122 +0000 UTC m=+157.423315983 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 10:00:47 crc kubenswrapper[4824]: I1006 10:00:47.629554 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-s8j95" event={"ID":"ffceea22-78f2-4032-a23a-71b4fb7ea83e","Type":"ContainerStarted","Data":"6774b169ac7407f329273f50fdabe9aeaee017a158d8d812c9817e54c9c6a4ff"} Oct 06 10:00:47 crc kubenswrapper[4824]: I1006 10:00:47.630668 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-s8j95" Oct 06 10:00:47 crc kubenswrapper[4824]: I1006 10:00:47.650504 4824 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-s8j95 container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.26:5443/healthz\": dial tcp 10.217.0.26:5443: connect: connection refused" start-of-body= Oct 06 10:00:47 crc kubenswrapper[4824]: I1006 10:00:47.650557 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-s8j95" podUID="ffceea22-78f2-4032-a23a-71b4fb7ea83e" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.26:5443/healthz\": dial tcp 10.217.0.26:5443: connect: connection refused" Oct 06 10:00:47 crc kubenswrapper[4824]: I1006 10:00:47.665875 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-8rxsm" event={"ID":"ff05ab6e-248f-445c-9856-c1a2868f8840","Type":"ContainerStarted","Data":"cd61a855be26b456f8698a30d62698bfcb0d1527215a2bb71e88c7ada4cb3b77"} Oct 06 10:00:47 crc kubenswrapper[4824]: I1006 10:00:47.667102 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8l8gd\" (UID: \"c2f6e84b-a28c-4ead-97f5-978b0639e5bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-8l8gd" Oct 06 10:00:47 crc kubenswrapper[4824]: E1006 10:00:47.669015 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-06 10:00:48.168961344 +0000 UTC m=+157.533384205 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8l8gd" (UID: "c2f6e84b-a28c-4ead-97f5-978b0639e5bf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 10:00:47 crc kubenswrapper[4824]: I1006 10:00:47.690881 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wlbrp" podStartSLOduration=136.690852313 podStartE2EDuration="2m16.690852313s" podCreationTimestamp="2025-10-06 09:58:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:00:47.665441497 +0000 UTC m=+157.029864358" watchObservedRunningTime="2025-10-06 10:00:47.690852313 +0000 UTC m=+157.055275174" Oct 06 10:00:47 crc kubenswrapper[4824]: I1006 10:00:47.708168 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-4hw5p" event={"ID":"9da4a453-8e05-4360-b74c-ad1c4706b005","Type":"ContainerStarted","Data":"4d621d82814a5fdb93da99e140999509034835761cdf380074735fa10e6a5ac2"} Oct 06 10:00:47 crc kubenswrapper[4824]: I1006 10:00:47.716314 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-vpg8w" event={"ID":"883a0642-0302-4fc5-b361-e25f1aba0878","Type":"ContainerStarted","Data":"fc399b26c176f0cfaeac61ece862a13e4ff3a99df7c235d2d18134ccfafa197f"} Oct 06 10:00:47 crc kubenswrapper[4824]: I1006 10:00:47.734475 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-x4fdc" event={"ID":"a47e3600-79be-46c3-a1b8-6678d8da80c5","Type":"ContainerStarted","Data":"9ef427ad8277b630f0f1b1ab9efb6a27bfb35bc3b2909c7421d1d3a9c4cf0a01"} Oct 06 10:00:47 crc kubenswrapper[4824]: I1006 10:00:47.761518 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vvtbh" event={"ID":"2a0286c5-9866-4a1c-9855-49cea5d04d69","Type":"ContainerStarted","Data":"4a0f99eb6dddd6b848d986ade6d7e9bf8bb87b4fdc53bf439924cc3a6eb9fac0"} Oct 06 10:00:47 crc kubenswrapper[4824]: I1006 10:00:47.778451 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wk4bd" Oct 06 10:00:47 crc kubenswrapper[4824]: I1006 10:00:47.779470 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 10:00:47 crc kubenswrapper[4824]: E1006 10:00:47.781206 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-06 10:00:48.281165729 +0000 UTC m=+157.645588750 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 10:00:47 crc kubenswrapper[4824]: I1006 10:00:47.785545 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-bjzzk" Oct 06 10:00:47 crc kubenswrapper[4824]: I1006 10:00:47.800472 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-2xznh" Oct 06 10:00:47 crc kubenswrapper[4824]: I1006 10:00:47.828324 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-s8j95" podStartSLOduration=135.828297199 podStartE2EDuration="2m15.828297199s" podCreationTimestamp="2025-10-06 09:58:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:00:47.755964198 +0000 UTC m=+157.120387059" watchObservedRunningTime="2025-10-06 10:00:47.828297199 +0000 UTC m=+157.192720060" Oct 06 10:00:47 crc kubenswrapper[4824]: I1006 10:00:47.829810 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-vpg8w" podStartSLOduration=7.829800526 podStartE2EDuration="7.829800526s" podCreationTimestamp="2025-10-06 10:00:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:00:47.823670506 +0000 UTC m=+157.188093377" watchObservedRunningTime="2025-10-06 10:00:47.829800526 +0000 UTC m=+157.194223387" Oct 06 10:00:47 crc kubenswrapper[4824]: I1006 10:00:47.857545 4824 patch_prober.go:28] interesting pod/router-default-5444994796-7qwsv container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 06 10:00:47 crc kubenswrapper[4824]: [-]has-synced failed: reason withheld Oct 06 10:00:47 crc kubenswrapper[4824]: [+]process-running ok Oct 06 10:00:47 crc kubenswrapper[4824]: healthz check failed Oct 06 10:00:47 crc kubenswrapper[4824]: I1006 10:00:47.857604 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-7qwsv" podUID="d6e255a4-d3e7-4cef-bf40-f98aa11b3e3b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 06 10:00:47 crc kubenswrapper[4824]: I1006 10:00:47.885331 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8l8gd\" (UID: \"c2f6e84b-a28c-4ead-97f5-978b0639e5bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-8l8gd" Oct 06 10:00:47 crc kubenswrapper[4824]: E1006 10:00:47.889293 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-06 10:00:48.389269042 +0000 UTC m=+157.753692083 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8l8gd" (UID: "c2f6e84b-a28c-4ead-97f5-978b0639e5bf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 10:00:47 crc kubenswrapper[4824]: I1006 10:00:47.943823 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-8rxsm" podStartSLOduration=135.943800356 podStartE2EDuration="2m15.943800356s" podCreationTimestamp="2025-10-06 09:58:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:00:47.862284598 +0000 UTC m=+157.226707459" watchObservedRunningTime="2025-10-06 10:00:47.943800356 +0000 UTC m=+157.308223217" Oct 06 10:00:47 crc kubenswrapper[4824]: I1006 10:00:47.986637 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 10:00:47 crc kubenswrapper[4824]: E1006 10:00:47.987123 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-06 10:00:48.487097102 +0000 UTC m=+157.851519963 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 10:00:47 crc kubenswrapper[4824]: I1006 10:00:47.993187 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-x4fdc" podStartSLOduration=136.993163492 podStartE2EDuration="2m16.993163492s" podCreationTimestamp="2025-10-06 09:58:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:00:47.991377118 +0000 UTC m=+157.355799989" watchObservedRunningTime="2025-10-06 10:00:47.993163492 +0000 UTC m=+157.357586353" Oct 06 10:00:47 crc kubenswrapper[4824]: I1006 10:00:47.994010 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-4hw5p" podStartSLOduration=136.994003833 podStartE2EDuration="2m16.994003833s" podCreationTimestamp="2025-10-06 09:58:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:00:47.942618706 +0000 UTC m=+157.307041567" watchObservedRunningTime="2025-10-06 10:00:47.994003833 +0000 UTC m=+157.358426684" Oct 06 10:00:48 crc kubenswrapper[4824]: I1006 10:00:48.088651 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8l8gd\" (UID: \"c2f6e84b-a28c-4ead-97f5-978b0639e5bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-8l8gd" Oct 06 10:00:48 crc kubenswrapper[4824]: E1006 10:00:48.089101 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-06 10:00:48.589086134 +0000 UTC m=+157.953508995 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8l8gd" (UID: "c2f6e84b-a28c-4ead-97f5-978b0639e5bf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 10:00:48 crc kubenswrapper[4824]: I1006 10:00:48.177017 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-tlvv7"] Oct 06 10:00:48 crc kubenswrapper[4824]: I1006 10:00:48.190231 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tlvv7" Oct 06 10:00:48 crc kubenswrapper[4824]: I1006 10:00:48.191143 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 10:00:48 crc kubenswrapper[4824]: E1006 10:00:48.192012 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-06 10:00:48.69197152 +0000 UTC m=+158.056394371 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 10:00:48 crc kubenswrapper[4824]: I1006 10:00:48.192119 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8l8gd\" (UID: \"c2f6e84b-a28c-4ead-97f5-978b0639e5bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-8l8gd" Oct 06 10:00:48 crc kubenswrapper[4824]: E1006 10:00:48.192872 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-06 10:00:48.692855252 +0000 UTC m=+158.057278113 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8l8gd" (UID: "c2f6e84b-a28c-4ead-97f5-978b0639e5bf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 10:00:48 crc kubenswrapper[4824]: I1006 10:00:48.194112 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-tlvv7"] Oct 06 10:00:48 crc kubenswrapper[4824]: I1006 10:00:48.208050 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Oct 06 10:00:48 crc kubenswrapper[4824]: I1006 10:00:48.208968 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vvtbh" podStartSLOduration=136.208951228 podStartE2EDuration="2m16.208951228s" podCreationTimestamp="2025-10-06 09:58:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:00:48.208255391 +0000 UTC m=+157.572678252" watchObservedRunningTime="2025-10-06 10:00:48.208951228 +0000 UTC m=+157.573374089" Oct 06 10:00:48 crc kubenswrapper[4824]: I1006 10:00:48.294688 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 10:00:48 crc kubenswrapper[4824]: I1006 10:00:48.294972 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/030ddedc-a25b-47f6-88b6-c1286dc1c910-catalog-content\") pod \"community-operators-tlvv7\" (UID: \"030ddedc-a25b-47f6-88b6-c1286dc1c910\") " pod="openshift-marketplace/community-operators-tlvv7" Oct 06 10:00:48 crc kubenswrapper[4824]: I1006 10:00:48.295114 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/030ddedc-a25b-47f6-88b6-c1286dc1c910-utilities\") pod \"community-operators-tlvv7\" (UID: \"030ddedc-a25b-47f6-88b6-c1286dc1c910\") " pod="openshift-marketplace/community-operators-tlvv7" Oct 06 10:00:48 crc kubenswrapper[4824]: I1006 10:00:48.295155 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-975sb\" (UniqueName: \"kubernetes.io/projected/030ddedc-a25b-47f6-88b6-c1286dc1c910-kube-api-access-975sb\") pod \"community-operators-tlvv7\" (UID: \"030ddedc-a25b-47f6-88b6-c1286dc1c910\") " pod="openshift-marketplace/community-operators-tlvv7" Oct 06 10:00:48 crc kubenswrapper[4824]: E1006 10:00:48.295265 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-06 10:00:48.795246464 +0000 UTC m=+158.159669325 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 10:00:48 crc kubenswrapper[4824]: I1006 10:00:48.352496 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-r52zf"] Oct 06 10:00:48 crc kubenswrapper[4824]: I1006 10:00:48.353477 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-r52zf" Oct 06 10:00:48 crc kubenswrapper[4824]: I1006 10:00:48.360494 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Oct 06 10:00:48 crc kubenswrapper[4824]: I1006 10:00:48.397535 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8l8gd\" (UID: \"c2f6e84b-a28c-4ead-97f5-978b0639e5bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-8l8gd" Oct 06 10:00:48 crc kubenswrapper[4824]: I1006 10:00:48.397603 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/06d3772b-e467-41a1-bf9b-9d6520ab68ef-catalog-content\") pod \"certified-operators-r52zf\" (UID: \"06d3772b-e467-41a1-bf9b-9d6520ab68ef\") " pod="openshift-marketplace/certified-operators-r52zf" Oct 06 10:00:48 crc kubenswrapper[4824]: I1006 10:00:48.397684 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/06d3772b-e467-41a1-bf9b-9d6520ab68ef-utilities\") pod \"certified-operators-r52zf\" (UID: \"06d3772b-e467-41a1-bf9b-9d6520ab68ef\") " pod="openshift-marketplace/certified-operators-r52zf" Oct 06 10:00:48 crc kubenswrapper[4824]: I1006 10:00:48.397709 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fhs24\" (UniqueName: \"kubernetes.io/projected/06d3772b-e467-41a1-bf9b-9d6520ab68ef-kube-api-access-fhs24\") pod \"certified-operators-r52zf\" (UID: \"06d3772b-e467-41a1-bf9b-9d6520ab68ef\") " pod="openshift-marketplace/certified-operators-r52zf" Oct 06 10:00:48 crc kubenswrapper[4824]: I1006 10:00:48.397730 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/030ddedc-a25b-47f6-88b6-c1286dc1c910-utilities\") pod \"community-operators-tlvv7\" (UID: \"030ddedc-a25b-47f6-88b6-c1286dc1c910\") " pod="openshift-marketplace/community-operators-tlvv7" Oct 06 10:00:48 crc kubenswrapper[4824]: I1006 10:00:48.397765 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-975sb\" (UniqueName: \"kubernetes.io/projected/030ddedc-a25b-47f6-88b6-c1286dc1c910-kube-api-access-975sb\") pod \"community-operators-tlvv7\" (UID: \"030ddedc-a25b-47f6-88b6-c1286dc1c910\") " pod="openshift-marketplace/community-operators-tlvv7" Oct 06 10:00:48 crc kubenswrapper[4824]: I1006 10:00:48.397787 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/030ddedc-a25b-47f6-88b6-c1286dc1c910-catalog-content\") pod \"community-operators-tlvv7\" (UID: \"030ddedc-a25b-47f6-88b6-c1286dc1c910\") " pod="openshift-marketplace/community-operators-tlvv7" Oct 06 10:00:48 crc kubenswrapper[4824]: I1006 10:00:48.398264 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/030ddedc-a25b-47f6-88b6-c1286dc1c910-catalog-content\") pod \"community-operators-tlvv7\" (UID: \"030ddedc-a25b-47f6-88b6-c1286dc1c910\") " pod="openshift-marketplace/community-operators-tlvv7" Oct 06 10:00:48 crc kubenswrapper[4824]: E1006 10:00:48.398429 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-06 10:00:48.898406635 +0000 UTC m=+158.262829496 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8l8gd" (UID: "c2f6e84b-a28c-4ead-97f5-978b0639e5bf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 10:00:48 crc kubenswrapper[4824]: I1006 10:00:48.398500 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/030ddedc-a25b-47f6-88b6-c1286dc1c910-utilities\") pod \"community-operators-tlvv7\" (UID: \"030ddedc-a25b-47f6-88b6-c1286dc1c910\") " pod="openshift-marketplace/community-operators-tlvv7" Oct 06 10:00:48 crc kubenswrapper[4824]: I1006 10:00:48.426404 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-r52zf"] Oct 06 10:00:48 crc kubenswrapper[4824]: I1006 10:00:48.441159 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-975sb\" (UniqueName: \"kubernetes.io/projected/030ddedc-a25b-47f6-88b6-c1286dc1c910-kube-api-access-975sb\") pod \"community-operators-tlvv7\" (UID: \"030ddedc-a25b-47f6-88b6-c1286dc1c910\") " pod="openshift-marketplace/community-operators-tlvv7" Oct 06 10:00:48 crc kubenswrapper[4824]: I1006 10:00:48.499240 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 10:00:48 crc kubenswrapper[4824]: I1006 10:00:48.499898 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/06d3772b-e467-41a1-bf9b-9d6520ab68ef-catalog-content\") pod \"certified-operators-r52zf\" (UID: \"06d3772b-e467-41a1-bf9b-9d6520ab68ef\") " pod="openshift-marketplace/certified-operators-r52zf" Oct 06 10:00:48 crc kubenswrapper[4824]: I1006 10:00:48.499987 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/06d3772b-e467-41a1-bf9b-9d6520ab68ef-utilities\") pod \"certified-operators-r52zf\" (UID: \"06d3772b-e467-41a1-bf9b-9d6520ab68ef\") " pod="openshift-marketplace/certified-operators-r52zf" Oct 06 10:00:48 crc kubenswrapper[4824]: I1006 10:00:48.500010 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fhs24\" (UniqueName: \"kubernetes.io/projected/06d3772b-e467-41a1-bf9b-9d6520ab68ef-kube-api-access-fhs24\") pod \"certified-operators-r52zf\" (UID: \"06d3772b-e467-41a1-bf9b-9d6520ab68ef\") " pod="openshift-marketplace/certified-operators-r52zf" Oct 06 10:00:48 crc kubenswrapper[4824]: I1006 10:00:48.501144 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/06d3772b-e467-41a1-bf9b-9d6520ab68ef-catalog-content\") pod \"certified-operators-r52zf\" (UID: \"06d3772b-e467-41a1-bf9b-9d6520ab68ef\") " pod="openshift-marketplace/certified-operators-r52zf" Oct 06 10:00:48 crc kubenswrapper[4824]: E1006 10:00:48.501258 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-06 10:00:49.001213039 +0000 UTC m=+158.365635950 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 10:00:48 crc kubenswrapper[4824]: I1006 10:00:48.501396 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/06d3772b-e467-41a1-bf9b-9d6520ab68ef-utilities\") pod \"certified-operators-r52zf\" (UID: \"06d3772b-e467-41a1-bf9b-9d6520ab68ef\") " pod="openshift-marketplace/certified-operators-r52zf" Oct 06 10:00:48 crc kubenswrapper[4824]: I1006 10:00:48.556594 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fhs24\" (UniqueName: \"kubernetes.io/projected/06d3772b-e467-41a1-bf9b-9d6520ab68ef-kube-api-access-fhs24\") pod \"certified-operators-r52zf\" (UID: \"06d3772b-e467-41a1-bf9b-9d6520ab68ef\") " pod="openshift-marketplace/certified-operators-r52zf" Oct 06 10:00:48 crc kubenswrapper[4824]: I1006 10:00:48.563192 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-zlxdl"] Oct 06 10:00:48 crc kubenswrapper[4824]: I1006 10:00:48.564136 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zlxdl" Oct 06 10:00:48 crc kubenswrapper[4824]: I1006 10:00:48.569604 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tlvv7" Oct 06 10:00:48 crc kubenswrapper[4824]: I1006 10:00:48.590767 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zlxdl"] Oct 06 10:00:48 crc kubenswrapper[4824]: I1006 10:00:48.602534 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8l8gd\" (UID: \"c2f6e84b-a28c-4ead-97f5-978b0639e5bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-8l8gd" Oct 06 10:00:48 crc kubenswrapper[4824]: I1006 10:00:48.602602 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/527511c8-2487-4416-a689-9ee161130297-catalog-content\") pod \"community-operators-zlxdl\" (UID: \"527511c8-2487-4416-a689-9ee161130297\") " pod="openshift-marketplace/community-operators-zlxdl" Oct 06 10:00:48 crc kubenswrapper[4824]: I1006 10:00:48.602642 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vc8nz\" (UniqueName: \"kubernetes.io/projected/527511c8-2487-4416-a689-9ee161130297-kube-api-access-vc8nz\") pod \"community-operators-zlxdl\" (UID: \"527511c8-2487-4416-a689-9ee161130297\") " pod="openshift-marketplace/community-operators-zlxdl" Oct 06 10:00:48 crc kubenswrapper[4824]: I1006 10:00:48.602720 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/527511c8-2487-4416-a689-9ee161130297-utilities\") pod \"community-operators-zlxdl\" (UID: \"527511c8-2487-4416-a689-9ee161130297\") " pod="openshift-marketplace/community-operators-zlxdl" Oct 06 10:00:48 crc kubenswrapper[4824]: E1006 10:00:48.603134 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-06 10:00:49.10310835 +0000 UTC m=+158.467531211 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8l8gd" (UID: "c2f6e84b-a28c-4ead-97f5-978b0639e5bf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 10:00:48 crc kubenswrapper[4824]: I1006 10:00:48.688971 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-r52zf" Oct 06 10:00:48 crc kubenswrapper[4824]: I1006 10:00:48.707881 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 10:00:48 crc kubenswrapper[4824]: I1006 10:00:48.708220 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/527511c8-2487-4416-a689-9ee161130297-utilities\") pod \"community-operators-zlxdl\" (UID: \"527511c8-2487-4416-a689-9ee161130297\") " pod="openshift-marketplace/community-operators-zlxdl" Oct 06 10:00:48 crc kubenswrapper[4824]: I1006 10:00:48.708269 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/527511c8-2487-4416-a689-9ee161130297-catalog-content\") pod \"community-operators-zlxdl\" (UID: \"527511c8-2487-4416-a689-9ee161130297\") " pod="openshift-marketplace/community-operators-zlxdl" Oct 06 10:00:48 crc kubenswrapper[4824]: I1006 10:00:48.708296 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vc8nz\" (UniqueName: \"kubernetes.io/projected/527511c8-2487-4416-a689-9ee161130297-kube-api-access-vc8nz\") pod \"community-operators-zlxdl\" (UID: \"527511c8-2487-4416-a689-9ee161130297\") " pod="openshift-marketplace/community-operators-zlxdl" Oct 06 10:00:48 crc kubenswrapper[4824]: E1006 10:00:48.708691 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-06 10:00:49.20867227 +0000 UTC m=+158.573095131 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 10:00:48 crc kubenswrapper[4824]: I1006 10:00:48.709070 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/527511c8-2487-4416-a689-9ee161130297-utilities\") pod \"community-operators-zlxdl\" (UID: \"527511c8-2487-4416-a689-9ee161130297\") " pod="openshift-marketplace/community-operators-zlxdl" Oct 06 10:00:48 crc kubenswrapper[4824]: I1006 10:00:48.709272 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/527511c8-2487-4416-a689-9ee161130297-catalog-content\") pod \"community-operators-zlxdl\" (UID: \"527511c8-2487-4416-a689-9ee161130297\") " pod="openshift-marketplace/community-operators-zlxdl" Oct 06 10:00:48 crc kubenswrapper[4824]: I1006 10:00:48.772824 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vc8nz\" (UniqueName: \"kubernetes.io/projected/527511c8-2487-4416-a689-9ee161130297-kube-api-access-vc8nz\") pod \"community-operators-zlxdl\" (UID: \"527511c8-2487-4416-a689-9ee161130297\") " pod="openshift-marketplace/community-operators-zlxdl" Oct 06 10:00:48 crc kubenswrapper[4824]: I1006 10:00:48.790134 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-dtcm6"] Oct 06 10:00:48 crc kubenswrapper[4824]: I1006 10:00:48.791257 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dtcm6" Oct 06 10:00:48 crc kubenswrapper[4824]: I1006 10:00:48.809320 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8l8gd\" (UID: \"c2f6e84b-a28c-4ead-97f5-978b0639e5bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-8l8gd" Oct 06 10:00:48 crc kubenswrapper[4824]: E1006 10:00:48.809680 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-06 10:00:49.309666238 +0000 UTC m=+158.674089099 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8l8gd" (UID: "c2f6e84b-a28c-4ead-97f5-978b0639e5bf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 10:00:48 crc kubenswrapper[4824]: I1006 10:00:48.810702 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dtcm6"] Oct 06 10:00:48 crc kubenswrapper[4824]: I1006 10:00:48.810872 4824 generic.go:334] "Generic (PLEG): container finished" podID="2fa5ee3e-29f4-43b2-a154-3c988754897a" containerID="2bde3a0a1204621439bf1209fc525e9d74c8e021a0751b5bf6570f3dfe8b6cb8" exitCode=0 Oct 06 10:00:48 crc kubenswrapper[4824]: I1006 10:00:48.811645 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zzrfg" event={"ID":"2fa5ee3e-29f4-43b2-a154-3c988754897a","Type":"ContainerDied","Data":"2bde3a0a1204621439bf1209fc525e9d74c8e021a0751b5bf6570f3dfe8b6cb8"} Oct 06 10:00:48 crc kubenswrapper[4824]: I1006 10:00:48.826703 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-qpxrp" event={"ID":"92937768-de56-4cda-816c-9d1be4c07f0f","Type":"ContainerStarted","Data":"648e05b90a44df2fbaf820dce8de2d8d6e3e54d47092bbb59ed38c452306d450"} Oct 06 10:00:48 crc kubenswrapper[4824]: I1006 10:00:48.826771 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-qpxrp" event={"ID":"92937768-de56-4cda-816c-9d1be4c07f0f","Type":"ContainerStarted","Data":"fb0bd76a4ad220541e539347f8feb9866f1c42696cf28f83450915be6276f8c0"} Oct 06 10:00:48 crc kubenswrapper[4824]: I1006 10:00:48.833395 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-v6gnw" Oct 06 10:00:48 crc kubenswrapper[4824]: I1006 10:00:48.855213 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-dr4mr" event={"ID":"c682e90b-3a5a-4b87-996b-8365b0936bda","Type":"ContainerStarted","Data":"930e0d9dd69a840749612c7d23a9d67d4e5d1609af5a1a659e617339ba98b4fa"} Oct 06 10:00:48 crc kubenswrapper[4824]: I1006 10:00:48.855685 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-dr4mr" Oct 06 10:00:48 crc kubenswrapper[4824]: I1006 10:00:48.859016 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-q76nt" event={"ID":"911b7e6b-1df0-4ef3-822e-84925bd6bbdc","Type":"ContainerStarted","Data":"14ee05911d75fb6cdd0ae0e71a4b6aa3a695f2e5593ef3bce82389c824f77593"} Oct 06 10:00:48 crc kubenswrapper[4824]: I1006 10:00:48.859070 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-q76nt" event={"ID":"911b7e6b-1df0-4ef3-822e-84925bd6bbdc","Type":"ContainerStarted","Data":"b240a5e776f7c418d3a5c489296a7b9b65f5f78d3d2ed1c94c6ac9b6d3084bbc"} Oct 06 10:00:48 crc kubenswrapper[4824]: I1006 10:00:48.872806 4824 patch_prober.go:28] interesting pod/router-default-5444994796-7qwsv container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 06 10:00:48 crc kubenswrapper[4824]: [-]has-synced failed: reason withheld Oct 06 10:00:48 crc kubenswrapper[4824]: [+]process-running ok Oct 06 10:00:48 crc kubenswrapper[4824]: healthz check failed Oct 06 10:00:48 crc kubenswrapper[4824]: I1006 10:00:48.872881 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-7qwsv" podUID="d6e255a4-d3e7-4cef-bf40-f98aa11b3e3b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 06 10:00:48 crc kubenswrapper[4824]: I1006 10:00:48.887668 4824 generic.go:334] "Generic (PLEG): container finished" podID="26536a6b-3639-46ad-b024-5e21b7118fe7" containerID="04591b6d2b8e78446287d500ea05b4d620e1226e9a5b376c57340ea2b03971f9" exitCode=0 Oct 06 10:00:48 crc kubenswrapper[4824]: I1006 10:00:48.887947 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29329080-9kpdv" event={"ID":"26536a6b-3639-46ad-b024-5e21b7118fe7","Type":"ContainerDied","Data":"04591b6d2b8e78446287d500ea05b4d620e1226e9a5b376c57340ea2b03971f9"} Oct 06 10:00:48 crc kubenswrapper[4824]: I1006 10:00:48.911472 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 10:00:48 crc kubenswrapper[4824]: I1006 10:00:48.912023 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dqjms\" (UniqueName: \"kubernetes.io/projected/88764df0-c119-4e11-bd94-6818bc8ef83c-kube-api-access-dqjms\") pod \"certified-operators-dtcm6\" (UID: \"88764df0-c119-4e11-bd94-6818bc8ef83c\") " pod="openshift-marketplace/certified-operators-dtcm6" Oct 06 10:00:48 crc kubenswrapper[4824]: I1006 10:00:48.912089 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/88764df0-c119-4e11-bd94-6818bc8ef83c-catalog-content\") pod \"certified-operators-dtcm6\" (UID: \"88764df0-c119-4e11-bd94-6818bc8ef83c\") " pod="openshift-marketplace/certified-operators-dtcm6" Oct 06 10:00:48 crc kubenswrapper[4824]: I1006 10:00:48.912153 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/88764df0-c119-4e11-bd94-6818bc8ef83c-utilities\") pod \"certified-operators-dtcm6\" (UID: \"88764df0-c119-4e11-bd94-6818bc8ef83c\") " pod="openshift-marketplace/certified-operators-dtcm6" Oct 06 10:00:48 crc kubenswrapper[4824]: E1006 10:00:48.912285 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-06 10:00:49.412268786 +0000 UTC m=+158.776691647 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 10:00:48 crc kubenswrapper[4824]: I1006 10:00:48.918485 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zlxdl" Oct 06 10:00:48 crc kubenswrapper[4824]: I1006 10:00:48.943135 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-qpxrp" podStartSLOduration=136.943111176 podStartE2EDuration="2m16.943111176s" podCreationTimestamp="2025-10-06 09:58:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:00:48.942553352 +0000 UTC m=+158.306976213" watchObservedRunningTime="2025-10-06 10:00:48.943111176 +0000 UTC m=+158.307534037" Oct 06 10:00:48 crc kubenswrapper[4824]: I1006 10:00:48.948617 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-56948" event={"ID":"678eeee3-60d8-4dbe-b815-10de5a46c4d1","Type":"ContainerStarted","Data":"39088013c08b523ae99819270220cf600c5e24c24869919becb18f999a6be23e"} Oct 06 10:00:48 crc kubenswrapper[4824]: I1006 10:00:48.965707 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-vn26d" event={"ID":"8e853926-f1e6-4571-863b-97918e660677","Type":"ContainerStarted","Data":"4c84b4b65134b9134f20ad304be92b1c7599e97904ee3281bd97741e96902761"} Oct 06 10:00:48 crc kubenswrapper[4824]: I1006 10:00:48.968415 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-8z68k" event={"ID":"af8fbf5a-f90e-4d22-ae65-05d8d8253308","Type":"ContainerStarted","Data":"119cba5ffd73d0ff173f47322893c5b169be0630ac8c94dbbc819afb9ec746eb"} Oct 06 10:00:48 crc kubenswrapper[4824]: I1006 10:00:48.969111 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-8z68k" Oct 06 10:00:48 crc kubenswrapper[4824]: I1006 10:00:48.986889 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-5blgj" event={"ID":"46e11b34-d0dd-4ea6-bed7-66eb707c7f92","Type":"ContainerStarted","Data":"c3c70c09fb0a27351fec65ba74a17bcab50aa3fc9b4b68c7f63a39e0019eb8ca"} Oct 06 10:00:49 crc kubenswrapper[4824]: I1006 10:00:49.021127 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/88764df0-c119-4e11-bd94-6818bc8ef83c-utilities\") pod \"certified-operators-dtcm6\" (UID: \"88764df0-c119-4e11-bd94-6818bc8ef83c\") " pod="openshift-marketplace/certified-operators-dtcm6" Oct 06 10:00:49 crc kubenswrapper[4824]: I1006 10:00:49.021188 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8l8gd\" (UID: \"c2f6e84b-a28c-4ead-97f5-978b0639e5bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-8l8gd" Oct 06 10:00:49 crc kubenswrapper[4824]: I1006 10:00:49.021394 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dqjms\" (UniqueName: \"kubernetes.io/projected/88764df0-c119-4e11-bd94-6818bc8ef83c-kube-api-access-dqjms\") pod \"certified-operators-dtcm6\" (UID: \"88764df0-c119-4e11-bd94-6818bc8ef83c\") " pod="openshift-marketplace/certified-operators-dtcm6" Oct 06 10:00:49 crc kubenswrapper[4824]: I1006 10:00:49.021454 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/88764df0-c119-4e11-bd94-6818bc8ef83c-catalog-content\") pod \"certified-operators-dtcm6\" (UID: \"88764df0-c119-4e11-bd94-6818bc8ef83c\") " pod="openshift-marketplace/certified-operators-dtcm6" Oct 06 10:00:49 crc kubenswrapper[4824]: I1006 10:00:49.023405 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/88764df0-c119-4e11-bd94-6818bc8ef83c-utilities\") pod \"certified-operators-dtcm6\" (UID: \"88764df0-c119-4e11-bd94-6818bc8ef83c\") " pod="openshift-marketplace/certified-operators-dtcm6" Oct 06 10:00:49 crc kubenswrapper[4824]: E1006 10:00:49.023730 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-06 10:00:49.523714402 +0000 UTC m=+158.888137263 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8l8gd" (UID: "c2f6e84b-a28c-4ead-97f5-978b0639e5bf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 10:00:49 crc kubenswrapper[4824]: I1006 10:00:49.029879 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6tt2t" event={"ID":"982129a4-11b4-4024-afb0-ac6024165c7a","Type":"ContainerStarted","Data":"412bb0a1f8f624bd658b925d61f37b66b2f1f351345d0ec7a69a79c3d73411ae"} Oct 06 10:00:49 crc kubenswrapper[4824]: I1006 10:00:49.043131 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/88764df0-c119-4e11-bd94-6818bc8ef83c-catalog-content\") pod \"certified-operators-dtcm6\" (UID: \"88764df0-c119-4e11-bd94-6818bc8ef83c\") " pod="openshift-marketplace/certified-operators-dtcm6" Oct 06 10:00:49 crc kubenswrapper[4824]: I1006 10:00:49.044952 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6tt2t" Oct 06 10:00:49 crc kubenswrapper[4824]: I1006 10:00:49.097143 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-56948" podStartSLOduration=137.09711903 podStartE2EDuration="2m17.09711903s" podCreationTimestamp="2025-10-06 09:58:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:00:49.021657552 +0000 UTC m=+158.386080413" watchObservedRunningTime="2025-10-06 10:00:49.09711903 +0000 UTC m=+158.461541891" Oct 06 10:00:49 crc kubenswrapper[4824]: I1006 10:00:49.104760 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-6kn7b" event={"ID":"7532dafc-fb61-4066-bb24-2470e4b2cfe6","Type":"ContainerStarted","Data":"88c3c811125696e57b52533912db729e8b06b144b251a1203b787108dd1ee62f"} Oct 06 10:00:49 crc kubenswrapper[4824]: I1006 10:00:49.124775 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 10:00:49 crc kubenswrapper[4824]: E1006 10:00:49.131014 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-06 10:00:49.630991435 +0000 UTC m=+158.995414296 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 10:00:49 crc kubenswrapper[4824]: I1006 10:00:49.132223 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dqjms\" (UniqueName: \"kubernetes.io/projected/88764df0-c119-4e11-bd94-6818bc8ef83c-kube-api-access-dqjms\") pod \"certified-operators-dtcm6\" (UID: \"88764df0-c119-4e11-bd94-6818bc8ef83c\") " pod="openshift-marketplace/certified-operators-dtcm6" Oct 06 10:00:49 crc kubenswrapper[4824]: I1006 10:00:49.133465 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dtcm6" Oct 06 10:00:49 crc kubenswrapper[4824]: I1006 10:00:49.134674 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6tt2t" Oct 06 10:00:49 crc kubenswrapper[4824]: I1006 10:00:49.173358 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-px5fp" event={"ID":"9613dce2-6a39-4480-8b7f-553288ebc45b","Type":"ContainerStarted","Data":"f454400d5dd8770bf6d0ba64cfbec09afc540e9c7f839fc8dc6f9db68c6a2d76"} Oct 06 10:00:49 crc kubenswrapper[4824]: I1006 10:00:49.174200 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-px5fp" Oct 06 10:00:49 crc kubenswrapper[4824]: I1006 10:00:49.206074 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vvtbh" event={"ID":"2a0286c5-9866-4a1c-9855-49cea5d04d69","Type":"ContainerStarted","Data":"40d1baa0a0dbc45e642b8275ac1f18ad1d43c733364aedd7a044e1e1b7bddba7"} Oct 06 10:00:49 crc kubenswrapper[4824]: I1006 10:00:49.226527 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-dr4mr" podStartSLOduration=9.226499298 podStartE2EDuration="9.226499298s" podCreationTimestamp="2025-10-06 10:00:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:00:49.221416273 +0000 UTC m=+158.585839134" watchObservedRunningTime="2025-10-06 10:00:49.226499298 +0000 UTC m=+158.590922159" Oct 06 10:00:49 crc kubenswrapper[4824]: I1006 10:00:49.232955 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8l8gd\" (UID: \"c2f6e84b-a28c-4ead-97f5-978b0639e5bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-8l8gd" Oct 06 10:00:49 crc kubenswrapper[4824]: E1006 10:00:49.234433 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-06 10:00:49.734419564 +0000 UTC m=+159.098842425 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8l8gd" (UID: "c2f6e84b-a28c-4ead-97f5-978b0639e5bf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 10:00:49 crc kubenswrapper[4824]: I1006 10:00:49.234739 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-xr4mg" event={"ID":"e81889bb-3225-458e-bce5-cfbafe6f4f92","Type":"ContainerStarted","Data":"33041b15810a51878f7b8f53a534a311703a348dfbbfab420f1ec5b3f4f454d3"} Oct 06 10:00:49 crc kubenswrapper[4824]: I1006 10:00:49.235624 4824 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-qq4zz container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.34:8080/healthz\": dial tcp 10.217.0.34:8080: connect: connection refused" start-of-body= Oct 06 10:00:49 crc kubenswrapper[4824]: I1006 10:00:49.235659 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-qq4zz" podUID="d2595750-aae8-4f17-8782-b7eccb0c2948" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.34:8080/healthz\": dial tcp 10.217.0.34:8080: connect: connection refused" Oct 06 10:00:49 crc kubenswrapper[4824]: I1006 10:00:49.321988 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-8z68k" podStartSLOduration=137.321944229 podStartE2EDuration="2m17.321944229s" podCreationTimestamp="2025-10-06 09:58:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:00:49.279700249 +0000 UTC m=+158.644123110" watchObservedRunningTime="2025-10-06 10:00:49.321944229 +0000 UTC m=+158.686367090" Oct 06 10:00:49 crc kubenswrapper[4824]: I1006 10:00:49.338856 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 10:00:49 crc kubenswrapper[4824]: E1006 10:00:49.352197 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-06 10:00:49.852149764 +0000 UTC m=+159.216572615 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 10:00:49 crc kubenswrapper[4824]: I1006 10:00:49.374347 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8l8gd\" (UID: \"c2f6e84b-a28c-4ead-97f5-978b0639e5bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-8l8gd" Oct 06 10:00:49 crc kubenswrapper[4824]: E1006 10:00:49.450392 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-06 10:00:49.950356744 +0000 UTC m=+159.314779605 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8l8gd" (UID: "c2f6e84b-a28c-4ead-97f5-978b0639e5bf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 10:00:49 crc kubenswrapper[4824]: I1006 10:00:49.465452 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-vn26d" podStartSLOduration=137.465422714 podStartE2EDuration="2m17.465422714s" podCreationTimestamp="2025-10-06 09:58:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:00:49.323408586 +0000 UTC m=+158.687831447" watchObservedRunningTime="2025-10-06 10:00:49.465422714 +0000 UTC m=+158.829845575" Oct 06 10:00:49 crc kubenswrapper[4824]: I1006 10:00:49.467548 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-q76nt" podStartSLOduration=138.467543747 podStartE2EDuration="2m18.467543747s" podCreationTimestamp="2025-10-06 09:58:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:00:49.397497481 +0000 UTC m=+158.761920342" watchObservedRunningTime="2025-10-06 10:00:49.467543747 +0000 UTC m=+158.831966598" Oct 06 10:00:49 crc kubenswrapper[4824]: I1006 10:00:49.481067 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 10:00:49 crc kubenswrapper[4824]: E1006 10:00:49.481522 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-06 10:00:49.981502801 +0000 UTC m=+159.345925662 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 10:00:49 crc kubenswrapper[4824]: I1006 10:00:49.520958 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-tlvv7"] Oct 06 10:00:49 crc kubenswrapper[4824]: I1006 10:00:49.524764 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-px5fp" podStartSLOduration=137.524735996 podStartE2EDuration="2m17.524735996s" podCreationTimestamp="2025-10-06 09:58:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:00:49.508867355 +0000 UTC m=+158.873290216" watchObservedRunningTime="2025-10-06 10:00:49.524735996 +0000 UTC m=+158.889158857" Oct 06 10:00:49 crc kubenswrapper[4824]: I1006 10:00:49.579998 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6tt2t" podStartSLOduration=137.579959696 podStartE2EDuration="2m17.579959696s" podCreationTimestamp="2025-10-06 09:58:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:00:49.578570302 +0000 UTC m=+158.942993153" watchObservedRunningTime="2025-10-06 10:00:49.579959696 +0000 UTC m=+158.944382557" Oct 06 10:00:49 crc kubenswrapper[4824]: I1006 10:00:49.593096 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8l8gd\" (UID: \"c2f6e84b-a28c-4ead-97f5-978b0639e5bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-8l8gd" Oct 06 10:00:49 crc kubenswrapper[4824]: E1006 10:00:49.593439 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-06 10:00:50.093425858 +0000 UTC m=+159.457848719 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8l8gd" (UID: "c2f6e84b-a28c-4ead-97f5-978b0639e5bf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 10:00:49 crc kubenswrapper[4824]: I1006 10:00:49.694873 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 10:00:49 crc kubenswrapper[4824]: E1006 10:00:49.695339 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-06 10:00:50.195318489 +0000 UTC m=+159.559741350 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 10:00:49 crc kubenswrapper[4824]: I1006 10:00:49.718931 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-6kn7b" podStartSLOduration=138.71890948 podStartE2EDuration="2m18.71890948s" podCreationTimestamp="2025-10-06 09:58:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:00:49.716345117 +0000 UTC m=+159.080767978" watchObservedRunningTime="2025-10-06 10:00:49.71890948 +0000 UTC m=+159.083332341" Oct 06 10:00:49 crc kubenswrapper[4824]: I1006 10:00:49.719588 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-xr4mg" podStartSLOduration=137.719582606 podStartE2EDuration="2m17.719582606s" podCreationTimestamp="2025-10-06 09:58:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:00:49.629831565 +0000 UTC m=+158.994254426" watchObservedRunningTime="2025-10-06 10:00:49.719582606 +0000 UTC m=+159.084005467" Oct 06 10:00:49 crc kubenswrapper[4824]: I1006 10:00:49.796807 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8l8gd\" (UID: \"c2f6e84b-a28c-4ead-97f5-978b0639e5bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-8l8gd" Oct 06 10:00:49 crc kubenswrapper[4824]: E1006 10:00:49.797572 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-06 10:00:50.297550268 +0000 UTC m=+159.661973129 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8l8gd" (UID: "c2f6e84b-a28c-4ead-97f5-978b0639e5bf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 10:00:49 crc kubenswrapper[4824]: I1006 10:00:49.864582 4824 patch_prober.go:28] interesting pod/router-default-5444994796-7qwsv container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 06 10:00:49 crc kubenswrapper[4824]: [-]has-synced failed: reason withheld Oct 06 10:00:49 crc kubenswrapper[4824]: [+]process-running ok Oct 06 10:00:49 crc kubenswrapper[4824]: healthz check failed Oct 06 10:00:49 crc kubenswrapper[4824]: I1006 10:00:49.865184 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-7qwsv" podUID="d6e255a4-d3e7-4cef-bf40-f98aa11b3e3b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 06 10:00:49 crc kubenswrapper[4824]: I1006 10:00:49.898950 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 10:00:49 crc kubenswrapper[4824]: E1006 10:00:49.899472 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-06 10:00:50.399451797 +0000 UTC m=+159.763874658 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 10:00:49 crc kubenswrapper[4824]: I1006 10:00:49.971142 4824 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-8z68k container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.21:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 06 10:00:49 crc kubenswrapper[4824]: I1006 10:00:49.971229 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-8z68k" podUID="af8fbf5a-f90e-4d22-ae65-05d8d8253308" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.21:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 06 10:00:50 crc kubenswrapper[4824]: I1006 10:00:50.001436 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8l8gd\" (UID: \"c2f6e84b-a28c-4ead-97f5-978b0639e5bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-8l8gd" Oct 06 10:00:50 crc kubenswrapper[4824]: E1006 10:00:50.001775 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-06 10:00:50.501760989 +0000 UTC m=+159.866183850 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8l8gd" (UID: "c2f6e84b-a28c-4ead-97f5-978b0639e5bf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 10:00:50 crc kubenswrapper[4824]: I1006 10:00:50.064120 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dtcm6"] Oct 06 10:00:50 crc kubenswrapper[4824]: I1006 10:00:50.081638 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-r52zf"] Oct 06 10:00:50 crc kubenswrapper[4824]: I1006 10:00:50.103088 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 10:00:50 crc kubenswrapper[4824]: E1006 10:00:50.103397 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-06 10:00:50.603376102 +0000 UTC m=+159.967798963 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 10:00:50 crc kubenswrapper[4824]: I1006 10:00:50.103472 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8l8gd\" (UID: \"c2f6e84b-a28c-4ead-97f5-978b0639e5bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-8l8gd" Oct 06 10:00:50 crc kubenswrapper[4824]: E1006 10:00:50.103824 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-06 10:00:50.603816913 +0000 UTC m=+159.968239774 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8l8gd" (UID: "c2f6e84b-a28c-4ead-97f5-978b0639e5bf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 10:00:50 crc kubenswrapper[4824]: I1006 10:00:50.156137 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-bgbpb"] Oct 06 10:00:50 crc kubenswrapper[4824]: I1006 10:00:50.157370 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bgbpb" Oct 06 10:00:50 crc kubenswrapper[4824]: I1006 10:00:50.160321 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Oct 06 10:00:50 crc kubenswrapper[4824]: I1006 10:00:50.175798 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bgbpb"] Oct 06 10:00:50 crc kubenswrapper[4824]: I1006 10:00:50.204906 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 10:00:50 crc kubenswrapper[4824]: E1006 10:00:50.205347 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-06 10:00:50.705326384 +0000 UTC m=+160.069749245 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 10:00:50 crc kubenswrapper[4824]: I1006 10:00:50.240094 4824 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-s8j95 container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.26:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 06 10:00:50 crc kubenswrapper[4824]: I1006 10:00:50.240568 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-s8j95" podUID="ffceea22-78f2-4032-a23a-71b4fb7ea83e" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.26:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 06 10:00:50 crc kubenswrapper[4824]: I1006 10:00:50.269239 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zlxdl"] Oct 06 10:00:50 crc kubenswrapper[4824]: I1006 10:00:50.274178 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-nk67f" event={"ID":"a1d6c12e-ff5a-4687-80ca-5699dd1bb9fe","Type":"ContainerStarted","Data":"7bab79b751bac81e8c3f7dd3255633f5a3f903dcd75906906aa26e2b4be11a45"} Oct 06 10:00:50 crc kubenswrapper[4824]: I1006 10:00:50.274235 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-nk67f" event={"ID":"a1d6c12e-ff5a-4687-80ca-5699dd1bb9fe","Type":"ContainerStarted","Data":"191a0d96cc168580324b5383675e0e557fa3ca8c33f3acae6b6641717f2f6b05"} Oct 06 10:00:50 crc kubenswrapper[4824]: I1006 10:00:50.280578 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dtcm6" event={"ID":"88764df0-c119-4e11-bd94-6818bc8ef83c","Type":"ContainerStarted","Data":"ab2c98e63ed8f8b9561d937d2f7127db8783130bfba458bb9a603e644acfa965"} Oct 06 10:00:50 crc kubenswrapper[4824]: W1006 10:00:50.281147 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod527511c8_2487_4416_a689_9ee161130297.slice/crio-e653732eeb0042c8b5b4e52c01ec69ddec3c1572905678874a2cdbd35dd0af59 WatchSource:0}: Error finding container e653732eeb0042c8b5b4e52c01ec69ddec3c1572905678874a2cdbd35dd0af59: Status 404 returned error can't find the container with id e653732eeb0042c8b5b4e52c01ec69ddec3c1572905678874a2cdbd35dd0af59 Oct 06 10:00:50 crc kubenswrapper[4824]: I1006 10:00:50.281711 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r52zf" event={"ID":"06d3772b-e467-41a1-bf9b-9d6520ab68ef","Type":"ContainerStarted","Data":"8f6bfad13da03714cd8ccfb513f43be501d6fe2cb2edbb314a684951d06f5ce2"} Oct 06 10:00:50 crc kubenswrapper[4824]: I1006 10:00:50.303600 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zzrfg" event={"ID":"2fa5ee3e-29f4-43b2-a154-3c988754897a","Type":"ContainerStarted","Data":"2b8ce8a34c80d7e39fef6ac53b0fee4fbf996d24e3c2ce747ba903ebef04d778"} Oct 06 10:00:50 crc kubenswrapper[4824]: I1006 10:00:50.305761 4824 generic.go:334] "Generic (PLEG): container finished" podID="030ddedc-a25b-47f6-88b6-c1286dc1c910" containerID="f05dfc3bc64501e41918c87609a67f03234892efdbd157bbcd331cb2f1e33136" exitCode=0 Oct 06 10:00:50 crc kubenswrapper[4824]: I1006 10:00:50.306960 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tlvv7" event={"ID":"030ddedc-a25b-47f6-88b6-c1286dc1c910","Type":"ContainerDied","Data":"f05dfc3bc64501e41918c87609a67f03234892efdbd157bbcd331cb2f1e33136"} Oct 06 10:00:50 crc kubenswrapper[4824]: I1006 10:00:50.307006 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tlvv7" event={"ID":"030ddedc-a25b-47f6-88b6-c1286dc1c910","Type":"ContainerStarted","Data":"b2cb07f58394a5ac75de56feaf694f27a1427db2fce7a24ff16b91d2de681dce"} Oct 06 10:00:50 crc kubenswrapper[4824]: I1006 10:00:50.309295 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/70324ded-62d7-47fc-a67a-110ecde1497c-catalog-content\") pod \"redhat-marketplace-bgbpb\" (UID: \"70324ded-62d7-47fc-a67a-110ecde1497c\") " pod="openshift-marketplace/redhat-marketplace-bgbpb" Oct 06 10:00:50 crc kubenswrapper[4824]: I1006 10:00:50.309345 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/70324ded-62d7-47fc-a67a-110ecde1497c-utilities\") pod \"redhat-marketplace-bgbpb\" (UID: \"70324ded-62d7-47fc-a67a-110ecde1497c\") " pod="openshift-marketplace/redhat-marketplace-bgbpb" Oct 06 10:00:50 crc kubenswrapper[4824]: I1006 10:00:50.309384 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8l8gd\" (UID: \"c2f6e84b-a28c-4ead-97f5-978b0639e5bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-8l8gd" Oct 06 10:00:50 crc kubenswrapper[4824]: I1006 10:00:50.309492 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v2pk8\" (UniqueName: \"kubernetes.io/projected/70324ded-62d7-47fc-a67a-110ecde1497c-kube-api-access-v2pk8\") pod \"redhat-marketplace-bgbpb\" (UID: \"70324ded-62d7-47fc-a67a-110ecde1497c\") " pod="openshift-marketplace/redhat-marketplace-bgbpb" Oct 06 10:00:50 crc kubenswrapper[4824]: E1006 10:00:50.309649 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-06 10:00:50.809635643 +0000 UTC m=+160.174058504 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8l8gd" (UID: "c2f6e84b-a28c-4ead-97f5-978b0639e5bf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 10:00:50 crc kubenswrapper[4824]: I1006 10:00:50.320937 4824 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 06 10:00:50 crc kubenswrapper[4824]: I1006 10:00:50.324894 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-8z68k" Oct 06 10:00:50 crc kubenswrapper[4824]: I1006 10:00:50.345021 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zzrfg" podStartSLOduration=138.344993995 podStartE2EDuration="2m18.344993995s" podCreationTimestamp="2025-10-06 09:58:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:00:50.344466612 +0000 UTC m=+159.708889473" watchObservedRunningTime="2025-10-06 10:00:50.344993995 +0000 UTC m=+159.709416856" Oct 06 10:00:50 crc kubenswrapper[4824]: I1006 10:00:50.412434 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 10:00:50 crc kubenswrapper[4824]: I1006 10:00:50.412664 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/70324ded-62d7-47fc-a67a-110ecde1497c-catalog-content\") pod \"redhat-marketplace-bgbpb\" (UID: \"70324ded-62d7-47fc-a67a-110ecde1497c\") " pod="openshift-marketplace/redhat-marketplace-bgbpb" Oct 06 10:00:50 crc kubenswrapper[4824]: I1006 10:00:50.412879 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/70324ded-62d7-47fc-a67a-110ecde1497c-utilities\") pod \"redhat-marketplace-bgbpb\" (UID: \"70324ded-62d7-47fc-a67a-110ecde1497c\") " pod="openshift-marketplace/redhat-marketplace-bgbpb" Oct 06 10:00:50 crc kubenswrapper[4824]: I1006 10:00:50.413106 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v2pk8\" (UniqueName: \"kubernetes.io/projected/70324ded-62d7-47fc-a67a-110ecde1497c-kube-api-access-v2pk8\") pod \"redhat-marketplace-bgbpb\" (UID: \"70324ded-62d7-47fc-a67a-110ecde1497c\") " pod="openshift-marketplace/redhat-marketplace-bgbpb" Oct 06 10:00:50 crc kubenswrapper[4824]: E1006 10:00:50.413775 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-06 10:00:50.913757719 +0000 UTC m=+160.278180580 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 10:00:50 crc kubenswrapper[4824]: I1006 10:00:50.429710 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/70324ded-62d7-47fc-a67a-110ecde1497c-catalog-content\") pod \"redhat-marketplace-bgbpb\" (UID: \"70324ded-62d7-47fc-a67a-110ecde1497c\") " pod="openshift-marketplace/redhat-marketplace-bgbpb" Oct 06 10:00:50 crc kubenswrapper[4824]: I1006 10:00:50.442902 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/70324ded-62d7-47fc-a67a-110ecde1497c-utilities\") pod \"redhat-marketplace-bgbpb\" (UID: \"70324ded-62d7-47fc-a67a-110ecde1497c\") " pod="openshift-marketplace/redhat-marketplace-bgbpb" Oct 06 10:00:50 crc kubenswrapper[4824]: I1006 10:00:50.515156 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8l8gd\" (UID: \"c2f6e84b-a28c-4ead-97f5-978b0639e5bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-8l8gd" Oct 06 10:00:50 crc kubenswrapper[4824]: E1006 10:00:50.519347 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-06 10:00:51.019282259 +0000 UTC m=+160.383705130 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8l8gd" (UID: "c2f6e84b-a28c-4ead-97f5-978b0639e5bf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 10:00:50 crc kubenswrapper[4824]: I1006 10:00:50.521447 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v2pk8\" (UniqueName: \"kubernetes.io/projected/70324ded-62d7-47fc-a67a-110ecde1497c-kube-api-access-v2pk8\") pod \"redhat-marketplace-bgbpb\" (UID: \"70324ded-62d7-47fc-a67a-110ecde1497c\") " pod="openshift-marketplace/redhat-marketplace-bgbpb" Oct 06 10:00:50 crc kubenswrapper[4824]: I1006 10:00:50.531504 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 06 10:00:50 crc kubenswrapper[4824]: I1006 10:00:50.533653 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 06 10:00:50 crc kubenswrapper[4824]: I1006 10:00:50.534661 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bgbpb" Oct 06 10:00:50 crc kubenswrapper[4824]: I1006 10:00:50.538429 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Oct 06 10:00:50 crc kubenswrapper[4824]: I1006 10:00:50.539591 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Oct 06 10:00:50 crc kubenswrapper[4824]: I1006 10:00:50.542184 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 06 10:00:50 crc kubenswrapper[4824]: I1006 10:00:50.557574 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-s8j95" Oct 06 10:00:50 crc kubenswrapper[4824]: I1006 10:00:50.568851 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-f7h6q"] Oct 06 10:00:50 crc kubenswrapper[4824]: I1006 10:00:50.571532 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-f7h6q" Oct 06 10:00:50 crc kubenswrapper[4824]: I1006 10:00:50.571835 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-f7h6q"] Oct 06 10:00:50 crc kubenswrapper[4824]: I1006 10:00:50.617099 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 10:00:50 crc kubenswrapper[4824]: E1006 10:00:50.617576 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-06 10:00:51.11755488 +0000 UTC m=+160.481977741 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 10:00:50 crc kubenswrapper[4824]: I1006 10:00:50.617621 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4c3973d0-2ae6-4999-bb44-bf6fb7496102-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"4c3973d0-2ae6-4999-bb44-bf6fb7496102\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 06 10:00:50 crc kubenswrapper[4824]: I1006 10:00:50.617689 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f3ef12c1-f3ad-4ead-aba0-a3901f80fc7f-utilities\") pod \"redhat-marketplace-f7h6q\" (UID: \"f3ef12c1-f3ad-4ead-aba0-a3901f80fc7f\") " pod="openshift-marketplace/redhat-marketplace-f7h6q" Oct 06 10:00:50 crc kubenswrapper[4824]: I1006 10:00:50.617716 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8l8gd\" (UID: \"c2f6e84b-a28c-4ead-97f5-978b0639e5bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-8l8gd" Oct 06 10:00:50 crc kubenswrapper[4824]: I1006 10:00:50.617737 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f3ef12c1-f3ad-4ead-aba0-a3901f80fc7f-catalog-content\") pod \"redhat-marketplace-f7h6q\" (UID: \"f3ef12c1-f3ad-4ead-aba0-a3901f80fc7f\") " pod="openshift-marketplace/redhat-marketplace-f7h6q" Oct 06 10:00:50 crc kubenswrapper[4824]: I1006 10:00:50.617756 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4c3973d0-2ae6-4999-bb44-bf6fb7496102-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"4c3973d0-2ae6-4999-bb44-bf6fb7496102\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 06 10:00:50 crc kubenswrapper[4824]: I1006 10:00:50.617784 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m4dlf\" (UniqueName: \"kubernetes.io/projected/f3ef12c1-f3ad-4ead-aba0-a3901f80fc7f-kube-api-access-m4dlf\") pod \"redhat-marketplace-f7h6q\" (UID: \"f3ef12c1-f3ad-4ead-aba0-a3901f80fc7f\") " pod="openshift-marketplace/redhat-marketplace-f7h6q" Oct 06 10:00:50 crc kubenswrapper[4824]: E1006 10:00:50.618151 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-06 10:00:51.118140305 +0000 UTC m=+160.482563166 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8l8gd" (UID: "c2f6e84b-a28c-4ead-97f5-978b0639e5bf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 10:00:50 crc kubenswrapper[4824]: I1006 10:00:50.719731 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 10:00:50 crc kubenswrapper[4824]: E1006 10:00:50.724273 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-06 10:00:51.224227089 +0000 UTC m=+160.588649950 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 10:00:50 crc kubenswrapper[4824]: I1006 10:00:50.724489 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f3ef12c1-f3ad-4ead-aba0-a3901f80fc7f-utilities\") pod \"redhat-marketplace-f7h6q\" (UID: \"f3ef12c1-f3ad-4ead-aba0-a3901f80fc7f\") " pod="openshift-marketplace/redhat-marketplace-f7h6q" Oct 06 10:00:50 crc kubenswrapper[4824]: I1006 10:00:50.724545 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8l8gd\" (UID: \"c2f6e84b-a28c-4ead-97f5-978b0639e5bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-8l8gd" Oct 06 10:00:50 crc kubenswrapper[4824]: I1006 10:00:50.724577 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f3ef12c1-f3ad-4ead-aba0-a3901f80fc7f-catalog-content\") pod \"redhat-marketplace-f7h6q\" (UID: \"f3ef12c1-f3ad-4ead-aba0-a3901f80fc7f\") " pod="openshift-marketplace/redhat-marketplace-f7h6q" Oct 06 10:00:50 crc kubenswrapper[4824]: I1006 10:00:50.724617 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4c3973d0-2ae6-4999-bb44-bf6fb7496102-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"4c3973d0-2ae6-4999-bb44-bf6fb7496102\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 06 10:00:50 crc kubenswrapper[4824]: I1006 10:00:50.724702 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m4dlf\" (UniqueName: \"kubernetes.io/projected/f3ef12c1-f3ad-4ead-aba0-a3901f80fc7f-kube-api-access-m4dlf\") pod \"redhat-marketplace-f7h6q\" (UID: \"f3ef12c1-f3ad-4ead-aba0-a3901f80fc7f\") " pod="openshift-marketplace/redhat-marketplace-f7h6q" Oct 06 10:00:50 crc kubenswrapper[4824]: I1006 10:00:50.724873 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4c3973d0-2ae6-4999-bb44-bf6fb7496102-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"4c3973d0-2ae6-4999-bb44-bf6fb7496102\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 06 10:00:50 crc kubenswrapper[4824]: E1006 10:00:50.725021 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-06 10:00:51.224961066 +0000 UTC m=+160.589383927 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8l8gd" (UID: "c2f6e84b-a28c-4ead-97f5-978b0639e5bf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 10:00:50 crc kubenswrapper[4824]: I1006 10:00:50.725134 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4c3973d0-2ae6-4999-bb44-bf6fb7496102-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"4c3973d0-2ae6-4999-bb44-bf6fb7496102\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 06 10:00:50 crc kubenswrapper[4824]: I1006 10:00:50.725331 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f3ef12c1-f3ad-4ead-aba0-a3901f80fc7f-catalog-content\") pod \"redhat-marketplace-f7h6q\" (UID: \"f3ef12c1-f3ad-4ead-aba0-a3901f80fc7f\") " pod="openshift-marketplace/redhat-marketplace-f7h6q" Oct 06 10:00:50 crc kubenswrapper[4824]: I1006 10:00:50.725333 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f3ef12c1-f3ad-4ead-aba0-a3901f80fc7f-utilities\") pod \"redhat-marketplace-f7h6q\" (UID: \"f3ef12c1-f3ad-4ead-aba0-a3901f80fc7f\") " pod="openshift-marketplace/redhat-marketplace-f7h6q" Oct 06 10:00:50 crc kubenswrapper[4824]: I1006 10:00:50.749479 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m4dlf\" (UniqueName: \"kubernetes.io/projected/f3ef12c1-f3ad-4ead-aba0-a3901f80fc7f-kube-api-access-m4dlf\") pod \"redhat-marketplace-f7h6q\" (UID: \"f3ef12c1-f3ad-4ead-aba0-a3901f80fc7f\") " pod="openshift-marketplace/redhat-marketplace-f7h6q" Oct 06 10:00:50 crc kubenswrapper[4824]: I1006 10:00:50.750199 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4c3973d0-2ae6-4999-bb44-bf6fb7496102-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"4c3973d0-2ae6-4999-bb44-bf6fb7496102\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 06 10:00:50 crc kubenswrapper[4824]: I1006 10:00:50.800235 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 06 10:00:50 crc kubenswrapper[4824]: I1006 10:00:50.808053 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-f7h6q" Oct 06 10:00:50 crc kubenswrapper[4824]: I1006 10:00:50.825953 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 10:00:50 crc kubenswrapper[4824]: E1006 10:00:50.826124 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-06 10:00:51.326091108 +0000 UTC m=+160.690513969 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 10:00:50 crc kubenswrapper[4824]: I1006 10:00:50.826610 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8l8gd\" (UID: \"c2f6e84b-a28c-4ead-97f5-978b0639e5bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-8l8gd" Oct 06 10:00:50 crc kubenswrapper[4824]: E1006 10:00:50.827028 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-06 10:00:51.327006751 +0000 UTC m=+160.691429612 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8l8gd" (UID: "c2f6e84b-a28c-4ead-97f5-978b0639e5bf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 10:00:50 crc kubenswrapper[4824]: I1006 10:00:50.837680 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29329080-9kpdv" Oct 06 10:00:50 crc kubenswrapper[4824]: I1006 10:00:50.837851 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bgbpb"] Oct 06 10:00:50 crc kubenswrapper[4824]: I1006 10:00:50.861656 4824 patch_prober.go:28] interesting pod/router-default-5444994796-7qwsv container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 06 10:00:50 crc kubenswrapper[4824]: [-]has-synced failed: reason withheld Oct 06 10:00:50 crc kubenswrapper[4824]: [+]process-running ok Oct 06 10:00:50 crc kubenswrapper[4824]: healthz check failed Oct 06 10:00:50 crc kubenswrapper[4824]: I1006 10:00:50.861923 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-7qwsv" podUID="d6e255a4-d3e7-4cef-bf40-f98aa11b3e3b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 06 10:00:50 crc kubenswrapper[4824]: I1006 10:00:50.927569 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 10:00:50 crc kubenswrapper[4824]: E1006 10:00:50.927804 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-06 10:00:51.427774073 +0000 UTC m=+160.792196934 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 10:00:50 crc kubenswrapper[4824]: I1006 10:00:50.927910 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/26536a6b-3639-46ad-b024-5e21b7118fe7-secret-volume\") pod \"26536a6b-3639-46ad-b024-5e21b7118fe7\" (UID: \"26536a6b-3639-46ad-b024-5e21b7118fe7\") " Oct 06 10:00:50 crc kubenswrapper[4824]: I1006 10:00:50.929608 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cjzws\" (UniqueName: \"kubernetes.io/projected/26536a6b-3639-46ad-b024-5e21b7118fe7-kube-api-access-cjzws\") pod \"26536a6b-3639-46ad-b024-5e21b7118fe7\" (UID: \"26536a6b-3639-46ad-b024-5e21b7118fe7\") " Oct 06 10:00:50 crc kubenswrapper[4824]: I1006 10:00:50.929756 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/26536a6b-3639-46ad-b024-5e21b7118fe7-config-volume\") pod \"26536a6b-3639-46ad-b024-5e21b7118fe7\" (UID: \"26536a6b-3639-46ad-b024-5e21b7118fe7\") " Oct 06 10:00:50 crc kubenswrapper[4824]: I1006 10:00:50.930638 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8l8gd\" (UID: \"c2f6e84b-a28c-4ead-97f5-978b0639e5bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-8l8gd" Oct 06 10:00:50 crc kubenswrapper[4824]: I1006 10:00:50.931157 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/26536a6b-3639-46ad-b024-5e21b7118fe7-config-volume" (OuterVolumeSpecName: "config-volume") pod "26536a6b-3639-46ad-b024-5e21b7118fe7" (UID: "26536a6b-3639-46ad-b024-5e21b7118fe7"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:00:50 crc kubenswrapper[4824]: I1006 10:00:50.931615 4824 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/26536a6b-3639-46ad-b024-5e21b7118fe7-config-volume\") on node \"crc\" DevicePath \"\"" Oct 06 10:00:50 crc kubenswrapper[4824]: E1006 10:00:50.932201 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-06 10:00:51.432185312 +0000 UTC m=+160.796608173 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8l8gd" (UID: "c2f6e84b-a28c-4ead-97f5-978b0639e5bf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 10:00:50 crc kubenswrapper[4824]: I1006 10:00:50.938650 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/26536a6b-3639-46ad-b024-5e21b7118fe7-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "26536a6b-3639-46ad-b024-5e21b7118fe7" (UID: "26536a6b-3639-46ad-b024-5e21b7118fe7"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:00:50 crc kubenswrapper[4824]: I1006 10:00:50.947376 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/26536a6b-3639-46ad-b024-5e21b7118fe7-kube-api-access-cjzws" (OuterVolumeSpecName: "kube-api-access-cjzws") pod "26536a6b-3639-46ad-b024-5e21b7118fe7" (UID: "26536a6b-3639-46ad-b024-5e21b7118fe7"). InnerVolumeSpecName "kube-api-access-cjzws". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:00:50 crc kubenswrapper[4824]: I1006 10:00:50.982105 4824 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Oct 06 10:00:51 crc kubenswrapper[4824]: I1006 10:00:51.033141 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 10:00:51 crc kubenswrapper[4824]: I1006 10:00:51.033561 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cjzws\" (UniqueName: \"kubernetes.io/projected/26536a6b-3639-46ad-b024-5e21b7118fe7-kube-api-access-cjzws\") on node \"crc\" DevicePath \"\"" Oct 06 10:00:51 crc kubenswrapper[4824]: I1006 10:00:51.033579 4824 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/26536a6b-3639-46ad-b024-5e21b7118fe7-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 06 10:00:51 crc kubenswrapper[4824]: E1006 10:00:51.033654 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-06 10:00:51.533628632 +0000 UTC m=+160.898051493 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 10:00:51 crc kubenswrapper[4824]: I1006 10:00:51.096733 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 06 10:00:51 crc kubenswrapper[4824]: I1006 10:00:51.136580 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8l8gd\" (UID: \"c2f6e84b-a28c-4ead-97f5-978b0639e5bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-8l8gd" Oct 06 10:00:51 crc kubenswrapper[4824]: E1006 10:00:51.136992 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-06 10:00:51.636960717 +0000 UTC m=+161.001383578 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8l8gd" (UID: "c2f6e84b-a28c-4ead-97f5-978b0639e5bf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 10:00:51 crc kubenswrapper[4824]: I1006 10:00:51.169236 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-f7h6q"] Oct 06 10:00:51 crc kubenswrapper[4824]: W1006 10:00:51.197157 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf3ef12c1_f3ad_4ead_aba0_a3901f80fc7f.slice/crio-ff7368a789c489bb4b27e52912320bd7d9f60525ae61173255c9eda795b11d75 WatchSource:0}: Error finding container ff7368a789c489bb4b27e52912320bd7d9f60525ae61173255c9eda795b11d75: Status 404 returned error can't find the container with id ff7368a789c489bb4b27e52912320bd7d9f60525ae61173255c9eda795b11d75 Oct 06 10:00:51 crc kubenswrapper[4824]: I1006 10:00:51.238420 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 10:00:51 crc kubenswrapper[4824]: E1006 10:00:51.238614 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-06 10:00:51.738581991 +0000 UTC m=+161.103004852 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 10:00:51 crc kubenswrapper[4824]: I1006 10:00:51.238745 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8l8gd\" (UID: \"c2f6e84b-a28c-4ead-97f5-978b0639e5bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-8l8gd" Oct 06 10:00:51 crc kubenswrapper[4824]: E1006 10:00:51.239175 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-06 10:00:51.739156735 +0000 UTC m=+161.103579596 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8l8gd" (UID: "c2f6e84b-a28c-4ead-97f5-978b0639e5bf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 10:00:51 crc kubenswrapper[4824]: I1006 10:00:51.327715 4824 generic.go:334] "Generic (PLEG): container finished" podID="527511c8-2487-4416-a689-9ee161130297" containerID="f4bfe51a6975baafcfaf9e53a1fd4a39a6bc9b04f30ac15711ae5aeeccfab3e7" exitCode=0 Oct 06 10:00:51 crc kubenswrapper[4824]: I1006 10:00:51.327815 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zlxdl" event={"ID":"527511c8-2487-4416-a689-9ee161130297","Type":"ContainerDied","Data":"f4bfe51a6975baafcfaf9e53a1fd4a39a6bc9b04f30ac15711ae5aeeccfab3e7"} Oct 06 10:00:51 crc kubenswrapper[4824]: I1006 10:00:51.327858 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zlxdl" event={"ID":"527511c8-2487-4416-a689-9ee161130297","Type":"ContainerStarted","Data":"e653732eeb0042c8b5b4e52c01ec69ddec3c1572905678874a2cdbd35dd0af59"} Oct 06 10:00:51 crc kubenswrapper[4824]: I1006 10:00:51.339510 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 10:00:51 crc kubenswrapper[4824]: E1006 10:00:51.341338 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-06 10:00:51.841315532 +0000 UTC m=+161.205738393 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 10:00:51 crc kubenswrapper[4824]: I1006 10:00:51.350166 4824 generic.go:334] "Generic (PLEG): container finished" podID="70324ded-62d7-47fc-a67a-110ecde1497c" containerID="a535c6214daf2963716066e8472758cee9b58fdecb11e0420742fd13986ddc34" exitCode=0 Oct 06 10:00:51 crc kubenswrapper[4824]: I1006 10:00:51.350257 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bgbpb" event={"ID":"70324ded-62d7-47fc-a67a-110ecde1497c","Type":"ContainerDied","Data":"a535c6214daf2963716066e8472758cee9b58fdecb11e0420742fd13986ddc34"} Oct 06 10:00:51 crc kubenswrapper[4824]: I1006 10:00:51.350292 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bgbpb" event={"ID":"70324ded-62d7-47fc-a67a-110ecde1497c","Type":"ContainerStarted","Data":"51ee4f7e0be054dee3cb8b7bcc662959901a5f2cf8f67435de0ef0d484a89694"} Oct 06 10:00:51 crc kubenswrapper[4824]: I1006 10:00:51.379636 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-nk67f" event={"ID":"a1d6c12e-ff5a-4687-80ca-5699dd1bb9fe","Type":"ContainerStarted","Data":"d44de9659517725fd9ccedf0296f5e2f6edd7dd00684c6cd1178dcb4b26cb413"} Oct 06 10:00:51 crc kubenswrapper[4824]: I1006 10:00:51.392173 4824 generic.go:334] "Generic (PLEG): container finished" podID="88764df0-c119-4e11-bd94-6818bc8ef83c" containerID="3426618e5cee07f375821734c951c2191a5c280c544c54ba3a509a664b4206f9" exitCode=0 Oct 06 10:00:51 crc kubenswrapper[4824]: I1006 10:00:51.392259 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dtcm6" event={"ID":"88764df0-c119-4e11-bd94-6818bc8ef83c","Type":"ContainerDied","Data":"3426618e5cee07f375821734c951c2191a5c280c544c54ba3a509a664b4206f9"} Oct 06 10:00:51 crc kubenswrapper[4824]: I1006 10:00:51.397724 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"4c3973d0-2ae6-4999-bb44-bf6fb7496102","Type":"ContainerStarted","Data":"92fad7e7da3e315a9844e88e942db7bc0b90b27e2c037bb8d6fc39e252926019"} Oct 06 10:00:51 crc kubenswrapper[4824]: I1006 10:00:51.400756 4824 generic.go:334] "Generic (PLEG): container finished" podID="06d3772b-e467-41a1-bf9b-9d6520ab68ef" containerID="447c2f2c33a8f910d33cae05ca6bed23a4f4b23c4f68908b9f2546f06c5be738" exitCode=0 Oct 06 10:00:51 crc kubenswrapper[4824]: I1006 10:00:51.400919 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r52zf" event={"ID":"06d3772b-e467-41a1-bf9b-9d6520ab68ef","Type":"ContainerDied","Data":"447c2f2c33a8f910d33cae05ca6bed23a4f4b23c4f68908b9f2546f06c5be738"} Oct 06 10:00:51 crc kubenswrapper[4824]: I1006 10:00:51.409942 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f7h6q" event={"ID":"f3ef12c1-f3ad-4ead-aba0-a3901f80fc7f","Type":"ContainerStarted","Data":"ff7368a789c489bb4b27e52912320bd7d9f60525ae61173255c9eda795b11d75"} Oct 06 10:00:51 crc kubenswrapper[4824]: I1006 10:00:51.412081 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29329080-9kpdv" event={"ID":"26536a6b-3639-46ad-b024-5e21b7118fe7","Type":"ContainerDied","Data":"b38020ad9a04fa2d400919ee281ddf257bfb04804b7a93f037bded057e20f8aa"} Oct 06 10:00:51 crc kubenswrapper[4824]: I1006 10:00:51.412115 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b38020ad9a04fa2d400919ee281ddf257bfb04804b7a93f037bded057e20f8aa" Oct 06 10:00:51 crc kubenswrapper[4824]: I1006 10:00:51.412683 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29329080-9kpdv" Oct 06 10:00:51 crc kubenswrapper[4824]: I1006 10:00:51.442543 4824 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-10-06T10:00:50.982134883Z","Handler":null,"Name":""} Oct 06 10:00:51 crc kubenswrapper[4824]: I1006 10:00:51.443318 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8l8gd\" (UID: \"c2f6e84b-a28c-4ead-97f5-978b0639e5bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-8l8gd" Oct 06 10:00:51 crc kubenswrapper[4824]: E1006 10:00:51.443683 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-06 10:00:51.943667253 +0000 UTC m=+161.308090114 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8l8gd" (UID: "c2f6e84b-a28c-4ead-97f5-978b0639e5bf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 10:00:51 crc kubenswrapper[4824]: I1006 10:00:51.452112 4824 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Oct 06 10:00:51 crc kubenswrapper[4824]: I1006 10:00:51.452146 4824 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Oct 06 10:00:51 crc kubenswrapper[4824]: I1006 10:00:51.544461 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 10:00:51 crc kubenswrapper[4824]: I1006 10:00:51.561921 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-5jqs9"] Oct 06 10:00:51 crc kubenswrapper[4824]: E1006 10:00:51.562187 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26536a6b-3639-46ad-b024-5e21b7118fe7" containerName="collect-profiles" Oct 06 10:00:51 crc kubenswrapper[4824]: I1006 10:00:51.562201 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="26536a6b-3639-46ad-b024-5e21b7118fe7" containerName="collect-profiles" Oct 06 10:00:51 crc kubenswrapper[4824]: I1006 10:00:51.562307 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="26536a6b-3639-46ad-b024-5e21b7118fe7" containerName="collect-profiles" Oct 06 10:00:51 crc kubenswrapper[4824]: I1006 10:00:51.563118 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5jqs9" Oct 06 10:00:51 crc kubenswrapper[4824]: I1006 10:00:51.570897 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Oct 06 10:00:51 crc kubenswrapper[4824]: I1006 10:00:51.580800 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 06 10:00:51 crc kubenswrapper[4824]: I1006 10:00:51.597297 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5jqs9"] Oct 06 10:00:51 crc kubenswrapper[4824]: I1006 10:00:51.642186 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-nk67f" podStartSLOduration=11.642164294 podStartE2EDuration="11.642164294s" podCreationTimestamp="2025-10-06 10:00:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:00:51.638024482 +0000 UTC m=+161.002447363" watchObservedRunningTime="2025-10-06 10:00:51.642164294 +0000 UTC m=+161.006587155" Oct 06 10:00:51 crc kubenswrapper[4824]: I1006 10:00:51.646561 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8l8gd\" (UID: \"c2f6e84b-a28c-4ead-97f5-978b0639e5bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-8l8gd" Oct 06 10:00:51 crc kubenswrapper[4824]: I1006 10:00:51.650298 4824 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 06 10:00:51 crc kubenswrapper[4824]: I1006 10:00:51.650331 4824 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8l8gd\" (UID: \"c2f6e84b-a28c-4ead-97f5-978b0639e5bf\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-8l8gd" Oct 06 10:00:51 crc kubenswrapper[4824]: I1006 10:00:51.681957 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8l8gd\" (UID: \"c2f6e84b-a28c-4ead-97f5-978b0639e5bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-8l8gd" Oct 06 10:00:51 crc kubenswrapper[4824]: I1006 10:00:51.748884 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c79214ab-ae6c-444b-bd4b-072c9a8bd37c-catalog-content\") pod \"redhat-operators-5jqs9\" (UID: \"c79214ab-ae6c-444b-bd4b-072c9a8bd37c\") " pod="openshift-marketplace/redhat-operators-5jqs9" Oct 06 10:00:51 crc kubenswrapper[4824]: I1006 10:00:51.749117 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4h5wb\" (UniqueName: \"kubernetes.io/projected/c79214ab-ae6c-444b-bd4b-072c9a8bd37c-kube-api-access-4h5wb\") pod \"redhat-operators-5jqs9\" (UID: \"c79214ab-ae6c-444b-bd4b-072c9a8bd37c\") " pod="openshift-marketplace/redhat-operators-5jqs9" Oct 06 10:00:51 crc kubenswrapper[4824]: I1006 10:00:51.749220 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c79214ab-ae6c-444b-bd4b-072c9a8bd37c-utilities\") pod \"redhat-operators-5jqs9\" (UID: \"c79214ab-ae6c-444b-bd4b-072c9a8bd37c\") " pod="openshift-marketplace/redhat-operators-5jqs9" Oct 06 10:00:51 crc kubenswrapper[4824]: I1006 10:00:51.853020 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c79214ab-ae6c-444b-bd4b-072c9a8bd37c-catalog-content\") pod \"redhat-operators-5jqs9\" (UID: \"c79214ab-ae6c-444b-bd4b-072c9a8bd37c\") " pod="openshift-marketplace/redhat-operators-5jqs9" Oct 06 10:00:51 crc kubenswrapper[4824]: I1006 10:00:51.853149 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4h5wb\" (UniqueName: \"kubernetes.io/projected/c79214ab-ae6c-444b-bd4b-072c9a8bd37c-kube-api-access-4h5wb\") pod \"redhat-operators-5jqs9\" (UID: \"c79214ab-ae6c-444b-bd4b-072c9a8bd37c\") " pod="openshift-marketplace/redhat-operators-5jqs9" Oct 06 10:00:51 crc kubenswrapper[4824]: I1006 10:00:51.853188 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c79214ab-ae6c-444b-bd4b-072c9a8bd37c-utilities\") pod \"redhat-operators-5jqs9\" (UID: \"c79214ab-ae6c-444b-bd4b-072c9a8bd37c\") " pod="openshift-marketplace/redhat-operators-5jqs9" Oct 06 10:00:51 crc kubenswrapper[4824]: I1006 10:00:51.853693 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c79214ab-ae6c-444b-bd4b-072c9a8bd37c-utilities\") pod \"redhat-operators-5jqs9\" (UID: \"c79214ab-ae6c-444b-bd4b-072c9a8bd37c\") " pod="openshift-marketplace/redhat-operators-5jqs9" Oct 06 10:00:51 crc kubenswrapper[4824]: I1006 10:00:51.853929 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c79214ab-ae6c-444b-bd4b-072c9a8bd37c-catalog-content\") pod \"redhat-operators-5jqs9\" (UID: \"c79214ab-ae6c-444b-bd4b-072c9a8bd37c\") " pod="openshift-marketplace/redhat-operators-5jqs9" Oct 06 10:00:51 crc kubenswrapper[4824]: I1006 10:00:51.858292 4824 patch_prober.go:28] interesting pod/router-default-5444994796-7qwsv container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 06 10:00:51 crc kubenswrapper[4824]: [-]has-synced failed: reason withheld Oct 06 10:00:51 crc kubenswrapper[4824]: [+]process-running ok Oct 06 10:00:51 crc kubenswrapper[4824]: healthz check failed Oct 06 10:00:51 crc kubenswrapper[4824]: I1006 10:00:51.858368 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-7qwsv" podUID="d6e255a4-d3e7-4cef-bf40-f98aa11b3e3b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 06 10:00:51 crc kubenswrapper[4824]: I1006 10:00:51.882165 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4h5wb\" (UniqueName: \"kubernetes.io/projected/c79214ab-ae6c-444b-bd4b-072c9a8bd37c-kube-api-access-4h5wb\") pod \"redhat-operators-5jqs9\" (UID: \"c79214ab-ae6c-444b-bd4b-072c9a8bd37c\") " pod="openshift-marketplace/redhat-operators-5jqs9" Oct 06 10:00:51 crc kubenswrapper[4824]: I1006 10:00:51.902244 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5jqs9" Oct 06 10:00:51 crc kubenswrapper[4824]: I1006 10:00:51.939492 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-8l8gd" Oct 06 10:00:51 crc kubenswrapper[4824]: I1006 10:00:51.948052 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-sst8j"] Oct 06 10:00:51 crc kubenswrapper[4824]: I1006 10:00:51.949193 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sst8j" Oct 06 10:00:51 crc kubenswrapper[4824]: I1006 10:00:51.960627 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-sst8j"] Oct 06 10:00:52 crc kubenswrapper[4824]: I1006 10:00:52.057750 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c470b33a-a2cd-46c8-b092-57d11b4b39c7-utilities\") pod \"redhat-operators-sst8j\" (UID: \"c470b33a-a2cd-46c8-b092-57d11b4b39c7\") " pod="openshift-marketplace/redhat-operators-sst8j" Oct 06 10:00:52 crc kubenswrapper[4824]: I1006 10:00:52.058234 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b2q9t\" (UniqueName: \"kubernetes.io/projected/c470b33a-a2cd-46c8-b092-57d11b4b39c7-kube-api-access-b2q9t\") pod \"redhat-operators-sst8j\" (UID: \"c470b33a-a2cd-46c8-b092-57d11b4b39c7\") " pod="openshift-marketplace/redhat-operators-sst8j" Oct 06 10:00:52 crc kubenswrapper[4824]: I1006 10:00:52.058269 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c470b33a-a2cd-46c8-b092-57d11b4b39c7-catalog-content\") pod \"redhat-operators-sst8j\" (UID: \"c470b33a-a2cd-46c8-b092-57d11b4b39c7\") " pod="openshift-marketplace/redhat-operators-sst8j" Oct 06 10:00:52 crc kubenswrapper[4824]: I1006 10:00:52.159744 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c470b33a-a2cd-46c8-b092-57d11b4b39c7-catalog-content\") pod \"redhat-operators-sst8j\" (UID: \"c470b33a-a2cd-46c8-b092-57d11b4b39c7\") " pod="openshift-marketplace/redhat-operators-sst8j" Oct 06 10:00:52 crc kubenswrapper[4824]: I1006 10:00:52.159823 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c470b33a-a2cd-46c8-b092-57d11b4b39c7-utilities\") pod \"redhat-operators-sst8j\" (UID: \"c470b33a-a2cd-46c8-b092-57d11b4b39c7\") " pod="openshift-marketplace/redhat-operators-sst8j" Oct 06 10:00:52 crc kubenswrapper[4824]: I1006 10:00:52.159893 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b2q9t\" (UniqueName: \"kubernetes.io/projected/c470b33a-a2cd-46c8-b092-57d11b4b39c7-kube-api-access-b2q9t\") pod \"redhat-operators-sst8j\" (UID: \"c470b33a-a2cd-46c8-b092-57d11b4b39c7\") " pod="openshift-marketplace/redhat-operators-sst8j" Oct 06 10:00:52 crc kubenswrapper[4824]: I1006 10:00:52.160883 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c470b33a-a2cd-46c8-b092-57d11b4b39c7-utilities\") pod \"redhat-operators-sst8j\" (UID: \"c470b33a-a2cd-46c8-b092-57d11b4b39c7\") " pod="openshift-marketplace/redhat-operators-sst8j" Oct 06 10:00:52 crc kubenswrapper[4824]: I1006 10:00:52.161318 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c470b33a-a2cd-46c8-b092-57d11b4b39c7-catalog-content\") pod \"redhat-operators-sst8j\" (UID: \"c470b33a-a2cd-46c8-b092-57d11b4b39c7\") " pod="openshift-marketplace/redhat-operators-sst8j" Oct 06 10:00:52 crc kubenswrapper[4824]: I1006 10:00:52.182952 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b2q9t\" (UniqueName: \"kubernetes.io/projected/c470b33a-a2cd-46c8-b092-57d11b4b39c7-kube-api-access-b2q9t\") pod \"redhat-operators-sst8j\" (UID: \"c470b33a-a2cd-46c8-b092-57d11b4b39c7\") " pod="openshift-marketplace/redhat-operators-sst8j" Oct 06 10:00:52 crc kubenswrapper[4824]: I1006 10:00:52.226096 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5jqs9"] Oct 06 10:00:52 crc kubenswrapper[4824]: I1006 10:00:52.302095 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sst8j" Oct 06 10:00:52 crc kubenswrapper[4824]: I1006 10:00:52.325758 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-8l8gd"] Oct 06 10:00:52 crc kubenswrapper[4824]: I1006 10:00:52.420480 4824 patch_prober.go:28] interesting pod/downloads-7954f5f757-s4ht9 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.7:8080/\": dial tcp 10.217.0.7:8080: connect: connection refused" start-of-body= Oct 06 10:00:52 crc kubenswrapper[4824]: I1006 10:00:52.420542 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-s4ht9" podUID="171abefb-ea19-40fa-8a22-4b38956d5e4b" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.7:8080/\": dial tcp 10.217.0.7:8080: connect: connection refused" Oct 06 10:00:52 crc kubenswrapper[4824]: I1006 10:00:52.421087 4824 patch_prober.go:28] interesting pod/downloads-7954f5f757-s4ht9 container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.7:8080/\": dial tcp 10.217.0.7:8080: connect: connection refused" start-of-body= Oct 06 10:00:52 crc kubenswrapper[4824]: I1006 10:00:52.421109 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-s4ht9" podUID="171abefb-ea19-40fa-8a22-4b38956d5e4b" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.7:8080/\": dial tcp 10.217.0.7:8080: connect: connection refused" Oct 06 10:00:52 crc kubenswrapper[4824]: I1006 10:00:52.460119 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5jqs9" event={"ID":"c79214ab-ae6c-444b-bd4b-072c9a8bd37c","Type":"ContainerStarted","Data":"a3d850b07f8ea0a2aad69710b67a0482fda2fa6524662838b582f01547234b46"} Oct 06 10:00:52 crc kubenswrapper[4824]: I1006 10:00:52.487773 4824 generic.go:334] "Generic (PLEG): container finished" podID="f3ef12c1-f3ad-4ead-aba0-a3901f80fc7f" containerID="fc6a773b82ac29844de4dfe214e868ae3d84e8217ea3e4f3f17b7ba83fe0134c" exitCode=0 Oct 06 10:00:52 crc kubenswrapper[4824]: I1006 10:00:52.487857 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f7h6q" event={"ID":"f3ef12c1-f3ad-4ead-aba0-a3901f80fc7f","Type":"ContainerDied","Data":"fc6a773b82ac29844de4dfe214e868ae3d84e8217ea3e4f3f17b7ba83fe0134c"} Oct 06 10:00:52 crc kubenswrapper[4824]: I1006 10:00:52.526462 4824 generic.go:334] "Generic (PLEG): container finished" podID="4c3973d0-2ae6-4999-bb44-bf6fb7496102" containerID="6c72661bebb39fbad2ef804fc6c1c974e48c4d1e189f5c1541f543c126ee4052" exitCode=0 Oct 06 10:00:52 crc kubenswrapper[4824]: I1006 10:00:52.526597 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"4c3973d0-2ae6-4999-bb44-bf6fb7496102","Type":"ContainerDied","Data":"6c72661bebb39fbad2ef804fc6c1c974e48c4d1e189f5c1541f543c126ee4052"} Oct 06 10:00:52 crc kubenswrapper[4824]: I1006 10:00:52.556004 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-8l8gd" event={"ID":"c2f6e84b-a28c-4ead-97f5-978b0639e5bf","Type":"ContainerStarted","Data":"bc9f21429286d2f02ba3385cf69c2113f5f39234e15a2aab5f3143545966c9ab"} Oct 06 10:00:52 crc kubenswrapper[4824]: I1006 10:00:52.782383 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-q76nt" Oct 06 10:00:52 crc kubenswrapper[4824]: I1006 10:00:52.782458 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-q76nt" Oct 06 10:00:52 crc kubenswrapper[4824]: I1006 10:00:52.787567 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-sst8j"] Oct 06 10:00:52 crc kubenswrapper[4824]: I1006 10:00:52.814257 4824 patch_prober.go:28] interesting pod/apiserver-76f77b778f-q76nt container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Oct 06 10:00:52 crc kubenswrapper[4824]: [+]log ok Oct 06 10:00:52 crc kubenswrapper[4824]: [+]etcd ok Oct 06 10:00:52 crc kubenswrapper[4824]: [+]poststarthook/start-apiserver-admission-initializer ok Oct 06 10:00:52 crc kubenswrapper[4824]: [+]poststarthook/generic-apiserver-start-informers ok Oct 06 10:00:52 crc kubenswrapper[4824]: [+]poststarthook/max-in-flight-filter ok Oct 06 10:00:52 crc kubenswrapper[4824]: [+]poststarthook/storage-object-count-tracker-hook ok Oct 06 10:00:52 crc kubenswrapper[4824]: [+]poststarthook/image.openshift.io-apiserver-caches ok Oct 06 10:00:52 crc kubenswrapper[4824]: [-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld Oct 06 10:00:52 crc kubenswrapper[4824]: [-]poststarthook/authorization.openshift.io-ensurenodebootstrap-sa failed: reason withheld Oct 06 10:00:52 crc kubenswrapper[4824]: [+]poststarthook/project.openshift.io-projectcache ok Oct 06 10:00:52 crc kubenswrapper[4824]: [+]poststarthook/project.openshift.io-projectauthorizationcache ok Oct 06 10:00:52 crc kubenswrapper[4824]: [+]poststarthook/openshift.io-startinformers ok Oct 06 10:00:52 crc kubenswrapper[4824]: [+]poststarthook/openshift.io-restmapperupdater ok Oct 06 10:00:52 crc kubenswrapper[4824]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Oct 06 10:00:52 crc kubenswrapper[4824]: livez check failed Oct 06 10:00:52 crc kubenswrapper[4824]: I1006 10:00:52.814319 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-q76nt" podUID="911b7e6b-1df0-4ef3-822e-84925bd6bbdc" containerName="openshift-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 06 10:00:52 crc kubenswrapper[4824]: W1006 10:00:52.815175 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc470b33a_a2cd_46c8_b092_57d11b4b39c7.slice/crio-cdebfbafa1fedc552ce52d0d8911fa4b901976e18169df50f8694c58d3aed9b4 WatchSource:0}: Error finding container cdebfbafa1fedc552ce52d0d8911fa4b901976e18169df50f8694c58d3aed9b4: Status 404 returned error can't find the container with id cdebfbafa1fedc552ce52d0d8911fa4b901976e18169df50f8694c58d3aed9b4 Oct 06 10:00:52 crc kubenswrapper[4824]: I1006 10:00:52.850246 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-7qwsv" Oct 06 10:00:52 crc kubenswrapper[4824]: I1006 10:00:52.855228 4824 patch_prober.go:28] interesting pod/router-default-5444994796-7qwsv container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 06 10:00:52 crc kubenswrapper[4824]: [-]has-synced failed: reason withheld Oct 06 10:00:52 crc kubenswrapper[4824]: [+]process-running ok Oct 06 10:00:52 crc kubenswrapper[4824]: healthz check failed Oct 06 10:00:52 crc kubenswrapper[4824]: I1006 10:00:52.855297 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-7qwsv" podUID="d6e255a4-d3e7-4cef-bf40-f98aa11b3e3b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 06 10:00:52 crc kubenswrapper[4824]: I1006 10:00:52.907607 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-fpxws" Oct 06 10:00:52 crc kubenswrapper[4824]: I1006 10:00:52.914493 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-fpxws" Oct 06 10:00:52 crc kubenswrapper[4824]: I1006 10:00:52.925182 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 06 10:00:52 crc kubenswrapper[4824]: I1006 10:00:52.926493 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 06 10:00:52 crc kubenswrapper[4824]: I1006 10:00:52.931373 4824 patch_prober.go:28] interesting pod/console-f9d7485db-fpxws container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.19:8443/health\": dial tcp 10.217.0.19:8443: connect: connection refused" start-of-body= Oct 06 10:00:52 crc kubenswrapper[4824]: I1006 10:00:52.931443 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-fpxws" podUID="722b7a46-52a1-49a7-a053-803334dc5964" containerName="console" probeResult="failure" output="Get \"https://10.217.0.19:8443/health\": dial tcp 10.217.0.19:8443: connect: connection refused" Oct 06 10:00:52 crc kubenswrapper[4824]: I1006 10:00:52.938935 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Oct 06 10:00:52 crc kubenswrapper[4824]: I1006 10:00:52.939400 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Oct 06 10:00:52 crc kubenswrapper[4824]: I1006 10:00:52.949131 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 06 10:00:52 crc kubenswrapper[4824]: I1006 10:00:52.982826 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/da7d73fc-5908-4185-b203-854619742975-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"da7d73fc-5908-4185-b203-854619742975\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 06 10:00:52 crc kubenswrapper[4824]: I1006 10:00:52.983012 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/da7d73fc-5908-4185-b203-854619742975-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"da7d73fc-5908-4185-b203-854619742975\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 06 10:00:53 crc kubenswrapper[4824]: I1006 10:00:53.084310 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/da7d73fc-5908-4185-b203-854619742975-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"da7d73fc-5908-4185-b203-854619742975\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 06 10:00:53 crc kubenswrapper[4824]: I1006 10:00:53.084491 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/da7d73fc-5908-4185-b203-854619742975-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"da7d73fc-5908-4185-b203-854619742975\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 06 10:00:53 crc kubenswrapper[4824]: I1006 10:00:53.084597 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/da7d73fc-5908-4185-b203-854619742975-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"da7d73fc-5908-4185-b203-854619742975\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 06 10:00:53 crc kubenswrapper[4824]: I1006 10:00:53.117967 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/da7d73fc-5908-4185-b203-854619742975-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"da7d73fc-5908-4185-b203-854619742975\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 06 10:00:53 crc kubenswrapper[4824]: I1006 10:00:53.262604 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 06 10:00:53 crc kubenswrapper[4824]: I1006 10:00:53.290607 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Oct 06 10:00:53 crc kubenswrapper[4824]: I1006 10:00:53.347871 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zzrfg" Oct 06 10:00:53 crc kubenswrapper[4824]: I1006 10:00:53.347941 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zzrfg" Oct 06 10:00:53 crc kubenswrapper[4824]: I1006 10:00:53.357803 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zzrfg" Oct 06 10:00:53 crc kubenswrapper[4824]: I1006 10:00:53.409952 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-qq4zz" Oct 06 10:00:53 crc kubenswrapper[4824]: I1006 10:00:53.569225 4824 generic.go:334] "Generic (PLEG): container finished" podID="c470b33a-a2cd-46c8-b092-57d11b4b39c7" containerID="f23e899e722c92f38881b2589bbbaa2f0b66cc00a4b51879a047488ff9dfdb22" exitCode=0 Oct 06 10:00:53 crc kubenswrapper[4824]: I1006 10:00:53.569613 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sst8j" event={"ID":"c470b33a-a2cd-46c8-b092-57d11b4b39c7","Type":"ContainerDied","Data":"f23e899e722c92f38881b2589bbbaa2f0b66cc00a4b51879a047488ff9dfdb22"} Oct 06 10:00:53 crc kubenswrapper[4824]: I1006 10:00:53.569653 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sst8j" event={"ID":"c470b33a-a2cd-46c8-b092-57d11b4b39c7","Type":"ContainerStarted","Data":"cdebfbafa1fedc552ce52d0d8911fa4b901976e18169df50f8694c58d3aed9b4"} Oct 06 10:00:53 crc kubenswrapper[4824]: I1006 10:00:53.579390 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-8l8gd" event={"ID":"c2f6e84b-a28c-4ead-97f5-978b0639e5bf","Type":"ContainerStarted","Data":"f379e7ee906ebe5d001d226f8d5629b56c3d3313684aac7bd4ea8a156413ffb5"} Oct 06 10:00:53 crc kubenswrapper[4824]: I1006 10:00:53.579541 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-8l8gd" Oct 06 10:00:53 crc kubenswrapper[4824]: I1006 10:00:53.584471 4824 generic.go:334] "Generic (PLEG): container finished" podID="c79214ab-ae6c-444b-bd4b-072c9a8bd37c" containerID="1ce8650bd8f98da7ce898f5bcd75ab5e21f63637a057f7de6cd1f89b8cb28fa4" exitCode=0 Oct 06 10:00:53 crc kubenswrapper[4824]: I1006 10:00:53.585068 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5jqs9" event={"ID":"c79214ab-ae6c-444b-bd4b-072c9a8bd37c","Type":"ContainerDied","Data":"1ce8650bd8f98da7ce898f5bcd75ab5e21f63637a057f7de6cd1f89b8cb28fa4"} Oct 06 10:00:53 crc kubenswrapper[4824]: I1006 10:00:53.603447 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zzrfg" Oct 06 10:00:53 crc kubenswrapper[4824]: I1006 10:00:53.624609 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 06 10:00:53 crc kubenswrapper[4824]: W1006 10:00:53.674309 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-podda7d73fc_5908_4185_b203_854619742975.slice/crio-d5f2c1b36877ad813f883cdb0a32e555e8bf7aedd396fe79a74b1f87053d5f83 WatchSource:0}: Error finding container d5f2c1b36877ad813f883cdb0a32e555e8bf7aedd396fe79a74b1f87053d5f83: Status 404 returned error can't find the container with id d5f2c1b36877ad813f883cdb0a32e555e8bf7aedd396fe79a74b1f87053d5f83 Oct 06 10:00:53 crc kubenswrapper[4824]: I1006 10:00:53.679460 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-8l8gd" podStartSLOduration=141.679436337 podStartE2EDuration="2m21.679436337s" podCreationTimestamp="2025-10-06 09:58:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:00:53.647881461 +0000 UTC m=+163.012304322" watchObservedRunningTime="2025-10-06 10:00:53.679436337 +0000 UTC m=+163.043859198" Oct 06 10:00:53 crc kubenswrapper[4824]: I1006 10:00:53.854858 4824 patch_prober.go:28] interesting pod/router-default-5444994796-7qwsv container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 06 10:00:53 crc kubenswrapper[4824]: [-]has-synced failed: reason withheld Oct 06 10:00:53 crc kubenswrapper[4824]: [+]process-running ok Oct 06 10:00:53 crc kubenswrapper[4824]: healthz check failed Oct 06 10:00:53 crc kubenswrapper[4824]: I1006 10:00:53.854922 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-7qwsv" podUID="d6e255a4-d3e7-4cef-bf40-f98aa11b3e3b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 06 10:00:54 crc kubenswrapper[4824]: I1006 10:00:54.039649 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 06 10:00:54 crc kubenswrapper[4824]: I1006 10:00:54.220315 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4c3973d0-2ae6-4999-bb44-bf6fb7496102-kube-api-access\") pod \"4c3973d0-2ae6-4999-bb44-bf6fb7496102\" (UID: \"4c3973d0-2ae6-4999-bb44-bf6fb7496102\") " Oct 06 10:00:54 crc kubenswrapper[4824]: I1006 10:00:54.220396 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4c3973d0-2ae6-4999-bb44-bf6fb7496102-kubelet-dir\") pod \"4c3973d0-2ae6-4999-bb44-bf6fb7496102\" (UID: \"4c3973d0-2ae6-4999-bb44-bf6fb7496102\") " Oct 06 10:00:54 crc kubenswrapper[4824]: I1006 10:00:54.220732 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7cf1d355-64cb-48a6-acfd-7d258b7afa10-metrics-certs\") pod \"network-metrics-daemon-gnw94\" (UID: \"7cf1d355-64cb-48a6-acfd-7d258b7afa10\") " pod="openshift-multus/network-metrics-daemon-gnw94" Oct 06 10:00:54 crc kubenswrapper[4824]: I1006 10:00:54.220785 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4c3973d0-2ae6-4999-bb44-bf6fb7496102-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "4c3973d0-2ae6-4999-bb44-bf6fb7496102" (UID: "4c3973d0-2ae6-4999-bb44-bf6fb7496102"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 06 10:00:54 crc kubenswrapper[4824]: I1006 10:00:54.226891 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4c3973d0-2ae6-4999-bb44-bf6fb7496102-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "4c3973d0-2ae6-4999-bb44-bf6fb7496102" (UID: "4c3973d0-2ae6-4999-bb44-bf6fb7496102"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:00:54 crc kubenswrapper[4824]: I1006 10:00:54.228943 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7cf1d355-64cb-48a6-acfd-7d258b7afa10-metrics-certs\") pod \"network-metrics-daemon-gnw94\" (UID: \"7cf1d355-64cb-48a6-acfd-7d258b7afa10\") " pod="openshift-multus/network-metrics-daemon-gnw94" Oct 06 10:00:54 crc kubenswrapper[4824]: I1006 10:00:54.296186 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gnw94" Oct 06 10:00:54 crc kubenswrapper[4824]: I1006 10:00:54.322956 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4c3973d0-2ae6-4999-bb44-bf6fb7496102-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 06 10:00:54 crc kubenswrapper[4824]: I1006 10:00:54.323007 4824 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4c3973d0-2ae6-4999-bb44-bf6fb7496102-kubelet-dir\") on node \"crc\" DevicePath \"\"" Oct 06 10:00:54 crc kubenswrapper[4824]: I1006 10:00:54.569293 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-gnw94"] Oct 06 10:00:54 crc kubenswrapper[4824]: I1006 10:00:54.599876 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 06 10:00:54 crc kubenswrapper[4824]: I1006 10:00:54.600065 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"4c3973d0-2ae6-4999-bb44-bf6fb7496102","Type":"ContainerDied","Data":"92fad7e7da3e315a9844e88e942db7bc0b90b27e2c037bb8d6fc39e252926019"} Oct 06 10:00:54 crc kubenswrapper[4824]: I1006 10:00:54.600124 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="92fad7e7da3e315a9844e88e942db7bc0b90b27e2c037bb8d6fc39e252926019" Oct 06 10:00:54 crc kubenswrapper[4824]: I1006 10:00:54.614917 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"da7d73fc-5908-4185-b203-854619742975","Type":"ContainerStarted","Data":"d5f2c1b36877ad813f883cdb0a32e555e8bf7aedd396fe79a74b1f87053d5f83"} Oct 06 10:00:54 crc kubenswrapper[4824]: I1006 10:00:54.854568 4824 patch_prober.go:28] interesting pod/router-default-5444994796-7qwsv container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 06 10:00:54 crc kubenswrapper[4824]: [-]has-synced failed: reason withheld Oct 06 10:00:54 crc kubenswrapper[4824]: [+]process-running ok Oct 06 10:00:54 crc kubenswrapper[4824]: healthz check failed Oct 06 10:00:54 crc kubenswrapper[4824]: I1006 10:00:54.854662 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-7qwsv" podUID="d6e255a4-d3e7-4cef-bf40-f98aa11b3e3b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 06 10:00:55 crc kubenswrapper[4824]: I1006 10:00:55.163742 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-dr4mr" Oct 06 10:00:55 crc kubenswrapper[4824]: I1006 10:00:55.637773 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-gnw94" event={"ID":"7cf1d355-64cb-48a6-acfd-7d258b7afa10","Type":"ContainerStarted","Data":"cbe7fd7dbe06b77e3b2f645c413732f743af45d97d2a1c3066dcfc98471c6189"} Oct 06 10:00:55 crc kubenswrapper[4824]: I1006 10:00:55.638195 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-gnw94" event={"ID":"7cf1d355-64cb-48a6-acfd-7d258b7afa10","Type":"ContainerStarted","Data":"06ac36558057b28dec5831671cc0824ab525d85f228e5447fad4f64022d4a43d"} Oct 06 10:00:55 crc kubenswrapper[4824]: I1006 10:00:55.640936 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"da7d73fc-5908-4185-b203-854619742975","Type":"ContainerStarted","Data":"88d1f51df9f7831cb7fbb3630e8797d8f4dc56604051efcd5c3eb7e6b406f51f"} Oct 06 10:00:55 crc kubenswrapper[4824]: I1006 10:00:55.657345 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=3.657322608 podStartE2EDuration="3.657322608s" podCreationTimestamp="2025-10-06 10:00:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:00:55.655649737 +0000 UTC m=+165.020072598" watchObservedRunningTime="2025-10-06 10:00:55.657322608 +0000 UTC m=+165.021745469" Oct 06 10:00:55 crc kubenswrapper[4824]: I1006 10:00:55.853670 4824 patch_prober.go:28] interesting pod/router-default-5444994796-7qwsv container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 06 10:00:55 crc kubenswrapper[4824]: [-]has-synced failed: reason withheld Oct 06 10:00:55 crc kubenswrapper[4824]: [+]process-running ok Oct 06 10:00:55 crc kubenswrapper[4824]: healthz check failed Oct 06 10:00:55 crc kubenswrapper[4824]: I1006 10:00:55.853934 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-7qwsv" podUID="d6e255a4-d3e7-4cef-bf40-f98aa11b3e3b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 06 10:00:56 crc kubenswrapper[4824]: I1006 10:00:56.657896 4824 generic.go:334] "Generic (PLEG): container finished" podID="da7d73fc-5908-4185-b203-854619742975" containerID="88d1f51df9f7831cb7fbb3630e8797d8f4dc56604051efcd5c3eb7e6b406f51f" exitCode=0 Oct 06 10:00:56 crc kubenswrapper[4824]: I1006 10:00:56.658045 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"da7d73fc-5908-4185-b203-854619742975","Type":"ContainerDied","Data":"88d1f51df9f7831cb7fbb3630e8797d8f4dc56604051efcd5c3eb7e6b406f51f"} Oct 06 10:00:56 crc kubenswrapper[4824]: I1006 10:00:56.661266 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-gnw94" event={"ID":"7cf1d355-64cb-48a6-acfd-7d258b7afa10","Type":"ContainerStarted","Data":"3763285e3fe363373352ae55f3683902ca2274292d7f270218afbb002bcf722c"} Oct 06 10:00:56 crc kubenswrapper[4824]: I1006 10:00:56.703497 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-gnw94" podStartSLOduration=145.703466802 podStartE2EDuration="2m25.703466802s" podCreationTimestamp="2025-10-06 09:58:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:00:56.699421603 +0000 UTC m=+166.063844464" watchObservedRunningTime="2025-10-06 10:00:56.703466802 +0000 UTC m=+166.067889673" Oct 06 10:00:56 crc kubenswrapper[4824]: I1006 10:00:56.852898 4824 patch_prober.go:28] interesting pod/router-default-5444994796-7qwsv container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 06 10:00:56 crc kubenswrapper[4824]: [-]has-synced failed: reason withheld Oct 06 10:00:56 crc kubenswrapper[4824]: [+]process-running ok Oct 06 10:00:56 crc kubenswrapper[4824]: healthz check failed Oct 06 10:00:56 crc kubenswrapper[4824]: I1006 10:00:56.853045 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-7qwsv" podUID="d6e255a4-d3e7-4cef-bf40-f98aa11b3e3b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 06 10:00:57 crc kubenswrapper[4824]: I1006 10:00:57.788464 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-q76nt" Oct 06 10:00:57 crc kubenswrapper[4824]: I1006 10:00:57.793814 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-q76nt" Oct 06 10:00:57 crc kubenswrapper[4824]: I1006 10:00:57.870909 4824 patch_prober.go:28] interesting pod/router-default-5444994796-7qwsv container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 06 10:00:57 crc kubenswrapper[4824]: [-]has-synced failed: reason withheld Oct 06 10:00:57 crc kubenswrapper[4824]: [+]process-running ok Oct 06 10:00:57 crc kubenswrapper[4824]: healthz check failed Oct 06 10:00:57 crc kubenswrapper[4824]: I1006 10:00:57.871474 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-7qwsv" podUID="d6e255a4-d3e7-4cef-bf40-f98aa11b3e3b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 06 10:00:58 crc kubenswrapper[4824]: I1006 10:00:58.853999 4824 patch_prober.go:28] interesting pod/router-default-5444994796-7qwsv container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 06 10:00:58 crc kubenswrapper[4824]: [-]has-synced failed: reason withheld Oct 06 10:00:58 crc kubenswrapper[4824]: [+]process-running ok Oct 06 10:00:58 crc kubenswrapper[4824]: healthz check failed Oct 06 10:00:58 crc kubenswrapper[4824]: I1006 10:00:58.854388 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-7qwsv" podUID="d6e255a4-d3e7-4cef-bf40-f98aa11b3e3b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 06 10:00:59 crc kubenswrapper[4824]: I1006 10:00:59.854364 4824 patch_prober.go:28] interesting pod/router-default-5444994796-7qwsv container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 06 10:00:59 crc kubenswrapper[4824]: [-]has-synced failed: reason withheld Oct 06 10:00:59 crc kubenswrapper[4824]: [+]process-running ok Oct 06 10:00:59 crc kubenswrapper[4824]: healthz check failed Oct 06 10:00:59 crc kubenswrapper[4824]: I1006 10:00:59.854451 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-7qwsv" podUID="d6e255a4-d3e7-4cef-bf40-f98aa11b3e3b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 06 10:01:00 crc kubenswrapper[4824]: I1006 10:01:00.853352 4824 patch_prober.go:28] interesting pod/router-default-5444994796-7qwsv container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 06 10:01:00 crc kubenswrapper[4824]: [-]has-synced failed: reason withheld Oct 06 10:01:00 crc kubenswrapper[4824]: [+]process-running ok Oct 06 10:01:00 crc kubenswrapper[4824]: healthz check failed Oct 06 10:01:00 crc kubenswrapper[4824]: I1006 10:01:00.853426 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-7qwsv" podUID="d6e255a4-d3e7-4cef-bf40-f98aa11b3e3b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 06 10:01:01 crc kubenswrapper[4824]: I1006 10:01:01.852628 4824 patch_prober.go:28] interesting pod/router-default-5444994796-7qwsv container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 06 10:01:01 crc kubenswrapper[4824]: [-]has-synced failed: reason withheld Oct 06 10:01:01 crc kubenswrapper[4824]: [+]process-running ok Oct 06 10:01:01 crc kubenswrapper[4824]: healthz check failed Oct 06 10:01:01 crc kubenswrapper[4824]: I1006 10:01:01.852695 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-7qwsv" podUID="d6e255a4-d3e7-4cef-bf40-f98aa11b3e3b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 06 10:01:02 crc kubenswrapper[4824]: I1006 10:01:02.421229 4824 patch_prober.go:28] interesting pod/downloads-7954f5f757-s4ht9 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.7:8080/\": dial tcp 10.217.0.7:8080: connect: connection refused" start-of-body= Oct 06 10:01:02 crc kubenswrapper[4824]: I1006 10:01:02.421302 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-s4ht9" podUID="171abefb-ea19-40fa-8a22-4b38956d5e4b" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.7:8080/\": dial tcp 10.217.0.7:8080: connect: connection refused" Oct 06 10:01:02 crc kubenswrapper[4824]: I1006 10:01:02.421329 4824 patch_prober.go:28] interesting pod/downloads-7954f5f757-s4ht9 container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.7:8080/\": dial tcp 10.217.0.7:8080: connect: connection refused" start-of-body= Oct 06 10:01:02 crc kubenswrapper[4824]: I1006 10:01:02.421403 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-s4ht9" podUID="171abefb-ea19-40fa-8a22-4b38956d5e4b" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.7:8080/\": dial tcp 10.217.0.7:8080: connect: connection refused" Oct 06 10:01:02 crc kubenswrapper[4824]: I1006 10:01:02.857114 4824 patch_prober.go:28] interesting pod/router-default-5444994796-7qwsv container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 06 10:01:02 crc kubenswrapper[4824]: [-]has-synced failed: reason withheld Oct 06 10:01:02 crc kubenswrapper[4824]: [+]process-running ok Oct 06 10:01:02 crc kubenswrapper[4824]: healthz check failed Oct 06 10:01:02 crc kubenswrapper[4824]: I1006 10:01:02.857583 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-7qwsv" podUID="d6e255a4-d3e7-4cef-bf40-f98aa11b3e3b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 06 10:01:02 crc kubenswrapper[4824]: I1006 10:01:02.908223 4824 patch_prober.go:28] interesting pod/console-f9d7485db-fpxws container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.19:8443/health\": dial tcp 10.217.0.19:8443: connect: connection refused" start-of-body= Oct 06 10:01:02 crc kubenswrapper[4824]: I1006 10:01:02.908295 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-fpxws" podUID="722b7a46-52a1-49a7-a053-803334dc5964" containerName="console" probeResult="failure" output="Get \"https://10.217.0.19:8443/health\": dial tcp 10.217.0.19:8443: connect: connection refused" Oct 06 10:01:03 crc kubenswrapper[4824]: I1006 10:01:03.854455 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-7qwsv" Oct 06 10:01:03 crc kubenswrapper[4824]: I1006 10:01:03.860326 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-7qwsv" Oct 06 10:01:04 crc kubenswrapper[4824]: I1006 10:01:04.543488 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 06 10:01:04 crc kubenswrapper[4824]: I1006 10:01:04.701165 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/da7d73fc-5908-4185-b203-854619742975-kubelet-dir\") pod \"da7d73fc-5908-4185-b203-854619742975\" (UID: \"da7d73fc-5908-4185-b203-854619742975\") " Oct 06 10:01:04 crc kubenswrapper[4824]: I1006 10:01:04.701254 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/da7d73fc-5908-4185-b203-854619742975-kube-api-access\") pod \"da7d73fc-5908-4185-b203-854619742975\" (UID: \"da7d73fc-5908-4185-b203-854619742975\") " Oct 06 10:01:04 crc kubenswrapper[4824]: I1006 10:01:04.701302 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/da7d73fc-5908-4185-b203-854619742975-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "da7d73fc-5908-4185-b203-854619742975" (UID: "da7d73fc-5908-4185-b203-854619742975"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 06 10:01:04 crc kubenswrapper[4824]: I1006 10:01:04.701853 4824 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/da7d73fc-5908-4185-b203-854619742975-kubelet-dir\") on node \"crc\" DevicePath \"\"" Oct 06 10:01:04 crc kubenswrapper[4824]: I1006 10:01:04.709334 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/da7d73fc-5908-4185-b203-854619742975-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "da7d73fc-5908-4185-b203-854619742975" (UID: "da7d73fc-5908-4185-b203-854619742975"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:01:04 crc kubenswrapper[4824]: I1006 10:01:04.746139 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"da7d73fc-5908-4185-b203-854619742975","Type":"ContainerDied","Data":"d5f2c1b36877ad813f883cdb0a32e555e8bf7aedd396fe79a74b1f87053d5f83"} Oct 06 10:01:04 crc kubenswrapper[4824]: I1006 10:01:04.746196 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d5f2c1b36877ad813f883cdb0a32e555e8bf7aedd396fe79a74b1f87053d5f83" Oct 06 10:01:04 crc kubenswrapper[4824]: I1006 10:01:04.746168 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 06 10:01:04 crc kubenswrapper[4824]: I1006 10:01:04.803030 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/da7d73fc-5908-4185-b203-854619742975-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 06 10:01:11 crc kubenswrapper[4824]: I1006 10:01:11.949332 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-8l8gd" Oct 06 10:01:12 crc kubenswrapper[4824]: I1006 10:01:12.435632 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-s4ht9" Oct 06 10:01:12 crc kubenswrapper[4824]: I1006 10:01:12.912737 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-fpxws" Oct 06 10:01:12 crc kubenswrapper[4824]: I1006 10:01:12.917496 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-fpxws" Oct 06 10:01:13 crc kubenswrapper[4824]: I1006 10:01:13.914905 4824 patch_prober.go:28] interesting pod/machine-config-daemon-khgzw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 10:01:13 crc kubenswrapper[4824]: I1006 10:01:13.914998 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 10:01:18 crc kubenswrapper[4824]: I1006 10:01:18.610900 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 10:01:21 crc kubenswrapper[4824]: E1006 10:01:21.462748 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Oct 06 10:01:21 crc kubenswrapper[4824]: E1006 10:01:21.463780 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-975sb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-tlvv7_openshift-marketplace(030ddedc-a25b-47f6-88b6-c1286dc1c910): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 06 10:01:21 crc kubenswrapper[4824]: E1006 10:01:21.464999 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-tlvv7" podUID="030ddedc-a25b-47f6-88b6-c1286dc1c910" Oct 06 10:01:22 crc kubenswrapper[4824]: E1006 10:01:22.943862 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-tlvv7" podUID="030ddedc-a25b-47f6-88b6-c1286dc1c910" Oct 06 10:01:23 crc kubenswrapper[4824]: E1006 10:01:23.050959 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Oct 06 10:01:23 crc kubenswrapper[4824]: E1006 10:01:23.051215 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-fhs24,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-r52zf_openshift-marketplace(06d3772b-e467-41a1-bf9b-9d6520ab68ef): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 06 10:01:23 crc kubenswrapper[4824]: E1006 10:01:23.052434 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-r52zf" podUID="06d3772b-e467-41a1-bf9b-9d6520ab68ef" Oct 06 10:01:23 crc kubenswrapper[4824]: E1006 10:01:23.069488 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Oct 06 10:01:23 crc kubenswrapper[4824]: E1006 10:01:23.069686 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-vc8nz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-zlxdl_openshift-marketplace(527511c8-2487-4416-a689-9ee161130297): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 06 10:01:23 crc kubenswrapper[4824]: E1006 10:01:23.071121 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-zlxdl" podUID="527511c8-2487-4416-a689-9ee161130297" Oct 06 10:01:23 crc kubenswrapper[4824]: E1006 10:01:23.079673 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Oct 06 10:01:23 crc kubenswrapper[4824]: E1006 10:01:23.079914 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-dqjms,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-dtcm6_openshift-marketplace(88764df0-c119-4e11-bd94-6818bc8ef83c): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 06 10:01:23 crc kubenswrapper[4824]: E1006 10:01:23.081350 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-dtcm6" podUID="88764df0-c119-4e11-bd94-6818bc8ef83c" Oct 06 10:01:23 crc kubenswrapper[4824]: I1006 10:01:23.085671 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-px5fp" Oct 06 10:01:26 crc kubenswrapper[4824]: E1006 10:01:26.787186 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-r52zf" podUID="06d3772b-e467-41a1-bf9b-9d6520ab68ef" Oct 06 10:01:26 crc kubenswrapper[4824]: E1006 10:01:26.787445 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-dtcm6" podUID="88764df0-c119-4e11-bd94-6818bc8ef83c" Oct 06 10:01:26 crc kubenswrapper[4824]: E1006 10:01:26.787518 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-zlxdl" podUID="527511c8-2487-4416-a689-9ee161130297" Oct 06 10:01:27 crc kubenswrapper[4824]: E1006 10:01:27.606642 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Oct 06 10:01:27 crc kubenswrapper[4824]: E1006 10:01:27.606840 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-v2pk8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-bgbpb_openshift-marketplace(70324ded-62d7-47fc-a67a-110ecde1497c): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 06 10:01:27 crc kubenswrapper[4824]: E1006 10:01:27.608281 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-bgbpb" podUID="70324ded-62d7-47fc-a67a-110ecde1497c" Oct 06 10:01:27 crc kubenswrapper[4824]: E1006 10:01:27.686357 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Oct 06 10:01:27 crc kubenswrapper[4824]: E1006 10:01:27.686795 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-m4dlf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-f7h6q_openshift-marketplace(f3ef12c1-f3ad-4ead-aba0-a3901f80fc7f): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 06 10:01:27 crc kubenswrapper[4824]: E1006 10:01:27.688290 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-f7h6q" podUID="f3ef12c1-f3ad-4ead-aba0-a3901f80fc7f" Oct 06 10:01:27 crc kubenswrapper[4824]: I1006 10:01:27.914098 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5jqs9" event={"ID":"c79214ab-ae6c-444b-bd4b-072c9a8bd37c","Type":"ContainerStarted","Data":"38f09ee01c275fb5b6c01f374e7ddad24fe9ee9659e3cd9f0b5373327c8892fc"} Oct 06 10:01:27 crc kubenswrapper[4824]: I1006 10:01:27.918494 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sst8j" event={"ID":"c470b33a-a2cd-46c8-b092-57d11b4b39c7","Type":"ContainerStarted","Data":"0505948a8d9c2922a4363c5632265c4db8f697a49af7a07e967e23cbbc9d7450"} Oct 06 10:01:27 crc kubenswrapper[4824]: E1006 10:01:27.921070 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-bgbpb" podUID="70324ded-62d7-47fc-a67a-110ecde1497c" Oct 06 10:01:27 crc kubenswrapper[4824]: E1006 10:01:27.921120 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-f7h6q" podUID="f3ef12c1-f3ad-4ead-aba0-a3901f80fc7f" Oct 06 10:01:28 crc kubenswrapper[4824]: I1006 10:01:28.973395 4824 generic.go:334] "Generic (PLEG): container finished" podID="c79214ab-ae6c-444b-bd4b-072c9a8bd37c" containerID="38f09ee01c275fb5b6c01f374e7ddad24fe9ee9659e3cd9f0b5373327c8892fc" exitCode=0 Oct 06 10:01:28 crc kubenswrapper[4824]: I1006 10:01:28.973504 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5jqs9" event={"ID":"c79214ab-ae6c-444b-bd4b-072c9a8bd37c","Type":"ContainerDied","Data":"38f09ee01c275fb5b6c01f374e7ddad24fe9ee9659e3cd9f0b5373327c8892fc"} Oct 06 10:01:28 crc kubenswrapper[4824]: I1006 10:01:28.976293 4824 generic.go:334] "Generic (PLEG): container finished" podID="c470b33a-a2cd-46c8-b092-57d11b4b39c7" containerID="0505948a8d9c2922a4363c5632265c4db8f697a49af7a07e967e23cbbc9d7450" exitCode=0 Oct 06 10:01:28 crc kubenswrapper[4824]: I1006 10:01:28.976320 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sst8j" event={"ID":"c470b33a-a2cd-46c8-b092-57d11b4b39c7","Type":"ContainerDied","Data":"0505948a8d9c2922a4363c5632265c4db8f697a49af7a07e967e23cbbc9d7450"} Oct 06 10:01:29 crc kubenswrapper[4824]: I1006 10:01:29.987932 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sst8j" event={"ID":"c470b33a-a2cd-46c8-b092-57d11b4b39c7","Type":"ContainerStarted","Data":"2342c530649742f64a0fffa453a8e5d0d67b15d08fc3539cba60e6fe27138d7e"} Oct 06 10:01:29 crc kubenswrapper[4824]: I1006 10:01:29.995825 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5jqs9" event={"ID":"c79214ab-ae6c-444b-bd4b-072c9a8bd37c","Type":"ContainerStarted","Data":"310ea3555fe50ef7415a359fc4b16c58fc944b5e04f0a977f12d6807fb9eb330"} Oct 06 10:01:30 crc kubenswrapper[4824]: I1006 10:01:30.022584 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-sst8j" podStartSLOduration=3.075973996 podStartE2EDuration="39.022543401s" podCreationTimestamp="2025-10-06 10:00:51 +0000 UTC" firstStartedPulling="2025-10-06 10:00:53.57319031 +0000 UTC m=+162.937613171" lastFinishedPulling="2025-10-06 10:01:29.519759725 +0000 UTC m=+198.884182576" observedRunningTime="2025-10-06 10:01:30.017476537 +0000 UTC m=+199.381899428" watchObservedRunningTime="2025-10-06 10:01:30.022543401 +0000 UTC m=+199.386966302" Oct 06 10:01:30 crc kubenswrapper[4824]: I1006 10:01:30.049350 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-5jqs9" podStartSLOduration=3.194753527 podStartE2EDuration="39.049321279s" podCreationTimestamp="2025-10-06 10:00:51 +0000 UTC" firstStartedPulling="2025-10-06 10:00:53.595145771 +0000 UTC m=+162.959568632" lastFinishedPulling="2025-10-06 10:01:29.449713463 +0000 UTC m=+198.814136384" observedRunningTime="2025-10-06 10:01:30.039458627 +0000 UTC m=+199.403881518" watchObservedRunningTime="2025-10-06 10:01:30.049321279 +0000 UTC m=+199.413744180" Oct 06 10:01:31 crc kubenswrapper[4824]: I1006 10:01:31.903717 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-5jqs9" Oct 06 10:01:31 crc kubenswrapper[4824]: I1006 10:01:31.903795 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-5jqs9" Oct 06 10:01:32 crc kubenswrapper[4824]: I1006 10:01:32.303134 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-sst8j" Oct 06 10:01:32 crc kubenswrapper[4824]: I1006 10:01:32.303617 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-sst8j" Oct 06 10:01:33 crc kubenswrapper[4824]: I1006 10:01:33.390623 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-5jqs9" podUID="c79214ab-ae6c-444b-bd4b-072c9a8bd37c" containerName="registry-server" probeResult="failure" output=< Oct 06 10:01:33 crc kubenswrapper[4824]: timeout: failed to connect service ":50051" within 1s Oct 06 10:01:33 crc kubenswrapper[4824]: > Oct 06 10:01:33 crc kubenswrapper[4824]: I1006 10:01:33.392151 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-sst8j" podUID="c470b33a-a2cd-46c8-b092-57d11b4b39c7" containerName="registry-server" probeResult="failure" output=< Oct 06 10:01:33 crc kubenswrapper[4824]: timeout: failed to connect service ":50051" within 1s Oct 06 10:01:33 crc kubenswrapper[4824]: > Oct 06 10:01:42 crc kubenswrapper[4824]: I1006 10:01:42.109599 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-5jqs9" Oct 06 10:01:42 crc kubenswrapper[4824]: I1006 10:01:42.175970 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-5jqs9" Oct 06 10:01:42 crc kubenswrapper[4824]: I1006 10:01:42.375691 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-sst8j" Oct 06 10:01:42 crc kubenswrapper[4824]: I1006 10:01:42.417719 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-sst8j" Oct 06 10:01:43 crc kubenswrapper[4824]: I1006 10:01:43.097862 4824 generic.go:334] "Generic (PLEG): container finished" podID="030ddedc-a25b-47f6-88b6-c1286dc1c910" containerID="4f1cda4ec404abec1f0abfc1c077db1856f8bcf4a45c9403fb951fb24e316277" exitCode=0 Oct 06 10:01:43 crc kubenswrapper[4824]: I1006 10:01:43.097938 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tlvv7" event={"ID":"030ddedc-a25b-47f6-88b6-c1286dc1c910","Type":"ContainerDied","Data":"4f1cda4ec404abec1f0abfc1c077db1856f8bcf4a45c9403fb951fb24e316277"} Oct 06 10:01:43 crc kubenswrapper[4824]: I1006 10:01:43.916033 4824 patch_prober.go:28] interesting pod/machine-config-daemon-khgzw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 10:01:43 crc kubenswrapper[4824]: I1006 10:01:43.916768 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 10:01:43 crc kubenswrapper[4824]: I1006 10:01:43.916844 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" Oct 06 10:01:43 crc kubenswrapper[4824]: I1006 10:01:43.917898 4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5512be36013baeca611ee92645a1d339da237b70d63aac43334beafe24513d44"} pod="openshift-machine-config-operator/machine-config-daemon-khgzw" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 06 10:01:43 crc kubenswrapper[4824]: I1006 10:01:43.918061 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" containerName="machine-config-daemon" containerID="cri-o://5512be36013baeca611ee92645a1d339da237b70d63aac43334beafe24513d44" gracePeriod=600 Oct 06 10:01:43 crc kubenswrapper[4824]: I1006 10:01:43.935655 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-sst8j"] Oct 06 10:01:44 crc kubenswrapper[4824]: I1006 10:01:44.111164 4824 generic.go:334] "Generic (PLEG): container finished" podID="88764df0-c119-4e11-bd94-6818bc8ef83c" containerID="0a30ca0f05eecfc1279b5e26a03155ef8d797e21fdd83d42c35aa685a8bb70ae" exitCode=0 Oct 06 10:01:44 crc kubenswrapper[4824]: I1006 10:01:44.111410 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dtcm6" event={"ID":"88764df0-c119-4e11-bd94-6818bc8ef83c","Type":"ContainerDied","Data":"0a30ca0f05eecfc1279b5e26a03155ef8d797e21fdd83d42c35aa685a8bb70ae"} Oct 06 10:01:44 crc kubenswrapper[4824]: I1006 10:01:44.115321 4824 generic.go:334] "Generic (PLEG): container finished" podID="06d3772b-e467-41a1-bf9b-9d6520ab68ef" containerID="dd59416586d18324918270d5729aa830d2868829f78adaf9e692b472eaf3f67d" exitCode=0 Oct 06 10:01:44 crc kubenswrapper[4824]: I1006 10:01:44.115368 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r52zf" event={"ID":"06d3772b-e467-41a1-bf9b-9d6520ab68ef","Type":"ContainerDied","Data":"dd59416586d18324918270d5729aa830d2868829f78adaf9e692b472eaf3f67d"} Oct 06 10:01:44 crc kubenswrapper[4824]: I1006 10:01:44.135678 4824 generic.go:334] "Generic (PLEG): container finished" podID="f3ef12c1-f3ad-4ead-aba0-a3901f80fc7f" containerID="6082a44ad6b26d8d1619132966e6e4b385079e36321830323e9b482f287f9b7e" exitCode=0 Oct 06 10:01:44 crc kubenswrapper[4824]: I1006 10:01:44.135770 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f7h6q" event={"ID":"f3ef12c1-f3ad-4ead-aba0-a3901f80fc7f","Type":"ContainerDied","Data":"6082a44ad6b26d8d1619132966e6e4b385079e36321830323e9b482f287f9b7e"} Oct 06 10:01:44 crc kubenswrapper[4824]: I1006 10:01:44.146080 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tlvv7" event={"ID":"030ddedc-a25b-47f6-88b6-c1286dc1c910","Type":"ContainerStarted","Data":"853c3143b0b0b58e6bebb8466a6f31993342218020b663e0e381e288cc558940"} Oct 06 10:01:44 crc kubenswrapper[4824]: I1006 10:01:44.158714 4824 generic.go:334] "Generic (PLEG): container finished" podID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" containerID="5512be36013baeca611ee92645a1d339da237b70d63aac43334beafe24513d44" exitCode=0 Oct 06 10:01:44 crc kubenswrapper[4824]: I1006 10:01:44.159251 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" event={"ID":"e1140c8c-93f1-4fce-8c06-a69561ff8a6d","Type":"ContainerDied","Data":"5512be36013baeca611ee92645a1d339da237b70d63aac43334beafe24513d44"} Oct 06 10:01:44 crc kubenswrapper[4824]: I1006 10:01:44.161450 4824 generic.go:334] "Generic (PLEG): container finished" podID="527511c8-2487-4416-a689-9ee161130297" containerID="5b50704f318215ea643719849c6a7433e5ba1be0efabfb94c3d69e6d92741990" exitCode=0 Oct 06 10:01:44 crc kubenswrapper[4824]: I1006 10:01:44.161497 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zlxdl" event={"ID":"527511c8-2487-4416-a689-9ee161130297","Type":"ContainerDied","Data":"5b50704f318215ea643719849c6a7433e5ba1be0efabfb94c3d69e6d92741990"} Oct 06 10:01:44 crc kubenswrapper[4824]: I1006 10:01:44.163652 4824 generic.go:334] "Generic (PLEG): container finished" podID="70324ded-62d7-47fc-a67a-110ecde1497c" containerID="f5d533e67f48f603c4640f3ea8a4d5b9e0e4fa5f3750daf25c5acd92a566d758" exitCode=0 Oct 06 10:01:44 crc kubenswrapper[4824]: I1006 10:01:44.163735 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bgbpb" event={"ID":"70324ded-62d7-47fc-a67a-110ecde1497c","Type":"ContainerDied","Data":"f5d533e67f48f603c4640f3ea8a4d5b9e0e4fa5f3750daf25c5acd92a566d758"} Oct 06 10:01:44 crc kubenswrapper[4824]: I1006 10:01:44.163890 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-sst8j" podUID="c470b33a-a2cd-46c8-b092-57d11b4b39c7" containerName="registry-server" containerID="cri-o://2342c530649742f64a0fffa453a8e5d0d67b15d08fc3539cba60e6fe27138d7e" gracePeriod=2 Oct 06 10:01:44 crc kubenswrapper[4824]: I1006 10:01:44.197786 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-tlvv7" podStartSLOduration=2.980230706 podStartE2EDuration="56.197764938s" podCreationTimestamp="2025-10-06 10:00:48 +0000 UTC" firstStartedPulling="2025-10-06 10:00:50.320559923 +0000 UTC m=+159.684982784" lastFinishedPulling="2025-10-06 10:01:43.538094155 +0000 UTC m=+212.902517016" observedRunningTime="2025-10-06 10:01:44.188242794 +0000 UTC m=+213.552665655" watchObservedRunningTime="2025-10-06 10:01:44.197764938 +0000 UTC m=+213.562187799" Oct 06 10:01:44 crc kubenswrapper[4824]: I1006 10:01:44.618453 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sst8j" Oct 06 10:01:44 crc kubenswrapper[4824]: I1006 10:01:44.702758 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b2q9t\" (UniqueName: \"kubernetes.io/projected/c470b33a-a2cd-46c8-b092-57d11b4b39c7-kube-api-access-b2q9t\") pod \"c470b33a-a2cd-46c8-b092-57d11b4b39c7\" (UID: \"c470b33a-a2cd-46c8-b092-57d11b4b39c7\") " Oct 06 10:01:44 crc kubenswrapper[4824]: I1006 10:01:44.703199 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c470b33a-a2cd-46c8-b092-57d11b4b39c7-utilities\") pod \"c470b33a-a2cd-46c8-b092-57d11b4b39c7\" (UID: \"c470b33a-a2cd-46c8-b092-57d11b4b39c7\") " Oct 06 10:01:44 crc kubenswrapper[4824]: I1006 10:01:44.703332 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c470b33a-a2cd-46c8-b092-57d11b4b39c7-catalog-content\") pod \"c470b33a-a2cd-46c8-b092-57d11b4b39c7\" (UID: \"c470b33a-a2cd-46c8-b092-57d11b4b39c7\") " Oct 06 10:01:44 crc kubenswrapper[4824]: I1006 10:01:44.704872 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c470b33a-a2cd-46c8-b092-57d11b4b39c7-utilities" (OuterVolumeSpecName: "utilities") pod "c470b33a-a2cd-46c8-b092-57d11b4b39c7" (UID: "c470b33a-a2cd-46c8-b092-57d11b4b39c7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:01:44 crc kubenswrapper[4824]: I1006 10:01:44.711686 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c470b33a-a2cd-46c8-b092-57d11b4b39c7-kube-api-access-b2q9t" (OuterVolumeSpecName: "kube-api-access-b2q9t") pod "c470b33a-a2cd-46c8-b092-57d11b4b39c7" (UID: "c470b33a-a2cd-46c8-b092-57d11b4b39c7"). InnerVolumeSpecName "kube-api-access-b2q9t". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:01:44 crc kubenswrapper[4824]: I1006 10:01:44.794121 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c470b33a-a2cd-46c8-b092-57d11b4b39c7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c470b33a-a2cd-46c8-b092-57d11b4b39c7" (UID: "c470b33a-a2cd-46c8-b092-57d11b4b39c7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:01:44 crc kubenswrapper[4824]: I1006 10:01:44.804793 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b2q9t\" (UniqueName: \"kubernetes.io/projected/c470b33a-a2cd-46c8-b092-57d11b4b39c7-kube-api-access-b2q9t\") on node \"crc\" DevicePath \"\"" Oct 06 10:01:44 crc kubenswrapper[4824]: I1006 10:01:44.804826 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c470b33a-a2cd-46c8-b092-57d11b4b39c7-utilities\") on node \"crc\" DevicePath \"\"" Oct 06 10:01:44 crc kubenswrapper[4824]: I1006 10:01:44.804839 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c470b33a-a2cd-46c8-b092-57d11b4b39c7-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 06 10:01:45 crc kubenswrapper[4824]: I1006 10:01:45.174912 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" event={"ID":"e1140c8c-93f1-4fce-8c06-a69561ff8a6d","Type":"ContainerStarted","Data":"849cbe616fe0e1fc3c60c65e540b59cf8cac7989da24b5c2d88ead6d16645186"} Oct 06 10:01:45 crc kubenswrapper[4824]: I1006 10:01:45.178167 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bgbpb" event={"ID":"70324ded-62d7-47fc-a67a-110ecde1497c","Type":"ContainerStarted","Data":"e4db6b7d450dc54a789cbfe6f13124becd9116d29f9d07848dbeb547dab40be5"} Oct 06 10:01:45 crc kubenswrapper[4824]: I1006 10:01:45.180870 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dtcm6" event={"ID":"88764df0-c119-4e11-bd94-6818bc8ef83c","Type":"ContainerStarted","Data":"7c90fa9d2a03c0dc8133d97648cb70462c17c4544e169550690565bbaceb9a37"} Oct 06 10:01:45 crc kubenswrapper[4824]: I1006 10:01:45.184329 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r52zf" event={"ID":"06d3772b-e467-41a1-bf9b-9d6520ab68ef","Type":"ContainerStarted","Data":"c13a247642fda90779a68772ce9c3288d749e24fea28b15a22612cd488351a46"} Oct 06 10:01:45 crc kubenswrapper[4824]: I1006 10:01:45.186613 4824 generic.go:334] "Generic (PLEG): container finished" podID="c470b33a-a2cd-46c8-b092-57d11b4b39c7" containerID="2342c530649742f64a0fffa453a8e5d0d67b15d08fc3539cba60e6fe27138d7e" exitCode=0 Oct 06 10:01:45 crc kubenswrapper[4824]: I1006 10:01:45.186678 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sst8j" event={"ID":"c470b33a-a2cd-46c8-b092-57d11b4b39c7","Type":"ContainerDied","Data":"2342c530649742f64a0fffa453a8e5d0d67b15d08fc3539cba60e6fe27138d7e"} Oct 06 10:01:45 crc kubenswrapper[4824]: I1006 10:01:45.186698 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sst8j" event={"ID":"c470b33a-a2cd-46c8-b092-57d11b4b39c7","Type":"ContainerDied","Data":"cdebfbafa1fedc552ce52d0d8911fa4b901976e18169df50f8694c58d3aed9b4"} Oct 06 10:01:45 crc kubenswrapper[4824]: I1006 10:01:45.186718 4824 scope.go:117] "RemoveContainer" containerID="2342c530649742f64a0fffa453a8e5d0d67b15d08fc3539cba60e6fe27138d7e" Oct 06 10:01:45 crc kubenswrapper[4824]: I1006 10:01:45.186767 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sst8j" Oct 06 10:01:45 crc kubenswrapper[4824]: I1006 10:01:45.190075 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f7h6q" event={"ID":"f3ef12c1-f3ad-4ead-aba0-a3901f80fc7f","Type":"ContainerStarted","Data":"094ad890fb1487a6eb107f25d839e12b279b722dbbf95ff97041daf9106bf5f2"} Oct 06 10:01:45 crc kubenswrapper[4824]: I1006 10:01:45.192170 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zlxdl" event={"ID":"527511c8-2487-4416-a689-9ee161130297","Type":"ContainerStarted","Data":"9a9c0b330cdc0b1de23ab79988dced381bf11758dd9f780a3cf3d6b29d647052"} Oct 06 10:01:45 crc kubenswrapper[4824]: I1006 10:01:45.214217 4824 scope.go:117] "RemoveContainer" containerID="0505948a8d9c2922a4363c5632265c4db8f697a49af7a07e967e23cbbc9d7450" Oct 06 10:01:45 crc kubenswrapper[4824]: I1006 10:01:45.219488 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-r52zf" podStartSLOduration=3.807827653 podStartE2EDuration="57.219460879s" podCreationTimestamp="2025-10-06 10:00:48 +0000 UTC" firstStartedPulling="2025-10-06 10:00:51.404102479 +0000 UTC m=+160.768525340" lastFinishedPulling="2025-10-06 10:01:44.815735705 +0000 UTC m=+214.180158566" observedRunningTime="2025-10-06 10:01:45.218903784 +0000 UTC m=+214.583326645" watchObservedRunningTime="2025-10-06 10:01:45.219460879 +0000 UTC m=+214.583883740" Oct 06 10:01:45 crc kubenswrapper[4824]: I1006 10:01:45.240579 4824 scope.go:117] "RemoveContainer" containerID="f23e899e722c92f38881b2589bbbaa2f0b66cc00a4b51879a047488ff9dfdb22" Oct 06 10:01:45 crc kubenswrapper[4824]: I1006 10:01:45.257495 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-f7h6q" podStartSLOduration=1.990661186 podStartE2EDuration="55.257469753s" podCreationTimestamp="2025-10-06 10:00:50 +0000 UTC" firstStartedPulling="2025-10-06 10:00:51.414377462 +0000 UTC m=+160.778800323" lastFinishedPulling="2025-10-06 10:01:44.681186019 +0000 UTC m=+214.045608890" observedRunningTime="2025-10-06 10:01:45.255273268 +0000 UTC m=+214.619696129" watchObservedRunningTime="2025-10-06 10:01:45.257469753 +0000 UTC m=+214.621892614" Oct 06 10:01:45 crc kubenswrapper[4824]: I1006 10:01:45.261341 4824 scope.go:117] "RemoveContainer" containerID="2342c530649742f64a0fffa453a8e5d0d67b15d08fc3539cba60e6fe27138d7e" Oct 06 10:01:45 crc kubenswrapper[4824]: E1006 10:01:45.265088 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2342c530649742f64a0fffa453a8e5d0d67b15d08fc3539cba60e6fe27138d7e\": container with ID starting with 2342c530649742f64a0fffa453a8e5d0d67b15d08fc3539cba60e6fe27138d7e not found: ID does not exist" containerID="2342c530649742f64a0fffa453a8e5d0d67b15d08fc3539cba60e6fe27138d7e" Oct 06 10:01:45 crc kubenswrapper[4824]: I1006 10:01:45.265132 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2342c530649742f64a0fffa453a8e5d0d67b15d08fc3539cba60e6fe27138d7e"} err="failed to get container status \"2342c530649742f64a0fffa453a8e5d0d67b15d08fc3539cba60e6fe27138d7e\": rpc error: code = NotFound desc = could not find container \"2342c530649742f64a0fffa453a8e5d0d67b15d08fc3539cba60e6fe27138d7e\": container with ID starting with 2342c530649742f64a0fffa453a8e5d0d67b15d08fc3539cba60e6fe27138d7e not found: ID does not exist" Oct 06 10:01:45 crc kubenswrapper[4824]: I1006 10:01:45.265160 4824 scope.go:117] "RemoveContainer" containerID="0505948a8d9c2922a4363c5632265c4db8f697a49af7a07e967e23cbbc9d7450" Oct 06 10:01:45 crc kubenswrapper[4824]: E1006 10:01:45.268073 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0505948a8d9c2922a4363c5632265c4db8f697a49af7a07e967e23cbbc9d7450\": container with ID starting with 0505948a8d9c2922a4363c5632265c4db8f697a49af7a07e967e23cbbc9d7450 not found: ID does not exist" containerID="0505948a8d9c2922a4363c5632265c4db8f697a49af7a07e967e23cbbc9d7450" Oct 06 10:01:45 crc kubenswrapper[4824]: I1006 10:01:45.268125 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0505948a8d9c2922a4363c5632265c4db8f697a49af7a07e967e23cbbc9d7450"} err="failed to get container status \"0505948a8d9c2922a4363c5632265c4db8f697a49af7a07e967e23cbbc9d7450\": rpc error: code = NotFound desc = could not find container \"0505948a8d9c2922a4363c5632265c4db8f697a49af7a07e967e23cbbc9d7450\": container with ID starting with 0505948a8d9c2922a4363c5632265c4db8f697a49af7a07e967e23cbbc9d7450 not found: ID does not exist" Oct 06 10:01:45 crc kubenswrapper[4824]: I1006 10:01:45.268144 4824 scope.go:117] "RemoveContainer" containerID="f23e899e722c92f38881b2589bbbaa2f0b66cc00a4b51879a047488ff9dfdb22" Oct 06 10:01:45 crc kubenswrapper[4824]: E1006 10:01:45.268481 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f23e899e722c92f38881b2589bbbaa2f0b66cc00a4b51879a047488ff9dfdb22\": container with ID starting with f23e899e722c92f38881b2589bbbaa2f0b66cc00a4b51879a047488ff9dfdb22 not found: ID does not exist" containerID="f23e899e722c92f38881b2589bbbaa2f0b66cc00a4b51879a047488ff9dfdb22" Oct 06 10:01:45 crc kubenswrapper[4824]: I1006 10:01:45.268509 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f23e899e722c92f38881b2589bbbaa2f0b66cc00a4b51879a047488ff9dfdb22"} err="failed to get container status \"f23e899e722c92f38881b2589bbbaa2f0b66cc00a4b51879a047488ff9dfdb22\": rpc error: code = NotFound desc = could not find container \"f23e899e722c92f38881b2589bbbaa2f0b66cc00a4b51879a047488ff9dfdb22\": container with ID starting with f23e899e722c92f38881b2589bbbaa2f0b66cc00a4b51879a047488ff9dfdb22 not found: ID does not exist" Oct 06 10:01:45 crc kubenswrapper[4824]: I1006 10:01:45.278116 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-bgbpb" podStartSLOduration=2.022073065 podStartE2EDuration="55.27809144s" podCreationTimestamp="2025-10-06 10:00:50 +0000 UTC" firstStartedPulling="2025-10-06 10:00:51.358713771 +0000 UTC m=+160.723136632" lastFinishedPulling="2025-10-06 10:01:44.614732146 +0000 UTC m=+213.979155007" observedRunningTime="2025-10-06 10:01:45.27284412 +0000 UTC m=+214.637266981" watchObservedRunningTime="2025-10-06 10:01:45.27809144 +0000 UTC m=+214.642514301" Oct 06 10:01:45 crc kubenswrapper[4824]: I1006 10:01:45.294569 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-dtcm6" podStartSLOduration=4.097914601 podStartE2EDuration="57.294544463s" podCreationTimestamp="2025-10-06 10:00:48 +0000 UTC" firstStartedPulling="2025-10-06 10:00:51.397756653 +0000 UTC m=+160.762179514" lastFinishedPulling="2025-10-06 10:01:44.594386515 +0000 UTC m=+213.958809376" observedRunningTime="2025-10-06 10:01:45.292285168 +0000 UTC m=+214.656708029" watchObservedRunningTime="2025-10-06 10:01:45.294544463 +0000 UTC m=+214.658967314" Oct 06 10:01:45 crc kubenswrapper[4824]: I1006 10:01:45.319368 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-zlxdl" podStartSLOduration=3.783331455 podStartE2EDuration="57.319345443s" podCreationTimestamp="2025-10-06 10:00:48 +0000 UTC" firstStartedPulling="2025-10-06 10:00:51.336504784 +0000 UTC m=+160.700927645" lastFinishedPulling="2025-10-06 10:01:44.872518772 +0000 UTC m=+214.236941633" observedRunningTime="2025-10-06 10:01:45.318222155 +0000 UTC m=+214.682645016" watchObservedRunningTime="2025-10-06 10:01:45.319345443 +0000 UTC m=+214.683768304" Oct 06 10:01:45 crc kubenswrapper[4824]: I1006 10:01:45.330521 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-sst8j"] Oct 06 10:01:45 crc kubenswrapper[4824]: I1006 10:01:45.334586 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-sst8j"] Oct 06 10:01:47 crc kubenswrapper[4824]: I1006 10:01:47.284697 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c470b33a-a2cd-46c8-b092-57d11b4b39c7" path="/var/lib/kubelet/pods/c470b33a-a2cd-46c8-b092-57d11b4b39c7/volumes" Oct 06 10:01:48 crc kubenswrapper[4824]: I1006 10:01:48.570548 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-tlvv7" Oct 06 10:01:48 crc kubenswrapper[4824]: I1006 10:01:48.570996 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-tlvv7" Oct 06 10:01:48 crc kubenswrapper[4824]: I1006 10:01:48.630210 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-tlvv7" Oct 06 10:01:48 crc kubenswrapper[4824]: I1006 10:01:48.690645 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-r52zf" Oct 06 10:01:48 crc kubenswrapper[4824]: I1006 10:01:48.690719 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-r52zf" Oct 06 10:01:48 crc kubenswrapper[4824]: I1006 10:01:48.736639 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-r52zf" Oct 06 10:01:48 crc kubenswrapper[4824]: I1006 10:01:48.920148 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-zlxdl" Oct 06 10:01:48 crc kubenswrapper[4824]: I1006 10:01:48.920222 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-zlxdl" Oct 06 10:01:48 crc kubenswrapper[4824]: I1006 10:01:48.964319 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-zlxdl" Oct 06 10:01:49 crc kubenswrapper[4824]: I1006 10:01:49.134059 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-dtcm6" Oct 06 10:01:49 crc kubenswrapper[4824]: I1006 10:01:49.134114 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-dtcm6" Oct 06 10:01:49 crc kubenswrapper[4824]: I1006 10:01:49.174199 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-dtcm6" Oct 06 10:01:49 crc kubenswrapper[4824]: I1006 10:01:49.268472 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-dtcm6" Oct 06 10:01:49 crc kubenswrapper[4824]: I1006 10:01:49.281604 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-r52zf" Oct 06 10:01:49 crc kubenswrapper[4824]: I1006 10:01:49.281666 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-tlvv7" Oct 06 10:01:50 crc kubenswrapper[4824]: I1006 10:01:50.535212 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-bgbpb" Oct 06 10:01:50 crc kubenswrapper[4824]: I1006 10:01:50.536775 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-bgbpb" Oct 06 10:01:50 crc kubenswrapper[4824]: I1006 10:01:50.600329 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-bgbpb" Oct 06 10:01:50 crc kubenswrapper[4824]: I1006 10:01:50.808526 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-f7h6q" Oct 06 10:01:50 crc kubenswrapper[4824]: I1006 10:01:50.808581 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-f7h6q" Oct 06 10:01:50 crc kubenswrapper[4824]: I1006 10:01:50.877255 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-f7h6q" Oct 06 10:01:51 crc kubenswrapper[4824]: I1006 10:01:51.287212 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-bgbpb" Oct 06 10:01:51 crc kubenswrapper[4824]: I1006 10:01:51.302688 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-f7h6q" Oct 06 10:01:51 crc kubenswrapper[4824]: I1006 10:01:51.339460 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-dtcm6"] Oct 06 10:01:51 crc kubenswrapper[4824]: I1006 10:01:51.340300 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-dtcm6" podUID="88764df0-c119-4e11-bd94-6818bc8ef83c" containerName="registry-server" containerID="cri-o://7c90fa9d2a03c0dc8133d97648cb70462c17c4544e169550690565bbaceb9a37" gracePeriod=2 Oct 06 10:01:51 crc kubenswrapper[4824]: I1006 10:01:51.789288 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dtcm6" Oct 06 10:01:51 crc kubenswrapper[4824]: I1006 10:01:51.980587 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dqjms\" (UniqueName: \"kubernetes.io/projected/88764df0-c119-4e11-bd94-6818bc8ef83c-kube-api-access-dqjms\") pod \"88764df0-c119-4e11-bd94-6818bc8ef83c\" (UID: \"88764df0-c119-4e11-bd94-6818bc8ef83c\") " Oct 06 10:01:51 crc kubenswrapper[4824]: I1006 10:01:51.980663 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/88764df0-c119-4e11-bd94-6818bc8ef83c-utilities\") pod \"88764df0-c119-4e11-bd94-6818bc8ef83c\" (UID: \"88764df0-c119-4e11-bd94-6818bc8ef83c\") " Oct 06 10:01:51 crc kubenswrapper[4824]: I1006 10:01:51.980774 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/88764df0-c119-4e11-bd94-6818bc8ef83c-catalog-content\") pod \"88764df0-c119-4e11-bd94-6818bc8ef83c\" (UID: \"88764df0-c119-4e11-bd94-6818bc8ef83c\") " Oct 06 10:01:51 crc kubenswrapper[4824]: I1006 10:01:51.982845 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/88764df0-c119-4e11-bd94-6818bc8ef83c-utilities" (OuterVolumeSpecName: "utilities") pod "88764df0-c119-4e11-bd94-6818bc8ef83c" (UID: "88764df0-c119-4e11-bd94-6818bc8ef83c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:01:51 crc kubenswrapper[4824]: I1006 10:01:51.990360 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/88764df0-c119-4e11-bd94-6818bc8ef83c-kube-api-access-dqjms" (OuterVolumeSpecName: "kube-api-access-dqjms") pod "88764df0-c119-4e11-bd94-6818bc8ef83c" (UID: "88764df0-c119-4e11-bd94-6818bc8ef83c"). InnerVolumeSpecName "kube-api-access-dqjms". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:01:52 crc kubenswrapper[4824]: I1006 10:01:52.036142 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/88764df0-c119-4e11-bd94-6818bc8ef83c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "88764df0-c119-4e11-bd94-6818bc8ef83c" (UID: "88764df0-c119-4e11-bd94-6818bc8ef83c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:01:52 crc kubenswrapper[4824]: I1006 10:01:52.082433 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dqjms\" (UniqueName: \"kubernetes.io/projected/88764df0-c119-4e11-bd94-6818bc8ef83c-kube-api-access-dqjms\") on node \"crc\" DevicePath \"\"" Oct 06 10:01:52 crc kubenswrapper[4824]: I1006 10:01:52.082480 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/88764df0-c119-4e11-bd94-6818bc8ef83c-utilities\") on node \"crc\" DevicePath \"\"" Oct 06 10:01:52 crc kubenswrapper[4824]: I1006 10:01:52.082499 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/88764df0-c119-4e11-bd94-6818bc8ef83c-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 06 10:01:52 crc kubenswrapper[4824]: I1006 10:01:52.239944 4824 generic.go:334] "Generic (PLEG): container finished" podID="88764df0-c119-4e11-bd94-6818bc8ef83c" containerID="7c90fa9d2a03c0dc8133d97648cb70462c17c4544e169550690565bbaceb9a37" exitCode=0 Oct 06 10:01:52 crc kubenswrapper[4824]: I1006 10:01:52.240076 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dtcm6" Oct 06 10:01:52 crc kubenswrapper[4824]: I1006 10:01:52.240101 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dtcm6" event={"ID":"88764df0-c119-4e11-bd94-6818bc8ef83c","Type":"ContainerDied","Data":"7c90fa9d2a03c0dc8133d97648cb70462c17c4544e169550690565bbaceb9a37"} Oct 06 10:01:52 crc kubenswrapper[4824]: I1006 10:01:52.242205 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dtcm6" event={"ID":"88764df0-c119-4e11-bd94-6818bc8ef83c","Type":"ContainerDied","Data":"ab2c98e63ed8f8b9561d937d2f7127db8783130bfba458bb9a603e644acfa965"} Oct 06 10:01:52 crc kubenswrapper[4824]: I1006 10:01:52.242360 4824 scope.go:117] "RemoveContainer" containerID="7c90fa9d2a03c0dc8133d97648cb70462c17c4544e169550690565bbaceb9a37" Oct 06 10:01:52 crc kubenswrapper[4824]: I1006 10:01:52.264212 4824 scope.go:117] "RemoveContainer" containerID="0a30ca0f05eecfc1279b5e26a03155ef8d797e21fdd83d42c35aa685a8bb70ae" Oct 06 10:01:52 crc kubenswrapper[4824]: I1006 10:01:52.290122 4824 scope.go:117] "RemoveContainer" containerID="3426618e5cee07f375821734c951c2191a5c280c544c54ba3a509a664b4206f9" Oct 06 10:01:52 crc kubenswrapper[4824]: I1006 10:01:52.294198 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-dtcm6"] Oct 06 10:01:52 crc kubenswrapper[4824]: I1006 10:01:52.302371 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-dtcm6"] Oct 06 10:01:52 crc kubenswrapper[4824]: I1006 10:01:52.314923 4824 scope.go:117] "RemoveContainer" containerID="7c90fa9d2a03c0dc8133d97648cb70462c17c4544e169550690565bbaceb9a37" Oct 06 10:01:52 crc kubenswrapper[4824]: E1006 10:01:52.315315 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7c90fa9d2a03c0dc8133d97648cb70462c17c4544e169550690565bbaceb9a37\": container with ID starting with 7c90fa9d2a03c0dc8133d97648cb70462c17c4544e169550690565bbaceb9a37 not found: ID does not exist" containerID="7c90fa9d2a03c0dc8133d97648cb70462c17c4544e169550690565bbaceb9a37" Oct 06 10:01:52 crc kubenswrapper[4824]: I1006 10:01:52.315371 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7c90fa9d2a03c0dc8133d97648cb70462c17c4544e169550690565bbaceb9a37"} err="failed to get container status \"7c90fa9d2a03c0dc8133d97648cb70462c17c4544e169550690565bbaceb9a37\": rpc error: code = NotFound desc = could not find container \"7c90fa9d2a03c0dc8133d97648cb70462c17c4544e169550690565bbaceb9a37\": container with ID starting with 7c90fa9d2a03c0dc8133d97648cb70462c17c4544e169550690565bbaceb9a37 not found: ID does not exist" Oct 06 10:01:52 crc kubenswrapper[4824]: I1006 10:01:52.315430 4824 scope.go:117] "RemoveContainer" containerID="0a30ca0f05eecfc1279b5e26a03155ef8d797e21fdd83d42c35aa685a8bb70ae" Oct 06 10:01:52 crc kubenswrapper[4824]: E1006 10:01:52.316214 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0a30ca0f05eecfc1279b5e26a03155ef8d797e21fdd83d42c35aa685a8bb70ae\": container with ID starting with 0a30ca0f05eecfc1279b5e26a03155ef8d797e21fdd83d42c35aa685a8bb70ae not found: ID does not exist" containerID="0a30ca0f05eecfc1279b5e26a03155ef8d797e21fdd83d42c35aa685a8bb70ae" Oct 06 10:01:52 crc kubenswrapper[4824]: I1006 10:01:52.316271 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0a30ca0f05eecfc1279b5e26a03155ef8d797e21fdd83d42c35aa685a8bb70ae"} err="failed to get container status \"0a30ca0f05eecfc1279b5e26a03155ef8d797e21fdd83d42c35aa685a8bb70ae\": rpc error: code = NotFound desc = could not find container \"0a30ca0f05eecfc1279b5e26a03155ef8d797e21fdd83d42c35aa685a8bb70ae\": container with ID starting with 0a30ca0f05eecfc1279b5e26a03155ef8d797e21fdd83d42c35aa685a8bb70ae not found: ID does not exist" Oct 06 10:01:52 crc kubenswrapper[4824]: I1006 10:01:52.316316 4824 scope.go:117] "RemoveContainer" containerID="3426618e5cee07f375821734c951c2191a5c280c544c54ba3a509a664b4206f9" Oct 06 10:01:52 crc kubenswrapper[4824]: E1006 10:01:52.316711 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3426618e5cee07f375821734c951c2191a5c280c544c54ba3a509a664b4206f9\": container with ID starting with 3426618e5cee07f375821734c951c2191a5c280c544c54ba3a509a664b4206f9 not found: ID does not exist" containerID="3426618e5cee07f375821734c951c2191a5c280c544c54ba3a509a664b4206f9" Oct 06 10:01:52 crc kubenswrapper[4824]: I1006 10:01:52.316755 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3426618e5cee07f375821734c951c2191a5c280c544c54ba3a509a664b4206f9"} err="failed to get container status \"3426618e5cee07f375821734c951c2191a5c280c544c54ba3a509a664b4206f9\": rpc error: code = NotFound desc = could not find container \"3426618e5cee07f375821734c951c2191a5c280c544c54ba3a509a664b4206f9\": container with ID starting with 3426618e5cee07f375821734c951c2191a5c280c544c54ba3a509a664b4206f9 not found: ID does not exist" Oct 06 10:01:53 crc kubenswrapper[4824]: I1006 10:01:53.281056 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="88764df0-c119-4e11-bd94-6818bc8ef83c" path="/var/lib/kubelet/pods/88764df0-c119-4e11-bd94-6818bc8ef83c/volumes" Oct 06 10:01:53 crc kubenswrapper[4824]: I1006 10:01:53.730857 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-f7h6q"] Oct 06 10:01:53 crc kubenswrapper[4824]: I1006 10:01:53.731203 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-f7h6q" podUID="f3ef12c1-f3ad-4ead-aba0-a3901f80fc7f" containerName="registry-server" containerID="cri-o://094ad890fb1487a6eb107f25d839e12b279b722dbbf95ff97041daf9106bf5f2" gracePeriod=2 Oct 06 10:01:54 crc kubenswrapper[4824]: I1006 10:01:54.219553 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-f7h6q" Oct 06 10:01:54 crc kubenswrapper[4824]: I1006 10:01:54.267403 4824 generic.go:334] "Generic (PLEG): container finished" podID="f3ef12c1-f3ad-4ead-aba0-a3901f80fc7f" containerID="094ad890fb1487a6eb107f25d839e12b279b722dbbf95ff97041daf9106bf5f2" exitCode=0 Oct 06 10:01:54 crc kubenswrapper[4824]: I1006 10:01:54.267474 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f7h6q" event={"ID":"f3ef12c1-f3ad-4ead-aba0-a3901f80fc7f","Type":"ContainerDied","Data":"094ad890fb1487a6eb107f25d839e12b279b722dbbf95ff97041daf9106bf5f2"} Oct 06 10:01:54 crc kubenswrapper[4824]: I1006 10:01:54.267522 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f7h6q" event={"ID":"f3ef12c1-f3ad-4ead-aba0-a3901f80fc7f","Type":"ContainerDied","Data":"ff7368a789c489bb4b27e52912320bd7d9f60525ae61173255c9eda795b11d75"} Oct 06 10:01:54 crc kubenswrapper[4824]: I1006 10:01:54.267548 4824 scope.go:117] "RemoveContainer" containerID="094ad890fb1487a6eb107f25d839e12b279b722dbbf95ff97041daf9106bf5f2" Oct 06 10:01:54 crc kubenswrapper[4824]: I1006 10:01:54.267943 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-f7h6q" Oct 06 10:01:54 crc kubenswrapper[4824]: I1006 10:01:54.302904 4824 scope.go:117] "RemoveContainer" containerID="6082a44ad6b26d8d1619132966e6e4b385079e36321830323e9b482f287f9b7e" Oct 06 10:01:54 crc kubenswrapper[4824]: I1006 10:01:54.325873 4824 scope.go:117] "RemoveContainer" containerID="fc6a773b82ac29844de4dfe214e868ae3d84e8217ea3e4f3f17b7ba83fe0134c" Oct 06 10:01:54 crc kubenswrapper[4824]: I1006 10:01:54.343247 4824 scope.go:117] "RemoveContainer" containerID="094ad890fb1487a6eb107f25d839e12b279b722dbbf95ff97041daf9106bf5f2" Oct 06 10:01:54 crc kubenswrapper[4824]: E1006 10:01:54.343732 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"094ad890fb1487a6eb107f25d839e12b279b722dbbf95ff97041daf9106bf5f2\": container with ID starting with 094ad890fb1487a6eb107f25d839e12b279b722dbbf95ff97041daf9106bf5f2 not found: ID does not exist" containerID="094ad890fb1487a6eb107f25d839e12b279b722dbbf95ff97041daf9106bf5f2" Oct 06 10:01:54 crc kubenswrapper[4824]: I1006 10:01:54.343802 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"094ad890fb1487a6eb107f25d839e12b279b722dbbf95ff97041daf9106bf5f2"} err="failed to get container status \"094ad890fb1487a6eb107f25d839e12b279b722dbbf95ff97041daf9106bf5f2\": rpc error: code = NotFound desc = could not find container \"094ad890fb1487a6eb107f25d839e12b279b722dbbf95ff97041daf9106bf5f2\": container with ID starting with 094ad890fb1487a6eb107f25d839e12b279b722dbbf95ff97041daf9106bf5f2 not found: ID does not exist" Oct 06 10:01:54 crc kubenswrapper[4824]: I1006 10:01:54.343838 4824 scope.go:117] "RemoveContainer" containerID="6082a44ad6b26d8d1619132966e6e4b385079e36321830323e9b482f287f9b7e" Oct 06 10:01:54 crc kubenswrapper[4824]: E1006 10:01:54.344202 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6082a44ad6b26d8d1619132966e6e4b385079e36321830323e9b482f287f9b7e\": container with ID starting with 6082a44ad6b26d8d1619132966e6e4b385079e36321830323e9b482f287f9b7e not found: ID does not exist" containerID="6082a44ad6b26d8d1619132966e6e4b385079e36321830323e9b482f287f9b7e" Oct 06 10:01:54 crc kubenswrapper[4824]: I1006 10:01:54.344257 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6082a44ad6b26d8d1619132966e6e4b385079e36321830323e9b482f287f9b7e"} err="failed to get container status \"6082a44ad6b26d8d1619132966e6e4b385079e36321830323e9b482f287f9b7e\": rpc error: code = NotFound desc = could not find container \"6082a44ad6b26d8d1619132966e6e4b385079e36321830323e9b482f287f9b7e\": container with ID starting with 6082a44ad6b26d8d1619132966e6e4b385079e36321830323e9b482f287f9b7e not found: ID does not exist" Oct 06 10:01:54 crc kubenswrapper[4824]: I1006 10:01:54.344277 4824 scope.go:117] "RemoveContainer" containerID="fc6a773b82ac29844de4dfe214e868ae3d84e8217ea3e4f3f17b7ba83fe0134c" Oct 06 10:01:54 crc kubenswrapper[4824]: E1006 10:01:54.344607 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fc6a773b82ac29844de4dfe214e868ae3d84e8217ea3e4f3f17b7ba83fe0134c\": container with ID starting with fc6a773b82ac29844de4dfe214e868ae3d84e8217ea3e4f3f17b7ba83fe0134c not found: ID does not exist" containerID="fc6a773b82ac29844de4dfe214e868ae3d84e8217ea3e4f3f17b7ba83fe0134c" Oct 06 10:01:54 crc kubenswrapper[4824]: I1006 10:01:54.344660 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fc6a773b82ac29844de4dfe214e868ae3d84e8217ea3e4f3f17b7ba83fe0134c"} err="failed to get container status \"fc6a773b82ac29844de4dfe214e868ae3d84e8217ea3e4f3f17b7ba83fe0134c\": rpc error: code = NotFound desc = could not find container \"fc6a773b82ac29844de4dfe214e868ae3d84e8217ea3e4f3f17b7ba83fe0134c\": container with ID starting with fc6a773b82ac29844de4dfe214e868ae3d84e8217ea3e4f3f17b7ba83fe0134c not found: ID does not exist" Oct 06 10:01:54 crc kubenswrapper[4824]: I1006 10:01:54.414821 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f3ef12c1-f3ad-4ead-aba0-a3901f80fc7f-utilities\") pod \"f3ef12c1-f3ad-4ead-aba0-a3901f80fc7f\" (UID: \"f3ef12c1-f3ad-4ead-aba0-a3901f80fc7f\") " Oct 06 10:01:54 crc kubenswrapper[4824]: I1006 10:01:54.414936 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m4dlf\" (UniqueName: \"kubernetes.io/projected/f3ef12c1-f3ad-4ead-aba0-a3901f80fc7f-kube-api-access-m4dlf\") pod \"f3ef12c1-f3ad-4ead-aba0-a3901f80fc7f\" (UID: \"f3ef12c1-f3ad-4ead-aba0-a3901f80fc7f\") " Oct 06 10:01:54 crc kubenswrapper[4824]: I1006 10:01:54.415054 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f3ef12c1-f3ad-4ead-aba0-a3901f80fc7f-catalog-content\") pod \"f3ef12c1-f3ad-4ead-aba0-a3901f80fc7f\" (UID: \"f3ef12c1-f3ad-4ead-aba0-a3901f80fc7f\") " Oct 06 10:01:54 crc kubenswrapper[4824]: I1006 10:01:54.416497 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f3ef12c1-f3ad-4ead-aba0-a3901f80fc7f-utilities" (OuterVolumeSpecName: "utilities") pod "f3ef12c1-f3ad-4ead-aba0-a3901f80fc7f" (UID: "f3ef12c1-f3ad-4ead-aba0-a3901f80fc7f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:01:54 crc kubenswrapper[4824]: I1006 10:01:54.427167 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f3ef12c1-f3ad-4ead-aba0-a3901f80fc7f-kube-api-access-m4dlf" (OuterVolumeSpecName: "kube-api-access-m4dlf") pod "f3ef12c1-f3ad-4ead-aba0-a3901f80fc7f" (UID: "f3ef12c1-f3ad-4ead-aba0-a3901f80fc7f"). InnerVolumeSpecName "kube-api-access-m4dlf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:01:54 crc kubenswrapper[4824]: I1006 10:01:54.438586 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f3ef12c1-f3ad-4ead-aba0-a3901f80fc7f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f3ef12c1-f3ad-4ead-aba0-a3901f80fc7f" (UID: "f3ef12c1-f3ad-4ead-aba0-a3901f80fc7f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:01:54 crc kubenswrapper[4824]: I1006 10:01:54.518212 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f3ef12c1-f3ad-4ead-aba0-a3901f80fc7f-utilities\") on node \"crc\" DevicePath \"\"" Oct 06 10:01:54 crc kubenswrapper[4824]: I1006 10:01:54.518787 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m4dlf\" (UniqueName: \"kubernetes.io/projected/f3ef12c1-f3ad-4ead-aba0-a3901f80fc7f-kube-api-access-m4dlf\") on node \"crc\" DevicePath \"\"" Oct 06 10:01:54 crc kubenswrapper[4824]: I1006 10:01:54.518921 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f3ef12c1-f3ad-4ead-aba0-a3901f80fc7f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 06 10:01:54 crc kubenswrapper[4824]: I1006 10:01:54.602596 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-f7h6q"] Oct 06 10:01:54 crc kubenswrapper[4824]: I1006 10:01:54.604898 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-f7h6q"] Oct 06 10:01:55 crc kubenswrapper[4824]: I1006 10:01:55.282739 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f3ef12c1-f3ad-4ead-aba0-a3901f80fc7f" path="/var/lib/kubelet/pods/f3ef12c1-f3ad-4ead-aba0-a3901f80fc7f/volumes" Oct 06 10:01:58 crc kubenswrapper[4824]: I1006 10:01:58.974322 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-zlxdl" Oct 06 10:02:00 crc kubenswrapper[4824]: I1006 10:02:00.131549 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zlxdl"] Oct 06 10:02:00 crc kubenswrapper[4824]: I1006 10:02:00.132432 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-zlxdl" podUID="527511c8-2487-4416-a689-9ee161130297" containerName="registry-server" containerID="cri-o://9a9c0b330cdc0b1de23ab79988dced381bf11758dd9f780a3cf3d6b29d647052" gracePeriod=2 Oct 06 10:02:00 crc kubenswrapper[4824]: I1006 10:02:00.326775 4824 generic.go:334] "Generic (PLEG): container finished" podID="527511c8-2487-4416-a689-9ee161130297" containerID="9a9c0b330cdc0b1de23ab79988dced381bf11758dd9f780a3cf3d6b29d647052" exitCode=0 Oct 06 10:02:00 crc kubenswrapper[4824]: I1006 10:02:00.326836 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zlxdl" event={"ID":"527511c8-2487-4416-a689-9ee161130297","Type":"ContainerDied","Data":"9a9c0b330cdc0b1de23ab79988dced381bf11758dd9f780a3cf3d6b29d647052"} Oct 06 10:02:00 crc kubenswrapper[4824]: I1006 10:02:00.514813 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zlxdl" Oct 06 10:02:00 crc kubenswrapper[4824]: I1006 10:02:00.611263 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vc8nz\" (UniqueName: \"kubernetes.io/projected/527511c8-2487-4416-a689-9ee161130297-kube-api-access-vc8nz\") pod \"527511c8-2487-4416-a689-9ee161130297\" (UID: \"527511c8-2487-4416-a689-9ee161130297\") " Oct 06 10:02:00 crc kubenswrapper[4824]: I1006 10:02:00.611889 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/527511c8-2487-4416-a689-9ee161130297-catalog-content\") pod \"527511c8-2487-4416-a689-9ee161130297\" (UID: \"527511c8-2487-4416-a689-9ee161130297\") " Oct 06 10:02:00 crc kubenswrapper[4824]: I1006 10:02:00.611925 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/527511c8-2487-4416-a689-9ee161130297-utilities\") pod \"527511c8-2487-4416-a689-9ee161130297\" (UID: \"527511c8-2487-4416-a689-9ee161130297\") " Oct 06 10:02:00 crc kubenswrapper[4824]: I1006 10:02:00.615534 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/527511c8-2487-4416-a689-9ee161130297-utilities" (OuterVolumeSpecName: "utilities") pod "527511c8-2487-4416-a689-9ee161130297" (UID: "527511c8-2487-4416-a689-9ee161130297"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:02:00 crc kubenswrapper[4824]: I1006 10:02:00.628173 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/527511c8-2487-4416-a689-9ee161130297-kube-api-access-vc8nz" (OuterVolumeSpecName: "kube-api-access-vc8nz") pod "527511c8-2487-4416-a689-9ee161130297" (UID: "527511c8-2487-4416-a689-9ee161130297"). InnerVolumeSpecName "kube-api-access-vc8nz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:02:00 crc kubenswrapper[4824]: I1006 10:02:00.671507 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/527511c8-2487-4416-a689-9ee161130297-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "527511c8-2487-4416-a689-9ee161130297" (UID: "527511c8-2487-4416-a689-9ee161130297"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:02:00 crc kubenswrapper[4824]: I1006 10:02:00.713371 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vc8nz\" (UniqueName: \"kubernetes.io/projected/527511c8-2487-4416-a689-9ee161130297-kube-api-access-vc8nz\") on node \"crc\" DevicePath \"\"" Oct 06 10:02:00 crc kubenswrapper[4824]: I1006 10:02:00.713417 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/527511c8-2487-4416-a689-9ee161130297-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 06 10:02:00 crc kubenswrapper[4824]: I1006 10:02:00.713429 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/527511c8-2487-4416-a689-9ee161130297-utilities\") on node \"crc\" DevicePath \"\"" Oct 06 10:02:01 crc kubenswrapper[4824]: I1006 10:02:01.334371 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zlxdl" event={"ID":"527511c8-2487-4416-a689-9ee161130297","Type":"ContainerDied","Data":"e653732eeb0042c8b5b4e52c01ec69ddec3c1572905678874a2cdbd35dd0af59"} Oct 06 10:02:01 crc kubenswrapper[4824]: I1006 10:02:01.334454 4824 scope.go:117] "RemoveContainer" containerID="9a9c0b330cdc0b1de23ab79988dced381bf11758dd9f780a3cf3d6b29d647052" Oct 06 10:02:01 crc kubenswrapper[4824]: I1006 10:02:01.334483 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zlxdl" Oct 06 10:02:01 crc kubenswrapper[4824]: I1006 10:02:01.357583 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zlxdl"] Oct 06 10:02:01 crc kubenswrapper[4824]: I1006 10:02:01.358640 4824 scope.go:117] "RemoveContainer" containerID="5b50704f318215ea643719849c6a7433e5ba1be0efabfb94c3d69e6d92741990" Oct 06 10:02:01 crc kubenswrapper[4824]: I1006 10:02:01.360490 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-zlxdl"] Oct 06 10:02:01 crc kubenswrapper[4824]: I1006 10:02:01.375367 4824 scope.go:117] "RemoveContainer" containerID="f4bfe51a6975baafcfaf9e53a1fd4a39a6bc9b04f30ac15711ae5aeeccfab3e7" Oct 06 10:02:02 crc kubenswrapper[4824]: I1006 10:02:02.314651 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-8z68k"] Oct 06 10:02:03 crc kubenswrapper[4824]: I1006 10:02:03.280242 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="527511c8-2487-4416-a689-9ee161130297" path="/var/lib/kubelet/pods/527511c8-2487-4416-a689-9ee161130297/volumes" Oct 06 10:02:27 crc kubenswrapper[4824]: I1006 10:02:27.368818 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-8z68k" podUID="af8fbf5a-f90e-4d22-ae65-05d8d8253308" containerName="oauth-openshift" containerID="cri-o://119cba5ffd73d0ff173f47322893c5b169be0630ac8c94dbbc819afb9ec746eb" gracePeriod=15 Oct 06 10:02:27 crc kubenswrapper[4824]: I1006 10:02:27.517616 4824 generic.go:334] "Generic (PLEG): container finished" podID="af8fbf5a-f90e-4d22-ae65-05d8d8253308" containerID="119cba5ffd73d0ff173f47322893c5b169be0630ac8c94dbbc819afb9ec746eb" exitCode=0 Oct 06 10:02:27 crc kubenswrapper[4824]: I1006 10:02:27.517712 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-8z68k" event={"ID":"af8fbf5a-f90e-4d22-ae65-05d8d8253308","Type":"ContainerDied","Data":"119cba5ffd73d0ff173f47322893c5b169be0630ac8c94dbbc819afb9ec746eb"} Oct 06 10:02:27 crc kubenswrapper[4824]: I1006 10:02:27.853392 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-8z68k" Oct 06 10:02:27 crc kubenswrapper[4824]: I1006 10:02:27.910634 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-6994c46f75-97hq2"] Oct 06 10:02:27 crc kubenswrapper[4824]: E1006 10:02:27.910963 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88764df0-c119-4e11-bd94-6818bc8ef83c" containerName="extract-utilities" Oct 06 10:02:27 crc kubenswrapper[4824]: I1006 10:02:27.911000 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="88764df0-c119-4e11-bd94-6818bc8ef83c" containerName="extract-utilities" Oct 06 10:02:27 crc kubenswrapper[4824]: E1006 10:02:27.911015 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="527511c8-2487-4416-a689-9ee161130297" containerName="extract-utilities" Oct 06 10:02:27 crc kubenswrapper[4824]: I1006 10:02:27.911023 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="527511c8-2487-4416-a689-9ee161130297" containerName="extract-utilities" Oct 06 10:02:27 crc kubenswrapper[4824]: E1006 10:02:27.911031 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da7d73fc-5908-4185-b203-854619742975" containerName="pruner" Oct 06 10:02:27 crc kubenswrapper[4824]: I1006 10:02:27.911041 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="da7d73fc-5908-4185-b203-854619742975" containerName="pruner" Oct 06 10:02:27 crc kubenswrapper[4824]: E1006 10:02:27.911054 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88764df0-c119-4e11-bd94-6818bc8ef83c" containerName="extract-content" Oct 06 10:02:27 crc kubenswrapper[4824]: I1006 10:02:27.911062 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="88764df0-c119-4e11-bd94-6818bc8ef83c" containerName="extract-content" Oct 06 10:02:27 crc kubenswrapper[4824]: E1006 10:02:27.911077 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3ef12c1-f3ad-4ead-aba0-a3901f80fc7f" containerName="extract-utilities" Oct 06 10:02:27 crc kubenswrapper[4824]: I1006 10:02:27.911085 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3ef12c1-f3ad-4ead-aba0-a3901f80fc7f" containerName="extract-utilities" Oct 06 10:02:27 crc kubenswrapper[4824]: E1006 10:02:27.911099 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c3973d0-2ae6-4999-bb44-bf6fb7496102" containerName="pruner" Oct 06 10:02:27 crc kubenswrapper[4824]: I1006 10:02:27.911107 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c3973d0-2ae6-4999-bb44-bf6fb7496102" containerName="pruner" Oct 06 10:02:27 crc kubenswrapper[4824]: E1006 10:02:27.911119 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="527511c8-2487-4416-a689-9ee161130297" containerName="registry-server" Oct 06 10:02:27 crc kubenswrapper[4824]: I1006 10:02:27.911130 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="527511c8-2487-4416-a689-9ee161130297" containerName="registry-server" Oct 06 10:02:27 crc kubenswrapper[4824]: E1006 10:02:27.911140 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3ef12c1-f3ad-4ead-aba0-a3901f80fc7f" containerName="extract-content" Oct 06 10:02:27 crc kubenswrapper[4824]: I1006 10:02:27.911149 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3ef12c1-f3ad-4ead-aba0-a3901f80fc7f" containerName="extract-content" Oct 06 10:02:27 crc kubenswrapper[4824]: E1006 10:02:27.911160 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c470b33a-a2cd-46c8-b092-57d11b4b39c7" containerName="registry-server" Oct 06 10:02:27 crc kubenswrapper[4824]: I1006 10:02:27.911168 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="c470b33a-a2cd-46c8-b092-57d11b4b39c7" containerName="registry-server" Oct 06 10:02:27 crc kubenswrapper[4824]: E1006 10:02:27.911179 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="527511c8-2487-4416-a689-9ee161130297" containerName="extract-content" Oct 06 10:02:27 crc kubenswrapper[4824]: I1006 10:02:27.911187 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="527511c8-2487-4416-a689-9ee161130297" containerName="extract-content" Oct 06 10:02:27 crc kubenswrapper[4824]: E1006 10:02:27.911202 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88764df0-c119-4e11-bd94-6818bc8ef83c" containerName="registry-server" Oct 06 10:02:27 crc kubenswrapper[4824]: I1006 10:02:27.911211 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="88764df0-c119-4e11-bd94-6818bc8ef83c" containerName="registry-server" Oct 06 10:02:27 crc kubenswrapper[4824]: E1006 10:02:27.911221 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3ef12c1-f3ad-4ead-aba0-a3901f80fc7f" containerName="registry-server" Oct 06 10:02:27 crc kubenswrapper[4824]: I1006 10:02:27.911229 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3ef12c1-f3ad-4ead-aba0-a3901f80fc7f" containerName="registry-server" Oct 06 10:02:27 crc kubenswrapper[4824]: E1006 10:02:27.911242 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c470b33a-a2cd-46c8-b092-57d11b4b39c7" containerName="extract-content" Oct 06 10:02:27 crc kubenswrapper[4824]: I1006 10:02:27.911251 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="c470b33a-a2cd-46c8-b092-57d11b4b39c7" containerName="extract-content" Oct 06 10:02:27 crc kubenswrapper[4824]: E1006 10:02:27.911266 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c470b33a-a2cd-46c8-b092-57d11b4b39c7" containerName="extract-utilities" Oct 06 10:02:27 crc kubenswrapper[4824]: I1006 10:02:27.911274 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="c470b33a-a2cd-46c8-b092-57d11b4b39c7" containerName="extract-utilities" Oct 06 10:02:27 crc kubenswrapper[4824]: E1006 10:02:27.911287 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af8fbf5a-f90e-4d22-ae65-05d8d8253308" containerName="oauth-openshift" Oct 06 10:02:27 crc kubenswrapper[4824]: I1006 10:02:27.911295 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="af8fbf5a-f90e-4d22-ae65-05d8d8253308" containerName="oauth-openshift" Oct 06 10:02:27 crc kubenswrapper[4824]: I1006 10:02:27.911417 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="da7d73fc-5908-4185-b203-854619742975" containerName="pruner" Oct 06 10:02:27 crc kubenswrapper[4824]: I1006 10:02:27.911431 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="f3ef12c1-f3ad-4ead-aba0-a3901f80fc7f" containerName="registry-server" Oct 06 10:02:27 crc kubenswrapper[4824]: I1006 10:02:27.911502 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="af8fbf5a-f90e-4d22-ae65-05d8d8253308" containerName="oauth-openshift" Oct 06 10:02:27 crc kubenswrapper[4824]: I1006 10:02:27.911521 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c3973d0-2ae6-4999-bb44-bf6fb7496102" containerName="pruner" Oct 06 10:02:27 crc kubenswrapper[4824]: I1006 10:02:27.911533 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="c470b33a-a2cd-46c8-b092-57d11b4b39c7" containerName="registry-server" Oct 06 10:02:27 crc kubenswrapper[4824]: I1006 10:02:27.911586 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="88764df0-c119-4e11-bd94-6818bc8ef83c" containerName="registry-server" Oct 06 10:02:27 crc kubenswrapper[4824]: I1006 10:02:27.911598 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="527511c8-2487-4416-a689-9ee161130297" containerName="registry-server" Oct 06 10:02:27 crc kubenswrapper[4824]: I1006 10:02:27.912161 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-6994c46f75-97hq2" Oct 06 10:02:27 crc kubenswrapper[4824]: I1006 10:02:27.926299 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-6994c46f75-97hq2"] Oct 06 10:02:27 crc kubenswrapper[4824]: I1006 10:02:27.950800 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/af8fbf5a-f90e-4d22-ae65-05d8d8253308-v4-0-config-user-template-provider-selection\") pod \"af8fbf5a-f90e-4d22-ae65-05d8d8253308\" (UID: \"af8fbf5a-f90e-4d22-ae65-05d8d8253308\") " Oct 06 10:02:27 crc kubenswrapper[4824]: I1006 10:02:27.950901 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/af8fbf5a-f90e-4d22-ae65-05d8d8253308-v4-0-config-system-trusted-ca-bundle\") pod \"af8fbf5a-f90e-4d22-ae65-05d8d8253308\" (UID: \"af8fbf5a-f90e-4d22-ae65-05d8d8253308\") " Oct 06 10:02:27 crc kubenswrapper[4824]: I1006 10:02:27.950947 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/af8fbf5a-f90e-4d22-ae65-05d8d8253308-v4-0-config-user-template-login\") pod \"af8fbf5a-f90e-4d22-ae65-05d8d8253308\" (UID: \"af8fbf5a-f90e-4d22-ae65-05d8d8253308\") " Oct 06 10:02:27 crc kubenswrapper[4824]: I1006 10:02:27.950996 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/af8fbf5a-f90e-4d22-ae65-05d8d8253308-audit-policies\") pod \"af8fbf5a-f90e-4d22-ae65-05d8d8253308\" (UID: \"af8fbf5a-f90e-4d22-ae65-05d8d8253308\") " Oct 06 10:02:27 crc kubenswrapper[4824]: I1006 10:02:27.951049 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/af8fbf5a-f90e-4d22-ae65-05d8d8253308-v4-0-config-user-template-error\") pod \"af8fbf5a-f90e-4d22-ae65-05d8d8253308\" (UID: \"af8fbf5a-f90e-4d22-ae65-05d8d8253308\") " Oct 06 10:02:27 crc kubenswrapper[4824]: I1006 10:02:27.951088 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4rsfz\" (UniqueName: \"kubernetes.io/projected/af8fbf5a-f90e-4d22-ae65-05d8d8253308-kube-api-access-4rsfz\") pod \"af8fbf5a-f90e-4d22-ae65-05d8d8253308\" (UID: \"af8fbf5a-f90e-4d22-ae65-05d8d8253308\") " Oct 06 10:02:27 crc kubenswrapper[4824]: I1006 10:02:27.951140 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/af8fbf5a-f90e-4d22-ae65-05d8d8253308-v4-0-config-user-idp-0-file-data\") pod \"af8fbf5a-f90e-4d22-ae65-05d8d8253308\" (UID: \"af8fbf5a-f90e-4d22-ae65-05d8d8253308\") " Oct 06 10:02:27 crc kubenswrapper[4824]: I1006 10:02:27.951178 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/af8fbf5a-f90e-4d22-ae65-05d8d8253308-v4-0-config-system-session\") pod \"af8fbf5a-f90e-4d22-ae65-05d8d8253308\" (UID: \"af8fbf5a-f90e-4d22-ae65-05d8d8253308\") " Oct 06 10:02:27 crc kubenswrapper[4824]: I1006 10:02:27.951228 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/af8fbf5a-f90e-4d22-ae65-05d8d8253308-v4-0-config-system-serving-cert\") pod \"af8fbf5a-f90e-4d22-ae65-05d8d8253308\" (UID: \"af8fbf5a-f90e-4d22-ae65-05d8d8253308\") " Oct 06 10:02:27 crc kubenswrapper[4824]: I1006 10:02:27.951275 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/af8fbf5a-f90e-4d22-ae65-05d8d8253308-v4-0-config-system-router-certs\") pod \"af8fbf5a-f90e-4d22-ae65-05d8d8253308\" (UID: \"af8fbf5a-f90e-4d22-ae65-05d8d8253308\") " Oct 06 10:02:27 crc kubenswrapper[4824]: I1006 10:02:27.951329 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/af8fbf5a-f90e-4d22-ae65-05d8d8253308-v4-0-config-system-service-ca\") pod \"af8fbf5a-f90e-4d22-ae65-05d8d8253308\" (UID: \"af8fbf5a-f90e-4d22-ae65-05d8d8253308\") " Oct 06 10:02:27 crc kubenswrapper[4824]: I1006 10:02:27.951366 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/af8fbf5a-f90e-4d22-ae65-05d8d8253308-audit-dir\") pod \"af8fbf5a-f90e-4d22-ae65-05d8d8253308\" (UID: \"af8fbf5a-f90e-4d22-ae65-05d8d8253308\") " Oct 06 10:02:27 crc kubenswrapper[4824]: I1006 10:02:27.951398 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/af8fbf5a-f90e-4d22-ae65-05d8d8253308-v4-0-config-system-ocp-branding-template\") pod \"af8fbf5a-f90e-4d22-ae65-05d8d8253308\" (UID: \"af8fbf5a-f90e-4d22-ae65-05d8d8253308\") " Oct 06 10:02:27 crc kubenswrapper[4824]: I1006 10:02:27.951448 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/af8fbf5a-f90e-4d22-ae65-05d8d8253308-v4-0-config-system-cliconfig\") pod \"af8fbf5a-f90e-4d22-ae65-05d8d8253308\" (UID: \"af8fbf5a-f90e-4d22-ae65-05d8d8253308\") " Oct 06 10:02:27 crc kubenswrapper[4824]: I1006 10:02:27.952264 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/af8fbf5a-f90e-4d22-ae65-05d8d8253308-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "af8fbf5a-f90e-4d22-ae65-05d8d8253308" (UID: "af8fbf5a-f90e-4d22-ae65-05d8d8253308"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 06 10:02:27 crc kubenswrapper[4824]: I1006 10:02:27.952561 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/af8fbf5a-f90e-4d22-ae65-05d8d8253308-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "af8fbf5a-f90e-4d22-ae65-05d8d8253308" (UID: "af8fbf5a-f90e-4d22-ae65-05d8d8253308"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:02:27 crc kubenswrapper[4824]: I1006 10:02:27.952713 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/af8fbf5a-f90e-4d22-ae65-05d8d8253308-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "af8fbf5a-f90e-4d22-ae65-05d8d8253308" (UID: "af8fbf5a-f90e-4d22-ae65-05d8d8253308"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:02:27 crc kubenswrapper[4824]: I1006 10:02:27.956544 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/af8fbf5a-f90e-4d22-ae65-05d8d8253308-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "af8fbf5a-f90e-4d22-ae65-05d8d8253308" (UID: "af8fbf5a-f90e-4d22-ae65-05d8d8253308"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:02:27 crc kubenswrapper[4824]: I1006 10:02:27.957158 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/af8fbf5a-f90e-4d22-ae65-05d8d8253308-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "af8fbf5a-f90e-4d22-ae65-05d8d8253308" (UID: "af8fbf5a-f90e-4d22-ae65-05d8d8253308"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:02:27 crc kubenswrapper[4824]: I1006 10:02:27.971537 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af8fbf5a-f90e-4d22-ae65-05d8d8253308-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "af8fbf5a-f90e-4d22-ae65-05d8d8253308" (UID: "af8fbf5a-f90e-4d22-ae65-05d8d8253308"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:02:27 crc kubenswrapper[4824]: I1006 10:02:27.971750 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/af8fbf5a-f90e-4d22-ae65-05d8d8253308-kube-api-access-4rsfz" (OuterVolumeSpecName: "kube-api-access-4rsfz") pod "af8fbf5a-f90e-4d22-ae65-05d8d8253308" (UID: "af8fbf5a-f90e-4d22-ae65-05d8d8253308"). InnerVolumeSpecName "kube-api-access-4rsfz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:02:27 crc kubenswrapper[4824]: I1006 10:02:27.971741 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af8fbf5a-f90e-4d22-ae65-05d8d8253308-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "af8fbf5a-f90e-4d22-ae65-05d8d8253308" (UID: "af8fbf5a-f90e-4d22-ae65-05d8d8253308"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:02:27 crc kubenswrapper[4824]: I1006 10:02:27.971934 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af8fbf5a-f90e-4d22-ae65-05d8d8253308-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "af8fbf5a-f90e-4d22-ae65-05d8d8253308" (UID: "af8fbf5a-f90e-4d22-ae65-05d8d8253308"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:02:27 crc kubenswrapper[4824]: I1006 10:02:27.972207 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af8fbf5a-f90e-4d22-ae65-05d8d8253308-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "af8fbf5a-f90e-4d22-ae65-05d8d8253308" (UID: "af8fbf5a-f90e-4d22-ae65-05d8d8253308"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:02:27 crc kubenswrapper[4824]: I1006 10:02:27.972733 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af8fbf5a-f90e-4d22-ae65-05d8d8253308-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "af8fbf5a-f90e-4d22-ae65-05d8d8253308" (UID: "af8fbf5a-f90e-4d22-ae65-05d8d8253308"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:02:27 crc kubenswrapper[4824]: I1006 10:02:27.976873 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af8fbf5a-f90e-4d22-ae65-05d8d8253308-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "af8fbf5a-f90e-4d22-ae65-05d8d8253308" (UID: "af8fbf5a-f90e-4d22-ae65-05d8d8253308"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:02:27 crc kubenswrapper[4824]: I1006 10:02:27.977073 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af8fbf5a-f90e-4d22-ae65-05d8d8253308-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "af8fbf5a-f90e-4d22-ae65-05d8d8253308" (UID: "af8fbf5a-f90e-4d22-ae65-05d8d8253308"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:02:27 crc kubenswrapper[4824]: I1006 10:02:27.977279 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af8fbf5a-f90e-4d22-ae65-05d8d8253308-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "af8fbf5a-f90e-4d22-ae65-05d8d8253308" (UID: "af8fbf5a-f90e-4d22-ae65-05d8d8253308"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:02:28 crc kubenswrapper[4824]: I1006 10:02:28.052359 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/d8b69c1d-a376-44da-bac2-72407a8373df-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-6994c46f75-97hq2\" (UID: \"d8b69c1d-a376-44da-bac2-72407a8373df\") " pod="openshift-authentication/oauth-openshift-6994c46f75-97hq2" Oct 06 10:02:28 crc kubenswrapper[4824]: I1006 10:02:28.052405 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/d8b69c1d-a376-44da-bac2-72407a8373df-v4-0-config-system-cliconfig\") pod \"oauth-openshift-6994c46f75-97hq2\" (UID: \"d8b69c1d-a376-44da-bac2-72407a8373df\") " pod="openshift-authentication/oauth-openshift-6994c46f75-97hq2" Oct 06 10:02:28 crc kubenswrapper[4824]: I1006 10:02:28.052432 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/d8b69c1d-a376-44da-bac2-72407a8373df-v4-0-config-system-serving-cert\") pod \"oauth-openshift-6994c46f75-97hq2\" (UID: \"d8b69c1d-a376-44da-bac2-72407a8373df\") " pod="openshift-authentication/oauth-openshift-6994c46f75-97hq2" Oct 06 10:02:28 crc kubenswrapper[4824]: I1006 10:02:28.052455 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/d8b69c1d-a376-44da-bac2-72407a8373df-audit-policies\") pod \"oauth-openshift-6994c46f75-97hq2\" (UID: \"d8b69c1d-a376-44da-bac2-72407a8373df\") " pod="openshift-authentication/oauth-openshift-6994c46f75-97hq2" Oct 06 10:02:28 crc kubenswrapper[4824]: I1006 10:02:28.052476 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cdnwd\" (UniqueName: \"kubernetes.io/projected/d8b69c1d-a376-44da-bac2-72407a8373df-kube-api-access-cdnwd\") pod \"oauth-openshift-6994c46f75-97hq2\" (UID: \"d8b69c1d-a376-44da-bac2-72407a8373df\") " pod="openshift-authentication/oauth-openshift-6994c46f75-97hq2" Oct 06 10:02:28 crc kubenswrapper[4824]: I1006 10:02:28.052493 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/d8b69c1d-a376-44da-bac2-72407a8373df-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-6994c46f75-97hq2\" (UID: \"d8b69c1d-a376-44da-bac2-72407a8373df\") " pod="openshift-authentication/oauth-openshift-6994c46f75-97hq2" Oct 06 10:02:28 crc kubenswrapper[4824]: I1006 10:02:28.052517 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/d8b69c1d-a376-44da-bac2-72407a8373df-audit-dir\") pod \"oauth-openshift-6994c46f75-97hq2\" (UID: \"d8b69c1d-a376-44da-bac2-72407a8373df\") " pod="openshift-authentication/oauth-openshift-6994c46f75-97hq2" Oct 06 10:02:28 crc kubenswrapper[4824]: I1006 10:02:28.052534 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/d8b69c1d-a376-44da-bac2-72407a8373df-v4-0-config-system-service-ca\") pod \"oauth-openshift-6994c46f75-97hq2\" (UID: \"d8b69c1d-a376-44da-bac2-72407a8373df\") " pod="openshift-authentication/oauth-openshift-6994c46f75-97hq2" Oct 06 10:02:28 crc kubenswrapper[4824]: I1006 10:02:28.052549 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/d8b69c1d-a376-44da-bac2-72407a8373df-v4-0-config-system-router-certs\") pod \"oauth-openshift-6994c46f75-97hq2\" (UID: \"d8b69c1d-a376-44da-bac2-72407a8373df\") " pod="openshift-authentication/oauth-openshift-6994c46f75-97hq2" Oct 06 10:02:28 crc kubenswrapper[4824]: I1006 10:02:28.052565 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/d8b69c1d-a376-44da-bac2-72407a8373df-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-6994c46f75-97hq2\" (UID: \"d8b69c1d-a376-44da-bac2-72407a8373df\") " pod="openshift-authentication/oauth-openshift-6994c46f75-97hq2" Oct 06 10:02:28 crc kubenswrapper[4824]: I1006 10:02:28.052584 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d8b69c1d-a376-44da-bac2-72407a8373df-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-6994c46f75-97hq2\" (UID: \"d8b69c1d-a376-44da-bac2-72407a8373df\") " pod="openshift-authentication/oauth-openshift-6994c46f75-97hq2" Oct 06 10:02:28 crc kubenswrapper[4824]: I1006 10:02:28.052608 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/d8b69c1d-a376-44da-bac2-72407a8373df-v4-0-config-user-template-login\") pod \"oauth-openshift-6994c46f75-97hq2\" (UID: \"d8b69c1d-a376-44da-bac2-72407a8373df\") " pod="openshift-authentication/oauth-openshift-6994c46f75-97hq2" Oct 06 10:02:28 crc kubenswrapper[4824]: I1006 10:02:28.052633 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/d8b69c1d-a376-44da-bac2-72407a8373df-v4-0-config-system-session\") pod \"oauth-openshift-6994c46f75-97hq2\" (UID: \"d8b69c1d-a376-44da-bac2-72407a8373df\") " pod="openshift-authentication/oauth-openshift-6994c46f75-97hq2" Oct 06 10:02:28 crc kubenswrapper[4824]: I1006 10:02:28.052652 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/d8b69c1d-a376-44da-bac2-72407a8373df-v4-0-config-user-template-error\") pod \"oauth-openshift-6994c46f75-97hq2\" (UID: \"d8b69c1d-a376-44da-bac2-72407a8373df\") " pod="openshift-authentication/oauth-openshift-6994c46f75-97hq2" Oct 06 10:02:28 crc kubenswrapper[4824]: I1006 10:02:28.052701 4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/af8fbf5a-f90e-4d22-ae65-05d8d8253308-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 10:02:28 crc kubenswrapper[4824]: I1006 10:02:28.052712 4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/af8fbf5a-f90e-4d22-ae65-05d8d8253308-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Oct 06 10:02:28 crc kubenswrapper[4824]: I1006 10:02:28.052722 4824 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/af8fbf5a-f90e-4d22-ae65-05d8d8253308-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 06 10:02:28 crc kubenswrapper[4824]: I1006 10:02:28.052733 4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/af8fbf5a-f90e-4d22-ae65-05d8d8253308-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Oct 06 10:02:28 crc kubenswrapper[4824]: I1006 10:02:28.052742 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4rsfz\" (UniqueName: \"kubernetes.io/projected/af8fbf5a-f90e-4d22-ae65-05d8d8253308-kube-api-access-4rsfz\") on node \"crc\" DevicePath \"\"" Oct 06 10:02:28 crc kubenswrapper[4824]: I1006 10:02:28.052752 4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/af8fbf5a-f90e-4d22-ae65-05d8d8253308-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Oct 06 10:02:28 crc kubenswrapper[4824]: I1006 10:02:28.052803 4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/af8fbf5a-f90e-4d22-ae65-05d8d8253308-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Oct 06 10:02:28 crc kubenswrapper[4824]: I1006 10:02:28.052815 4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/af8fbf5a-f90e-4d22-ae65-05d8d8253308-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 06 10:02:28 crc kubenswrapper[4824]: I1006 10:02:28.052824 4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/af8fbf5a-f90e-4d22-ae65-05d8d8253308-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Oct 06 10:02:28 crc kubenswrapper[4824]: I1006 10:02:28.052834 4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/af8fbf5a-f90e-4d22-ae65-05d8d8253308-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Oct 06 10:02:28 crc kubenswrapper[4824]: I1006 10:02:28.052843 4824 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/af8fbf5a-f90e-4d22-ae65-05d8d8253308-audit-dir\") on node \"crc\" DevicePath \"\"" Oct 06 10:02:28 crc kubenswrapper[4824]: I1006 10:02:28.052854 4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/af8fbf5a-f90e-4d22-ae65-05d8d8253308-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Oct 06 10:02:28 crc kubenswrapper[4824]: I1006 10:02:28.052864 4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/af8fbf5a-f90e-4d22-ae65-05d8d8253308-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Oct 06 10:02:28 crc kubenswrapper[4824]: I1006 10:02:28.052874 4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/af8fbf5a-f90e-4d22-ae65-05d8d8253308-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Oct 06 10:02:28 crc kubenswrapper[4824]: I1006 10:02:28.154227 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/d8b69c1d-a376-44da-bac2-72407a8373df-v4-0-config-user-template-login\") pod \"oauth-openshift-6994c46f75-97hq2\" (UID: \"d8b69c1d-a376-44da-bac2-72407a8373df\") " pod="openshift-authentication/oauth-openshift-6994c46f75-97hq2" Oct 06 10:02:28 crc kubenswrapper[4824]: I1006 10:02:28.154338 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/d8b69c1d-a376-44da-bac2-72407a8373df-v4-0-config-system-session\") pod \"oauth-openshift-6994c46f75-97hq2\" (UID: \"d8b69c1d-a376-44da-bac2-72407a8373df\") " pod="openshift-authentication/oauth-openshift-6994c46f75-97hq2" Oct 06 10:02:28 crc kubenswrapper[4824]: I1006 10:02:28.154393 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/d8b69c1d-a376-44da-bac2-72407a8373df-v4-0-config-user-template-error\") pod \"oauth-openshift-6994c46f75-97hq2\" (UID: \"d8b69c1d-a376-44da-bac2-72407a8373df\") " pod="openshift-authentication/oauth-openshift-6994c46f75-97hq2" Oct 06 10:02:28 crc kubenswrapper[4824]: I1006 10:02:28.154479 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/d8b69c1d-a376-44da-bac2-72407a8373df-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-6994c46f75-97hq2\" (UID: \"d8b69c1d-a376-44da-bac2-72407a8373df\") " pod="openshift-authentication/oauth-openshift-6994c46f75-97hq2" Oct 06 10:02:28 crc kubenswrapper[4824]: I1006 10:02:28.154516 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/d8b69c1d-a376-44da-bac2-72407a8373df-v4-0-config-system-cliconfig\") pod \"oauth-openshift-6994c46f75-97hq2\" (UID: \"d8b69c1d-a376-44da-bac2-72407a8373df\") " pod="openshift-authentication/oauth-openshift-6994c46f75-97hq2" Oct 06 10:02:28 crc kubenswrapper[4824]: I1006 10:02:28.154578 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/d8b69c1d-a376-44da-bac2-72407a8373df-v4-0-config-system-serving-cert\") pod \"oauth-openshift-6994c46f75-97hq2\" (UID: \"d8b69c1d-a376-44da-bac2-72407a8373df\") " pod="openshift-authentication/oauth-openshift-6994c46f75-97hq2" Oct 06 10:02:28 crc kubenswrapper[4824]: I1006 10:02:28.154624 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/d8b69c1d-a376-44da-bac2-72407a8373df-audit-policies\") pod \"oauth-openshift-6994c46f75-97hq2\" (UID: \"d8b69c1d-a376-44da-bac2-72407a8373df\") " pod="openshift-authentication/oauth-openshift-6994c46f75-97hq2" Oct 06 10:02:28 crc kubenswrapper[4824]: I1006 10:02:28.154663 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cdnwd\" (UniqueName: \"kubernetes.io/projected/d8b69c1d-a376-44da-bac2-72407a8373df-kube-api-access-cdnwd\") pod \"oauth-openshift-6994c46f75-97hq2\" (UID: \"d8b69c1d-a376-44da-bac2-72407a8373df\") " pod="openshift-authentication/oauth-openshift-6994c46f75-97hq2" Oct 06 10:02:28 crc kubenswrapper[4824]: I1006 10:02:28.154704 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/d8b69c1d-a376-44da-bac2-72407a8373df-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-6994c46f75-97hq2\" (UID: \"d8b69c1d-a376-44da-bac2-72407a8373df\") " pod="openshift-authentication/oauth-openshift-6994c46f75-97hq2" Oct 06 10:02:28 crc kubenswrapper[4824]: I1006 10:02:28.154752 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/d8b69c1d-a376-44da-bac2-72407a8373df-audit-dir\") pod \"oauth-openshift-6994c46f75-97hq2\" (UID: \"d8b69c1d-a376-44da-bac2-72407a8373df\") " pod="openshift-authentication/oauth-openshift-6994c46f75-97hq2" Oct 06 10:02:28 crc kubenswrapper[4824]: I1006 10:02:28.154785 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/d8b69c1d-a376-44da-bac2-72407a8373df-v4-0-config-system-service-ca\") pod \"oauth-openshift-6994c46f75-97hq2\" (UID: \"d8b69c1d-a376-44da-bac2-72407a8373df\") " pod="openshift-authentication/oauth-openshift-6994c46f75-97hq2" Oct 06 10:02:28 crc kubenswrapper[4824]: I1006 10:02:28.154825 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/d8b69c1d-a376-44da-bac2-72407a8373df-v4-0-config-system-router-certs\") pod \"oauth-openshift-6994c46f75-97hq2\" (UID: \"d8b69c1d-a376-44da-bac2-72407a8373df\") " pod="openshift-authentication/oauth-openshift-6994c46f75-97hq2" Oct 06 10:02:28 crc kubenswrapper[4824]: I1006 10:02:28.154864 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/d8b69c1d-a376-44da-bac2-72407a8373df-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-6994c46f75-97hq2\" (UID: \"d8b69c1d-a376-44da-bac2-72407a8373df\") " pod="openshift-authentication/oauth-openshift-6994c46f75-97hq2" Oct 06 10:02:28 crc kubenswrapper[4824]: I1006 10:02:28.154903 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d8b69c1d-a376-44da-bac2-72407a8373df-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-6994c46f75-97hq2\" (UID: \"d8b69c1d-a376-44da-bac2-72407a8373df\") " pod="openshift-authentication/oauth-openshift-6994c46f75-97hq2" Oct 06 10:02:28 crc kubenswrapper[4824]: I1006 10:02:28.155567 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/d8b69c1d-a376-44da-bac2-72407a8373df-audit-dir\") pod \"oauth-openshift-6994c46f75-97hq2\" (UID: \"d8b69c1d-a376-44da-bac2-72407a8373df\") " pod="openshift-authentication/oauth-openshift-6994c46f75-97hq2" Oct 06 10:02:28 crc kubenswrapper[4824]: I1006 10:02:28.157152 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/d8b69c1d-a376-44da-bac2-72407a8373df-v4-0-config-system-cliconfig\") pod \"oauth-openshift-6994c46f75-97hq2\" (UID: \"d8b69c1d-a376-44da-bac2-72407a8373df\") " pod="openshift-authentication/oauth-openshift-6994c46f75-97hq2" Oct 06 10:02:28 crc kubenswrapper[4824]: I1006 10:02:28.157319 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/d8b69c1d-a376-44da-bac2-72407a8373df-v4-0-config-system-service-ca\") pod \"oauth-openshift-6994c46f75-97hq2\" (UID: \"d8b69c1d-a376-44da-bac2-72407a8373df\") " pod="openshift-authentication/oauth-openshift-6994c46f75-97hq2" Oct 06 10:02:28 crc kubenswrapper[4824]: I1006 10:02:28.157632 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/d8b69c1d-a376-44da-bac2-72407a8373df-audit-policies\") pod \"oauth-openshift-6994c46f75-97hq2\" (UID: \"d8b69c1d-a376-44da-bac2-72407a8373df\") " pod="openshift-authentication/oauth-openshift-6994c46f75-97hq2" Oct 06 10:02:28 crc kubenswrapper[4824]: I1006 10:02:28.158653 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d8b69c1d-a376-44da-bac2-72407a8373df-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-6994c46f75-97hq2\" (UID: \"d8b69c1d-a376-44da-bac2-72407a8373df\") " pod="openshift-authentication/oauth-openshift-6994c46f75-97hq2" Oct 06 10:02:28 crc kubenswrapper[4824]: I1006 10:02:28.161241 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/d8b69c1d-a376-44da-bac2-72407a8373df-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-6994c46f75-97hq2\" (UID: \"d8b69c1d-a376-44da-bac2-72407a8373df\") " pod="openshift-authentication/oauth-openshift-6994c46f75-97hq2" Oct 06 10:02:28 crc kubenswrapper[4824]: I1006 10:02:28.161929 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/d8b69c1d-a376-44da-bac2-72407a8373df-v4-0-config-system-session\") pod \"oauth-openshift-6994c46f75-97hq2\" (UID: \"d8b69c1d-a376-44da-bac2-72407a8373df\") " pod="openshift-authentication/oauth-openshift-6994c46f75-97hq2" Oct 06 10:02:28 crc kubenswrapper[4824]: I1006 10:02:28.161915 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/d8b69c1d-a376-44da-bac2-72407a8373df-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-6994c46f75-97hq2\" (UID: \"d8b69c1d-a376-44da-bac2-72407a8373df\") " pod="openshift-authentication/oauth-openshift-6994c46f75-97hq2" Oct 06 10:02:28 crc kubenswrapper[4824]: I1006 10:02:28.162116 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/d8b69c1d-a376-44da-bac2-72407a8373df-v4-0-config-user-template-error\") pod \"oauth-openshift-6994c46f75-97hq2\" (UID: \"d8b69c1d-a376-44da-bac2-72407a8373df\") " pod="openshift-authentication/oauth-openshift-6994c46f75-97hq2" Oct 06 10:02:28 crc kubenswrapper[4824]: I1006 10:02:28.162696 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/d8b69c1d-a376-44da-bac2-72407a8373df-v4-0-config-user-template-login\") pod \"oauth-openshift-6994c46f75-97hq2\" (UID: \"d8b69c1d-a376-44da-bac2-72407a8373df\") " pod="openshift-authentication/oauth-openshift-6994c46f75-97hq2" Oct 06 10:02:28 crc kubenswrapper[4824]: I1006 10:02:28.163483 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/d8b69c1d-a376-44da-bac2-72407a8373df-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-6994c46f75-97hq2\" (UID: \"d8b69c1d-a376-44da-bac2-72407a8373df\") " pod="openshift-authentication/oauth-openshift-6994c46f75-97hq2" Oct 06 10:02:28 crc kubenswrapper[4824]: I1006 10:02:28.164036 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/d8b69c1d-a376-44da-bac2-72407a8373df-v4-0-config-system-router-certs\") pod \"oauth-openshift-6994c46f75-97hq2\" (UID: \"d8b69c1d-a376-44da-bac2-72407a8373df\") " pod="openshift-authentication/oauth-openshift-6994c46f75-97hq2" Oct 06 10:02:28 crc kubenswrapper[4824]: I1006 10:02:28.166040 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/d8b69c1d-a376-44da-bac2-72407a8373df-v4-0-config-system-serving-cert\") pod \"oauth-openshift-6994c46f75-97hq2\" (UID: \"d8b69c1d-a376-44da-bac2-72407a8373df\") " pod="openshift-authentication/oauth-openshift-6994c46f75-97hq2" Oct 06 10:02:28 crc kubenswrapper[4824]: I1006 10:02:28.182934 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cdnwd\" (UniqueName: \"kubernetes.io/projected/d8b69c1d-a376-44da-bac2-72407a8373df-kube-api-access-cdnwd\") pod \"oauth-openshift-6994c46f75-97hq2\" (UID: \"d8b69c1d-a376-44da-bac2-72407a8373df\") " pod="openshift-authentication/oauth-openshift-6994c46f75-97hq2" Oct 06 10:02:28 crc kubenswrapper[4824]: I1006 10:02:28.231931 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-6994c46f75-97hq2" Oct 06 10:02:28 crc kubenswrapper[4824]: I1006 10:02:28.498138 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-6994c46f75-97hq2"] Oct 06 10:02:28 crc kubenswrapper[4824]: I1006 10:02:28.531964 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-8z68k" Oct 06 10:02:28 crc kubenswrapper[4824]: I1006 10:02:28.532022 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-8z68k" event={"ID":"af8fbf5a-f90e-4d22-ae65-05d8d8253308","Type":"ContainerDied","Data":"039afdf33c495c5167b8f389b7d84c3a275a2fd6aa47bc9c99e0195359cafe51"} Oct 06 10:02:28 crc kubenswrapper[4824]: I1006 10:02:28.532515 4824 scope.go:117] "RemoveContainer" containerID="119cba5ffd73d0ff173f47322893c5b169be0630ac8c94dbbc819afb9ec746eb" Oct 06 10:02:28 crc kubenswrapper[4824]: I1006 10:02:28.533426 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-6994c46f75-97hq2" event={"ID":"d8b69c1d-a376-44da-bac2-72407a8373df","Type":"ContainerStarted","Data":"794c9942b3f027c4a5db5e336668153a57b7368ac7d5feebba23c74ec3562956"} Oct 06 10:02:28 crc kubenswrapper[4824]: I1006 10:02:28.591598 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-8z68k"] Oct 06 10:02:28 crc kubenswrapper[4824]: I1006 10:02:28.594057 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-8z68k"] Oct 06 10:02:29 crc kubenswrapper[4824]: I1006 10:02:29.287030 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="af8fbf5a-f90e-4d22-ae65-05d8d8253308" path="/var/lib/kubelet/pods/af8fbf5a-f90e-4d22-ae65-05d8d8253308/volumes" Oct 06 10:02:29 crc kubenswrapper[4824]: I1006 10:02:29.544577 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-6994c46f75-97hq2" event={"ID":"d8b69c1d-a376-44da-bac2-72407a8373df","Type":"ContainerStarted","Data":"f485f6b5de1c618df6340623c2e547fac8a1ace939a46b757bce9ca33e2595b8"} Oct 06 10:02:29 crc kubenswrapper[4824]: I1006 10:02:29.544922 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-6994c46f75-97hq2" Oct 06 10:02:29 crc kubenswrapper[4824]: I1006 10:02:29.550716 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-6994c46f75-97hq2" Oct 06 10:02:29 crc kubenswrapper[4824]: I1006 10:02:29.626783 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-6994c46f75-97hq2" podStartSLOduration=27.626675141 podStartE2EDuration="27.626675141s" podCreationTimestamp="2025-10-06 10:02:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:02:29.597022923 +0000 UTC m=+258.961445824" watchObservedRunningTime="2025-10-06 10:02:29.626675141 +0000 UTC m=+258.991098052" Oct 06 10:02:50 crc kubenswrapper[4824]: I1006 10:02:50.101834 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-r52zf"] Oct 06 10:02:50 crc kubenswrapper[4824]: I1006 10:02:50.102844 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-r52zf" podUID="06d3772b-e467-41a1-bf9b-9d6520ab68ef" containerName="registry-server" containerID="cri-o://c13a247642fda90779a68772ce9c3288d749e24fea28b15a22612cd488351a46" gracePeriod=30 Oct 06 10:02:50 crc kubenswrapper[4824]: I1006 10:02:50.123318 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-tlvv7"] Oct 06 10:02:50 crc kubenswrapper[4824]: I1006 10:02:50.123639 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-tlvv7" podUID="030ddedc-a25b-47f6-88b6-c1286dc1c910" containerName="registry-server" containerID="cri-o://853c3143b0b0b58e6bebb8466a6f31993342218020b663e0e381e288cc558940" gracePeriod=30 Oct 06 10:02:50 crc kubenswrapper[4824]: I1006 10:02:50.134531 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-qq4zz"] Oct 06 10:02:50 crc kubenswrapper[4824]: I1006 10:02:50.134940 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-qq4zz" podUID="d2595750-aae8-4f17-8782-b7eccb0c2948" containerName="marketplace-operator" containerID="cri-o://d27b1d97570bdd6569bbf1dc7b40419c3ce223f7aac865ed7a1b5c65b7acfcaa" gracePeriod=30 Oct 06 10:02:50 crc kubenswrapper[4824]: I1006 10:02:50.139787 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-bgbpb"] Oct 06 10:02:50 crc kubenswrapper[4824]: I1006 10:02:50.140095 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-bgbpb" podUID="70324ded-62d7-47fc-a67a-110ecde1497c" containerName="registry-server" containerID="cri-o://e4db6b7d450dc54a789cbfe6f13124becd9116d29f9d07848dbeb547dab40be5" gracePeriod=30 Oct 06 10:02:50 crc kubenswrapper[4824]: I1006 10:02:50.148624 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-5jqs9"] Oct 06 10:02:50 crc kubenswrapper[4824]: I1006 10:02:50.148949 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-5jqs9" podUID="c79214ab-ae6c-444b-bd4b-072c9a8bd37c" containerName="registry-server" containerID="cri-o://310ea3555fe50ef7415a359fc4b16c58fc944b5e04f0a977f12d6807fb9eb330" gracePeriod=30 Oct 06 10:02:50 crc kubenswrapper[4824]: I1006 10:02:50.159787 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-db6vn"] Oct 06 10:02:50 crc kubenswrapper[4824]: I1006 10:02:50.160826 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-db6vn" Oct 06 10:02:50 crc kubenswrapper[4824]: I1006 10:02:50.179015 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-db6vn"] Oct 06 10:02:50 crc kubenswrapper[4824]: I1006 10:02:50.297582 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a3a514b8-f221-465e-bbb2-3c88bcb4f622-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-db6vn\" (UID: \"a3a514b8-f221-465e-bbb2-3c88bcb4f622\") " pod="openshift-marketplace/marketplace-operator-79b997595-db6vn" Oct 06 10:02:50 crc kubenswrapper[4824]: I1006 10:02:50.297669 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/a3a514b8-f221-465e-bbb2-3c88bcb4f622-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-db6vn\" (UID: \"a3a514b8-f221-465e-bbb2-3c88bcb4f622\") " pod="openshift-marketplace/marketplace-operator-79b997595-db6vn" Oct 06 10:02:50 crc kubenswrapper[4824]: I1006 10:02:50.297696 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k62ln\" (UniqueName: \"kubernetes.io/projected/a3a514b8-f221-465e-bbb2-3c88bcb4f622-kube-api-access-k62ln\") pod \"marketplace-operator-79b997595-db6vn\" (UID: \"a3a514b8-f221-465e-bbb2-3c88bcb4f622\") " pod="openshift-marketplace/marketplace-operator-79b997595-db6vn" Oct 06 10:02:50 crc kubenswrapper[4824]: I1006 10:02:50.399103 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a3a514b8-f221-465e-bbb2-3c88bcb4f622-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-db6vn\" (UID: \"a3a514b8-f221-465e-bbb2-3c88bcb4f622\") " pod="openshift-marketplace/marketplace-operator-79b997595-db6vn" Oct 06 10:02:50 crc kubenswrapper[4824]: I1006 10:02:50.399156 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/a3a514b8-f221-465e-bbb2-3c88bcb4f622-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-db6vn\" (UID: \"a3a514b8-f221-465e-bbb2-3c88bcb4f622\") " pod="openshift-marketplace/marketplace-operator-79b997595-db6vn" Oct 06 10:02:50 crc kubenswrapper[4824]: I1006 10:02:50.399177 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k62ln\" (UniqueName: \"kubernetes.io/projected/a3a514b8-f221-465e-bbb2-3c88bcb4f622-kube-api-access-k62ln\") pod \"marketplace-operator-79b997595-db6vn\" (UID: \"a3a514b8-f221-465e-bbb2-3c88bcb4f622\") " pod="openshift-marketplace/marketplace-operator-79b997595-db6vn" Oct 06 10:02:50 crc kubenswrapper[4824]: I1006 10:02:50.400494 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a3a514b8-f221-465e-bbb2-3c88bcb4f622-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-db6vn\" (UID: \"a3a514b8-f221-465e-bbb2-3c88bcb4f622\") " pod="openshift-marketplace/marketplace-operator-79b997595-db6vn" Oct 06 10:02:50 crc kubenswrapper[4824]: I1006 10:02:50.411619 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/a3a514b8-f221-465e-bbb2-3c88bcb4f622-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-db6vn\" (UID: \"a3a514b8-f221-465e-bbb2-3c88bcb4f622\") " pod="openshift-marketplace/marketplace-operator-79b997595-db6vn" Oct 06 10:02:50 crc kubenswrapper[4824]: I1006 10:02:50.417558 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k62ln\" (UniqueName: \"kubernetes.io/projected/a3a514b8-f221-465e-bbb2-3c88bcb4f622-kube-api-access-k62ln\") pod \"marketplace-operator-79b997595-db6vn\" (UID: \"a3a514b8-f221-465e-bbb2-3c88bcb4f622\") " pod="openshift-marketplace/marketplace-operator-79b997595-db6vn" Oct 06 10:02:50 crc kubenswrapper[4824]: I1006 10:02:50.480919 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-db6vn" Oct 06 10:02:50 crc kubenswrapper[4824]: E1006 10:02:50.538263 4824 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of e4db6b7d450dc54a789cbfe6f13124becd9116d29f9d07848dbeb547dab40be5 is running failed: container process not found" containerID="e4db6b7d450dc54a789cbfe6f13124becd9116d29f9d07848dbeb547dab40be5" cmd=["grpc_health_probe","-addr=:50051"] Oct 06 10:02:50 crc kubenswrapper[4824]: E1006 10:02:50.539024 4824 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of e4db6b7d450dc54a789cbfe6f13124becd9116d29f9d07848dbeb547dab40be5 is running failed: container process not found" containerID="e4db6b7d450dc54a789cbfe6f13124becd9116d29f9d07848dbeb547dab40be5" cmd=["grpc_health_probe","-addr=:50051"] Oct 06 10:02:50 crc kubenswrapper[4824]: E1006 10:02:50.539702 4824 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of e4db6b7d450dc54a789cbfe6f13124becd9116d29f9d07848dbeb547dab40be5 is running failed: container process not found" containerID="e4db6b7d450dc54a789cbfe6f13124becd9116d29f9d07848dbeb547dab40be5" cmd=["grpc_health_probe","-addr=:50051"] Oct 06 10:02:50 crc kubenswrapper[4824]: E1006 10:02:50.539918 4824 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of e4db6b7d450dc54a789cbfe6f13124becd9116d29f9d07848dbeb547dab40be5 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/redhat-marketplace-bgbpb" podUID="70324ded-62d7-47fc-a67a-110ecde1497c" containerName="registry-server" Oct 06 10:02:50 crc kubenswrapper[4824]: I1006 10:02:50.593480 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-r52zf" Oct 06 10:02:50 crc kubenswrapper[4824]: I1006 10:02:50.634533 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tlvv7" Oct 06 10:02:50 crc kubenswrapper[4824]: I1006 10:02:50.642052 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5jqs9" Oct 06 10:02:50 crc kubenswrapper[4824]: I1006 10:02:50.684207 4824 generic.go:334] "Generic (PLEG): container finished" podID="d2595750-aae8-4f17-8782-b7eccb0c2948" containerID="d27b1d97570bdd6569bbf1dc7b40419c3ce223f7aac865ed7a1b5c65b7acfcaa" exitCode=0 Oct 06 10:02:50 crc kubenswrapper[4824]: I1006 10:02:50.684317 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-qq4zz" event={"ID":"d2595750-aae8-4f17-8782-b7eccb0c2948","Type":"ContainerDied","Data":"d27b1d97570bdd6569bbf1dc7b40419c3ce223f7aac865ed7a1b5c65b7acfcaa"} Oct 06 10:02:50 crc kubenswrapper[4824]: I1006 10:02:50.688116 4824 generic.go:334] "Generic (PLEG): container finished" podID="70324ded-62d7-47fc-a67a-110ecde1497c" containerID="e4db6b7d450dc54a789cbfe6f13124becd9116d29f9d07848dbeb547dab40be5" exitCode=0 Oct 06 10:02:50 crc kubenswrapper[4824]: I1006 10:02:50.688191 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bgbpb" event={"ID":"70324ded-62d7-47fc-a67a-110ecde1497c","Type":"ContainerDied","Data":"e4db6b7d450dc54a789cbfe6f13124becd9116d29f9d07848dbeb547dab40be5"} Oct 06 10:02:50 crc kubenswrapper[4824]: I1006 10:02:50.693049 4824 generic.go:334] "Generic (PLEG): container finished" podID="c79214ab-ae6c-444b-bd4b-072c9a8bd37c" containerID="310ea3555fe50ef7415a359fc4b16c58fc944b5e04f0a977f12d6807fb9eb330" exitCode=0 Oct 06 10:02:50 crc kubenswrapper[4824]: I1006 10:02:50.693149 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5jqs9" Oct 06 10:02:50 crc kubenswrapper[4824]: I1006 10:02:50.693180 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5jqs9" event={"ID":"c79214ab-ae6c-444b-bd4b-072c9a8bd37c","Type":"ContainerDied","Data":"310ea3555fe50ef7415a359fc4b16c58fc944b5e04f0a977f12d6807fb9eb330"} Oct 06 10:02:50 crc kubenswrapper[4824]: I1006 10:02:50.693333 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5jqs9" event={"ID":"c79214ab-ae6c-444b-bd4b-072c9a8bd37c","Type":"ContainerDied","Data":"a3d850b07f8ea0a2aad69710b67a0482fda2fa6524662838b582f01547234b46"} Oct 06 10:02:50 crc kubenswrapper[4824]: I1006 10:02:50.693392 4824 scope.go:117] "RemoveContainer" containerID="310ea3555fe50ef7415a359fc4b16c58fc944b5e04f0a977f12d6807fb9eb330" Oct 06 10:02:50 crc kubenswrapper[4824]: I1006 10:02:50.697883 4824 generic.go:334] "Generic (PLEG): container finished" podID="06d3772b-e467-41a1-bf9b-9d6520ab68ef" containerID="c13a247642fda90779a68772ce9c3288d749e24fea28b15a22612cd488351a46" exitCode=0 Oct 06 10:02:50 crc kubenswrapper[4824]: I1006 10:02:50.697945 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r52zf" event={"ID":"06d3772b-e467-41a1-bf9b-9d6520ab68ef","Type":"ContainerDied","Data":"c13a247642fda90779a68772ce9c3288d749e24fea28b15a22612cd488351a46"} Oct 06 10:02:50 crc kubenswrapper[4824]: I1006 10:02:50.697998 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r52zf" event={"ID":"06d3772b-e467-41a1-bf9b-9d6520ab68ef","Type":"ContainerDied","Data":"8f6bfad13da03714cd8ccfb513f43be501d6fe2cb2edbb314a684951d06f5ce2"} Oct 06 10:02:50 crc kubenswrapper[4824]: I1006 10:02:50.698569 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-r52zf" Oct 06 10:02:50 crc kubenswrapper[4824]: I1006 10:02:50.699497 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-qq4zz" Oct 06 10:02:50 crc kubenswrapper[4824]: I1006 10:02:50.701367 4824 generic.go:334] "Generic (PLEG): container finished" podID="030ddedc-a25b-47f6-88b6-c1286dc1c910" containerID="853c3143b0b0b58e6bebb8466a6f31993342218020b663e0e381e288cc558940" exitCode=0 Oct 06 10:02:50 crc kubenswrapper[4824]: I1006 10:02:50.701404 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tlvv7" event={"ID":"030ddedc-a25b-47f6-88b6-c1286dc1c910","Type":"ContainerDied","Data":"853c3143b0b0b58e6bebb8466a6f31993342218020b663e0e381e288cc558940"} Oct 06 10:02:50 crc kubenswrapper[4824]: I1006 10:02:50.701425 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tlvv7" event={"ID":"030ddedc-a25b-47f6-88b6-c1286dc1c910","Type":"ContainerDied","Data":"b2cb07f58394a5ac75de56feaf694f27a1427db2fce7a24ff16b91d2de681dce"} Oct 06 10:02:50 crc kubenswrapper[4824]: I1006 10:02:50.701476 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tlvv7" Oct 06 10:02:50 crc kubenswrapper[4824]: I1006 10:02:50.702304 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c79214ab-ae6c-444b-bd4b-072c9a8bd37c-catalog-content\") pod \"c79214ab-ae6c-444b-bd4b-072c9a8bd37c\" (UID: \"c79214ab-ae6c-444b-bd4b-072c9a8bd37c\") " Oct 06 10:02:50 crc kubenswrapper[4824]: I1006 10:02:50.702355 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-975sb\" (UniqueName: \"kubernetes.io/projected/030ddedc-a25b-47f6-88b6-c1286dc1c910-kube-api-access-975sb\") pod \"030ddedc-a25b-47f6-88b6-c1286dc1c910\" (UID: \"030ddedc-a25b-47f6-88b6-c1286dc1c910\") " Oct 06 10:02:50 crc kubenswrapper[4824]: I1006 10:02:50.702395 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4h5wb\" (UniqueName: \"kubernetes.io/projected/c79214ab-ae6c-444b-bd4b-072c9a8bd37c-kube-api-access-4h5wb\") pod \"c79214ab-ae6c-444b-bd4b-072c9a8bd37c\" (UID: \"c79214ab-ae6c-444b-bd4b-072c9a8bd37c\") " Oct 06 10:02:50 crc kubenswrapper[4824]: I1006 10:02:50.702433 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/06d3772b-e467-41a1-bf9b-9d6520ab68ef-catalog-content\") pod \"06d3772b-e467-41a1-bf9b-9d6520ab68ef\" (UID: \"06d3772b-e467-41a1-bf9b-9d6520ab68ef\") " Oct 06 10:02:50 crc kubenswrapper[4824]: I1006 10:02:50.702471 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/06d3772b-e467-41a1-bf9b-9d6520ab68ef-utilities\") pod \"06d3772b-e467-41a1-bf9b-9d6520ab68ef\" (UID: \"06d3772b-e467-41a1-bf9b-9d6520ab68ef\") " Oct 06 10:02:50 crc kubenswrapper[4824]: I1006 10:02:50.702499 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fhs24\" (UniqueName: \"kubernetes.io/projected/06d3772b-e467-41a1-bf9b-9d6520ab68ef-kube-api-access-fhs24\") pod \"06d3772b-e467-41a1-bf9b-9d6520ab68ef\" (UID: \"06d3772b-e467-41a1-bf9b-9d6520ab68ef\") " Oct 06 10:02:50 crc kubenswrapper[4824]: I1006 10:02:50.702530 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/030ddedc-a25b-47f6-88b6-c1286dc1c910-utilities\") pod \"030ddedc-a25b-47f6-88b6-c1286dc1c910\" (UID: \"030ddedc-a25b-47f6-88b6-c1286dc1c910\") " Oct 06 10:02:50 crc kubenswrapper[4824]: I1006 10:02:50.702573 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c79214ab-ae6c-444b-bd4b-072c9a8bd37c-utilities\") pod \"c79214ab-ae6c-444b-bd4b-072c9a8bd37c\" (UID: \"c79214ab-ae6c-444b-bd4b-072c9a8bd37c\") " Oct 06 10:02:50 crc kubenswrapper[4824]: I1006 10:02:50.702613 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/030ddedc-a25b-47f6-88b6-c1286dc1c910-catalog-content\") pod \"030ddedc-a25b-47f6-88b6-c1286dc1c910\" (UID: \"030ddedc-a25b-47f6-88b6-c1286dc1c910\") " Oct 06 10:02:50 crc kubenswrapper[4824]: I1006 10:02:50.705450 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/030ddedc-a25b-47f6-88b6-c1286dc1c910-utilities" (OuterVolumeSpecName: "utilities") pod "030ddedc-a25b-47f6-88b6-c1286dc1c910" (UID: "030ddedc-a25b-47f6-88b6-c1286dc1c910"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:02:50 crc kubenswrapper[4824]: I1006 10:02:50.705596 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/06d3772b-e467-41a1-bf9b-9d6520ab68ef-utilities" (OuterVolumeSpecName: "utilities") pod "06d3772b-e467-41a1-bf9b-9d6520ab68ef" (UID: "06d3772b-e467-41a1-bf9b-9d6520ab68ef"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:02:50 crc kubenswrapper[4824]: I1006 10:02:50.705788 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c79214ab-ae6c-444b-bd4b-072c9a8bd37c-utilities" (OuterVolumeSpecName: "utilities") pod "c79214ab-ae6c-444b-bd4b-072c9a8bd37c" (UID: "c79214ab-ae6c-444b-bd4b-072c9a8bd37c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:02:50 crc kubenswrapper[4824]: I1006 10:02:50.710231 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/06d3772b-e467-41a1-bf9b-9d6520ab68ef-kube-api-access-fhs24" (OuterVolumeSpecName: "kube-api-access-fhs24") pod "06d3772b-e467-41a1-bf9b-9d6520ab68ef" (UID: "06d3772b-e467-41a1-bf9b-9d6520ab68ef"). InnerVolumeSpecName "kube-api-access-fhs24". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:02:50 crc kubenswrapper[4824]: I1006 10:02:50.714014 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c79214ab-ae6c-444b-bd4b-072c9a8bd37c-kube-api-access-4h5wb" (OuterVolumeSpecName: "kube-api-access-4h5wb") pod "c79214ab-ae6c-444b-bd4b-072c9a8bd37c" (UID: "c79214ab-ae6c-444b-bd4b-072c9a8bd37c"). InnerVolumeSpecName "kube-api-access-4h5wb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:02:50 crc kubenswrapper[4824]: I1006 10:02:50.715706 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/030ddedc-a25b-47f6-88b6-c1286dc1c910-kube-api-access-975sb" (OuterVolumeSpecName: "kube-api-access-975sb") pod "030ddedc-a25b-47f6-88b6-c1286dc1c910" (UID: "030ddedc-a25b-47f6-88b6-c1286dc1c910"). InnerVolumeSpecName "kube-api-access-975sb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:02:50 crc kubenswrapper[4824]: I1006 10:02:50.727000 4824 scope.go:117] "RemoveContainer" containerID="38f09ee01c275fb5b6c01f374e7ddad24fe9ee9659e3cd9f0b5373327c8892fc" Oct 06 10:02:50 crc kubenswrapper[4824]: I1006 10:02:50.757593 4824 scope.go:117] "RemoveContainer" containerID="1ce8650bd8f98da7ce898f5bcd75ab5e21f63637a057f7de6cd1f89b8cb28fa4" Oct 06 10:02:50 crc kubenswrapper[4824]: I1006 10:02:50.774301 4824 scope.go:117] "RemoveContainer" containerID="310ea3555fe50ef7415a359fc4b16c58fc944b5e04f0a977f12d6807fb9eb330" Oct 06 10:02:50 crc kubenswrapper[4824]: E1006 10:02:50.775125 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"310ea3555fe50ef7415a359fc4b16c58fc944b5e04f0a977f12d6807fb9eb330\": container with ID starting with 310ea3555fe50ef7415a359fc4b16c58fc944b5e04f0a977f12d6807fb9eb330 not found: ID does not exist" containerID="310ea3555fe50ef7415a359fc4b16c58fc944b5e04f0a977f12d6807fb9eb330" Oct 06 10:02:50 crc kubenswrapper[4824]: I1006 10:02:50.775185 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"310ea3555fe50ef7415a359fc4b16c58fc944b5e04f0a977f12d6807fb9eb330"} err="failed to get container status \"310ea3555fe50ef7415a359fc4b16c58fc944b5e04f0a977f12d6807fb9eb330\": rpc error: code = NotFound desc = could not find container \"310ea3555fe50ef7415a359fc4b16c58fc944b5e04f0a977f12d6807fb9eb330\": container with ID starting with 310ea3555fe50ef7415a359fc4b16c58fc944b5e04f0a977f12d6807fb9eb330 not found: ID does not exist" Oct 06 10:02:50 crc kubenswrapper[4824]: I1006 10:02:50.775226 4824 scope.go:117] "RemoveContainer" containerID="38f09ee01c275fb5b6c01f374e7ddad24fe9ee9659e3cd9f0b5373327c8892fc" Oct 06 10:02:50 crc kubenswrapper[4824]: E1006 10:02:50.775632 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"38f09ee01c275fb5b6c01f374e7ddad24fe9ee9659e3cd9f0b5373327c8892fc\": container with ID starting with 38f09ee01c275fb5b6c01f374e7ddad24fe9ee9659e3cd9f0b5373327c8892fc not found: ID does not exist" containerID="38f09ee01c275fb5b6c01f374e7ddad24fe9ee9659e3cd9f0b5373327c8892fc" Oct 06 10:02:50 crc kubenswrapper[4824]: I1006 10:02:50.775673 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"38f09ee01c275fb5b6c01f374e7ddad24fe9ee9659e3cd9f0b5373327c8892fc"} err="failed to get container status \"38f09ee01c275fb5b6c01f374e7ddad24fe9ee9659e3cd9f0b5373327c8892fc\": rpc error: code = NotFound desc = could not find container \"38f09ee01c275fb5b6c01f374e7ddad24fe9ee9659e3cd9f0b5373327c8892fc\": container with ID starting with 38f09ee01c275fb5b6c01f374e7ddad24fe9ee9659e3cd9f0b5373327c8892fc not found: ID does not exist" Oct 06 10:02:50 crc kubenswrapper[4824]: I1006 10:02:50.775695 4824 scope.go:117] "RemoveContainer" containerID="1ce8650bd8f98da7ce898f5bcd75ab5e21f63637a057f7de6cd1f89b8cb28fa4" Oct 06 10:02:50 crc kubenswrapper[4824]: E1006 10:02:50.776026 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1ce8650bd8f98da7ce898f5bcd75ab5e21f63637a057f7de6cd1f89b8cb28fa4\": container with ID starting with 1ce8650bd8f98da7ce898f5bcd75ab5e21f63637a057f7de6cd1f89b8cb28fa4 not found: ID does not exist" containerID="1ce8650bd8f98da7ce898f5bcd75ab5e21f63637a057f7de6cd1f89b8cb28fa4" Oct 06 10:02:50 crc kubenswrapper[4824]: I1006 10:02:50.776065 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ce8650bd8f98da7ce898f5bcd75ab5e21f63637a057f7de6cd1f89b8cb28fa4"} err="failed to get container status \"1ce8650bd8f98da7ce898f5bcd75ab5e21f63637a057f7de6cd1f89b8cb28fa4\": rpc error: code = NotFound desc = could not find container \"1ce8650bd8f98da7ce898f5bcd75ab5e21f63637a057f7de6cd1f89b8cb28fa4\": container with ID starting with 1ce8650bd8f98da7ce898f5bcd75ab5e21f63637a057f7de6cd1f89b8cb28fa4 not found: ID does not exist" Oct 06 10:02:50 crc kubenswrapper[4824]: I1006 10:02:50.776093 4824 scope.go:117] "RemoveContainer" containerID="c13a247642fda90779a68772ce9c3288d749e24fea28b15a22612cd488351a46" Oct 06 10:02:50 crc kubenswrapper[4824]: I1006 10:02:50.783777 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/06d3772b-e467-41a1-bf9b-9d6520ab68ef-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "06d3772b-e467-41a1-bf9b-9d6520ab68ef" (UID: "06d3772b-e467-41a1-bf9b-9d6520ab68ef"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:02:50 crc kubenswrapper[4824]: I1006 10:02:50.798777 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/030ddedc-a25b-47f6-88b6-c1286dc1c910-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "030ddedc-a25b-47f6-88b6-c1286dc1c910" (UID: "030ddedc-a25b-47f6-88b6-c1286dc1c910"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:02:50 crc kubenswrapper[4824]: I1006 10:02:50.798937 4824 scope.go:117] "RemoveContainer" containerID="dd59416586d18324918270d5729aa830d2868829f78adaf9e692b472eaf3f67d" Oct 06 10:02:50 crc kubenswrapper[4824]: I1006 10:02:50.803519 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/d2595750-aae8-4f17-8782-b7eccb0c2948-marketplace-operator-metrics\") pod \"d2595750-aae8-4f17-8782-b7eccb0c2948\" (UID: \"d2595750-aae8-4f17-8782-b7eccb0c2948\") " Oct 06 10:02:50 crc kubenswrapper[4824]: I1006 10:02:50.803634 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d2595750-aae8-4f17-8782-b7eccb0c2948-marketplace-trusted-ca\") pod \"d2595750-aae8-4f17-8782-b7eccb0c2948\" (UID: \"d2595750-aae8-4f17-8782-b7eccb0c2948\") " Oct 06 10:02:50 crc kubenswrapper[4824]: I1006 10:02:50.803761 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnnkq\" (UniqueName: \"kubernetes.io/projected/d2595750-aae8-4f17-8782-b7eccb0c2948-kube-api-access-rnnkq\") pod \"d2595750-aae8-4f17-8782-b7eccb0c2948\" (UID: \"d2595750-aae8-4f17-8782-b7eccb0c2948\") " Oct 06 10:02:50 crc kubenswrapper[4824]: I1006 10:02:50.804019 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/030ddedc-a25b-47f6-88b6-c1286dc1c910-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 06 10:02:50 crc kubenswrapper[4824]: I1006 10:02:50.804086 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-975sb\" (UniqueName: \"kubernetes.io/projected/030ddedc-a25b-47f6-88b6-c1286dc1c910-kube-api-access-975sb\") on node \"crc\" DevicePath \"\"" Oct 06 10:02:50 crc kubenswrapper[4824]: I1006 10:02:50.804155 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4h5wb\" (UniqueName: \"kubernetes.io/projected/c79214ab-ae6c-444b-bd4b-072c9a8bd37c-kube-api-access-4h5wb\") on node \"crc\" DevicePath \"\"" Oct 06 10:02:50 crc kubenswrapper[4824]: I1006 10:02:50.804217 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/06d3772b-e467-41a1-bf9b-9d6520ab68ef-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 06 10:02:50 crc kubenswrapper[4824]: I1006 10:02:50.804278 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/06d3772b-e467-41a1-bf9b-9d6520ab68ef-utilities\") on node \"crc\" DevicePath \"\"" Oct 06 10:02:50 crc kubenswrapper[4824]: I1006 10:02:50.804345 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fhs24\" (UniqueName: \"kubernetes.io/projected/06d3772b-e467-41a1-bf9b-9d6520ab68ef-kube-api-access-fhs24\") on node \"crc\" DevicePath \"\"" Oct 06 10:02:50 crc kubenswrapper[4824]: I1006 10:02:50.804401 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/030ddedc-a25b-47f6-88b6-c1286dc1c910-utilities\") on node \"crc\" DevicePath \"\"" Oct 06 10:02:50 crc kubenswrapper[4824]: I1006 10:02:50.804462 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c79214ab-ae6c-444b-bd4b-072c9a8bd37c-utilities\") on node \"crc\" DevicePath \"\"" Oct 06 10:02:50 crc kubenswrapper[4824]: I1006 10:02:50.805030 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d2595750-aae8-4f17-8782-b7eccb0c2948-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "d2595750-aae8-4f17-8782-b7eccb0c2948" (UID: "d2595750-aae8-4f17-8782-b7eccb0c2948"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:02:50 crc kubenswrapper[4824]: I1006 10:02:50.807347 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2595750-aae8-4f17-8782-b7eccb0c2948-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "d2595750-aae8-4f17-8782-b7eccb0c2948" (UID: "d2595750-aae8-4f17-8782-b7eccb0c2948"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:02:50 crc kubenswrapper[4824]: I1006 10:02:50.807822 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d2595750-aae8-4f17-8782-b7eccb0c2948-kube-api-access-rnnkq" (OuterVolumeSpecName: "kube-api-access-rnnkq") pod "d2595750-aae8-4f17-8782-b7eccb0c2948" (UID: "d2595750-aae8-4f17-8782-b7eccb0c2948"). InnerVolumeSpecName "kube-api-access-rnnkq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:02:50 crc kubenswrapper[4824]: I1006 10:02:50.812832 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c79214ab-ae6c-444b-bd4b-072c9a8bd37c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c79214ab-ae6c-444b-bd4b-072c9a8bd37c" (UID: "c79214ab-ae6c-444b-bd4b-072c9a8bd37c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:02:50 crc kubenswrapper[4824]: I1006 10:02:50.824658 4824 scope.go:117] "RemoveContainer" containerID="447c2f2c33a8f910d33cae05ca6bed23a4f4b23c4f68908b9f2546f06c5be738" Oct 06 10:02:50 crc kubenswrapper[4824]: I1006 10:02:50.840659 4824 scope.go:117] "RemoveContainer" containerID="c13a247642fda90779a68772ce9c3288d749e24fea28b15a22612cd488351a46" Oct 06 10:02:50 crc kubenswrapper[4824]: E1006 10:02:50.841157 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c13a247642fda90779a68772ce9c3288d749e24fea28b15a22612cd488351a46\": container with ID starting with c13a247642fda90779a68772ce9c3288d749e24fea28b15a22612cd488351a46 not found: ID does not exist" containerID="c13a247642fda90779a68772ce9c3288d749e24fea28b15a22612cd488351a46" Oct 06 10:02:50 crc kubenswrapper[4824]: I1006 10:02:50.841203 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c13a247642fda90779a68772ce9c3288d749e24fea28b15a22612cd488351a46"} err="failed to get container status \"c13a247642fda90779a68772ce9c3288d749e24fea28b15a22612cd488351a46\": rpc error: code = NotFound desc = could not find container \"c13a247642fda90779a68772ce9c3288d749e24fea28b15a22612cd488351a46\": container with ID starting with c13a247642fda90779a68772ce9c3288d749e24fea28b15a22612cd488351a46 not found: ID does not exist" Oct 06 10:02:50 crc kubenswrapper[4824]: I1006 10:02:50.841233 4824 scope.go:117] "RemoveContainer" containerID="dd59416586d18324918270d5729aa830d2868829f78adaf9e692b472eaf3f67d" Oct 06 10:02:50 crc kubenswrapper[4824]: E1006 10:02:50.841832 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dd59416586d18324918270d5729aa830d2868829f78adaf9e692b472eaf3f67d\": container with ID starting with dd59416586d18324918270d5729aa830d2868829f78adaf9e692b472eaf3f67d not found: ID does not exist" containerID="dd59416586d18324918270d5729aa830d2868829f78adaf9e692b472eaf3f67d" Oct 06 10:02:50 crc kubenswrapper[4824]: I1006 10:02:50.841857 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dd59416586d18324918270d5729aa830d2868829f78adaf9e692b472eaf3f67d"} err="failed to get container status \"dd59416586d18324918270d5729aa830d2868829f78adaf9e692b472eaf3f67d\": rpc error: code = NotFound desc = could not find container \"dd59416586d18324918270d5729aa830d2868829f78adaf9e692b472eaf3f67d\": container with ID starting with dd59416586d18324918270d5729aa830d2868829f78adaf9e692b472eaf3f67d not found: ID does not exist" Oct 06 10:02:50 crc kubenswrapper[4824]: I1006 10:02:50.841871 4824 scope.go:117] "RemoveContainer" containerID="447c2f2c33a8f910d33cae05ca6bed23a4f4b23c4f68908b9f2546f06c5be738" Oct 06 10:02:50 crc kubenswrapper[4824]: E1006 10:02:50.842376 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"447c2f2c33a8f910d33cae05ca6bed23a4f4b23c4f68908b9f2546f06c5be738\": container with ID starting with 447c2f2c33a8f910d33cae05ca6bed23a4f4b23c4f68908b9f2546f06c5be738 not found: ID does not exist" containerID="447c2f2c33a8f910d33cae05ca6bed23a4f4b23c4f68908b9f2546f06c5be738" Oct 06 10:02:50 crc kubenswrapper[4824]: I1006 10:02:50.842437 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"447c2f2c33a8f910d33cae05ca6bed23a4f4b23c4f68908b9f2546f06c5be738"} err="failed to get container status \"447c2f2c33a8f910d33cae05ca6bed23a4f4b23c4f68908b9f2546f06c5be738\": rpc error: code = NotFound desc = could not find container \"447c2f2c33a8f910d33cae05ca6bed23a4f4b23c4f68908b9f2546f06c5be738\": container with ID starting with 447c2f2c33a8f910d33cae05ca6bed23a4f4b23c4f68908b9f2546f06c5be738 not found: ID does not exist" Oct 06 10:02:50 crc kubenswrapper[4824]: I1006 10:02:50.842483 4824 scope.go:117] "RemoveContainer" containerID="853c3143b0b0b58e6bebb8466a6f31993342218020b663e0e381e288cc558940" Oct 06 10:02:50 crc kubenswrapper[4824]: I1006 10:02:50.865642 4824 scope.go:117] "RemoveContainer" containerID="4f1cda4ec404abec1f0abfc1c077db1856f8bcf4a45c9403fb951fb24e316277" Oct 06 10:02:50 crc kubenswrapper[4824]: I1006 10:02:50.880776 4824 scope.go:117] "RemoveContainer" containerID="f05dfc3bc64501e41918c87609a67f03234892efdbd157bbcd331cb2f1e33136" Oct 06 10:02:50 crc kubenswrapper[4824]: I1006 10:02:50.905818 4824 scope.go:117] "RemoveContainer" containerID="853c3143b0b0b58e6bebb8466a6f31993342218020b663e0e381e288cc558940" Oct 06 10:02:50 crc kubenswrapper[4824]: I1006 10:02:50.906059 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnnkq\" (UniqueName: \"kubernetes.io/projected/d2595750-aae8-4f17-8782-b7eccb0c2948-kube-api-access-rnnkq\") on node \"crc\" DevicePath \"\"" Oct 06 10:02:50 crc kubenswrapper[4824]: I1006 10:02:50.906086 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c79214ab-ae6c-444b-bd4b-072c9a8bd37c-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 06 10:02:50 crc kubenswrapper[4824]: I1006 10:02:50.906100 4824 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/d2595750-aae8-4f17-8782-b7eccb0c2948-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Oct 06 10:02:50 crc kubenswrapper[4824]: I1006 10:02:50.906115 4824 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d2595750-aae8-4f17-8782-b7eccb0c2948-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 06 10:02:50 crc kubenswrapper[4824]: E1006 10:02:50.912812 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"853c3143b0b0b58e6bebb8466a6f31993342218020b663e0e381e288cc558940\": container with ID starting with 853c3143b0b0b58e6bebb8466a6f31993342218020b663e0e381e288cc558940 not found: ID does not exist" containerID="853c3143b0b0b58e6bebb8466a6f31993342218020b663e0e381e288cc558940" Oct 06 10:02:50 crc kubenswrapper[4824]: I1006 10:02:50.912861 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"853c3143b0b0b58e6bebb8466a6f31993342218020b663e0e381e288cc558940"} err="failed to get container status \"853c3143b0b0b58e6bebb8466a6f31993342218020b663e0e381e288cc558940\": rpc error: code = NotFound desc = could not find container \"853c3143b0b0b58e6bebb8466a6f31993342218020b663e0e381e288cc558940\": container with ID starting with 853c3143b0b0b58e6bebb8466a6f31993342218020b663e0e381e288cc558940 not found: ID does not exist" Oct 06 10:02:50 crc kubenswrapper[4824]: I1006 10:02:50.912896 4824 scope.go:117] "RemoveContainer" containerID="4f1cda4ec404abec1f0abfc1c077db1856f8bcf4a45c9403fb951fb24e316277" Oct 06 10:02:50 crc kubenswrapper[4824]: E1006 10:02:50.913410 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4f1cda4ec404abec1f0abfc1c077db1856f8bcf4a45c9403fb951fb24e316277\": container with ID starting with 4f1cda4ec404abec1f0abfc1c077db1856f8bcf4a45c9403fb951fb24e316277 not found: ID does not exist" containerID="4f1cda4ec404abec1f0abfc1c077db1856f8bcf4a45c9403fb951fb24e316277" Oct 06 10:02:50 crc kubenswrapper[4824]: I1006 10:02:50.913438 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4f1cda4ec404abec1f0abfc1c077db1856f8bcf4a45c9403fb951fb24e316277"} err="failed to get container status \"4f1cda4ec404abec1f0abfc1c077db1856f8bcf4a45c9403fb951fb24e316277\": rpc error: code = NotFound desc = could not find container \"4f1cda4ec404abec1f0abfc1c077db1856f8bcf4a45c9403fb951fb24e316277\": container with ID starting with 4f1cda4ec404abec1f0abfc1c077db1856f8bcf4a45c9403fb951fb24e316277 not found: ID does not exist" Oct 06 10:02:50 crc kubenswrapper[4824]: I1006 10:02:50.913455 4824 scope.go:117] "RemoveContainer" containerID="f05dfc3bc64501e41918c87609a67f03234892efdbd157bbcd331cb2f1e33136" Oct 06 10:02:50 crc kubenswrapper[4824]: E1006 10:02:50.913668 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f05dfc3bc64501e41918c87609a67f03234892efdbd157bbcd331cb2f1e33136\": container with ID starting with f05dfc3bc64501e41918c87609a67f03234892efdbd157bbcd331cb2f1e33136 not found: ID does not exist" containerID="f05dfc3bc64501e41918c87609a67f03234892efdbd157bbcd331cb2f1e33136" Oct 06 10:02:50 crc kubenswrapper[4824]: I1006 10:02:50.913684 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f05dfc3bc64501e41918c87609a67f03234892efdbd157bbcd331cb2f1e33136"} err="failed to get container status \"f05dfc3bc64501e41918c87609a67f03234892efdbd157bbcd331cb2f1e33136\": rpc error: code = NotFound desc = could not find container \"f05dfc3bc64501e41918c87609a67f03234892efdbd157bbcd331cb2f1e33136\": container with ID starting with f05dfc3bc64501e41918c87609a67f03234892efdbd157bbcd331cb2f1e33136 not found: ID does not exist" Oct 06 10:02:50 crc kubenswrapper[4824]: I1006 10:02:50.943914 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-db6vn"] Oct 06 10:02:50 crc kubenswrapper[4824]: W1006 10:02:50.957687 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda3a514b8_f221_465e_bbb2_3c88bcb4f622.slice/crio-8bd4d2d603ddfefcb8d182cf2e1ae41833e736b86d88d155485806ffccfaca6d WatchSource:0}: Error finding container 8bd4d2d603ddfefcb8d182cf2e1ae41833e736b86d88d155485806ffccfaca6d: Status 404 returned error can't find the container with id 8bd4d2d603ddfefcb8d182cf2e1ae41833e736b86d88d155485806ffccfaca6d Oct 06 10:02:51 crc kubenswrapper[4824]: I1006 10:02:51.055893 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bgbpb" Oct 06 10:02:51 crc kubenswrapper[4824]: I1006 10:02:51.064259 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-5jqs9"] Oct 06 10:02:51 crc kubenswrapper[4824]: I1006 10:02:51.067852 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-5jqs9"] Oct 06 10:02:51 crc kubenswrapper[4824]: I1006 10:02:51.078077 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-r52zf"] Oct 06 10:02:51 crc kubenswrapper[4824]: I1006 10:02:51.085000 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-r52zf"] Oct 06 10:02:51 crc kubenswrapper[4824]: I1006 10:02:51.128259 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-tlvv7"] Oct 06 10:02:51 crc kubenswrapper[4824]: I1006 10:02:51.134947 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-tlvv7"] Oct 06 10:02:51 crc kubenswrapper[4824]: I1006 10:02:51.208797 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/70324ded-62d7-47fc-a67a-110ecde1497c-utilities\") pod \"70324ded-62d7-47fc-a67a-110ecde1497c\" (UID: \"70324ded-62d7-47fc-a67a-110ecde1497c\") " Oct 06 10:02:51 crc kubenswrapper[4824]: I1006 10:02:51.208889 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v2pk8\" (UniqueName: \"kubernetes.io/projected/70324ded-62d7-47fc-a67a-110ecde1497c-kube-api-access-v2pk8\") pod \"70324ded-62d7-47fc-a67a-110ecde1497c\" (UID: \"70324ded-62d7-47fc-a67a-110ecde1497c\") " Oct 06 10:02:51 crc kubenswrapper[4824]: I1006 10:02:51.208957 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/70324ded-62d7-47fc-a67a-110ecde1497c-catalog-content\") pod \"70324ded-62d7-47fc-a67a-110ecde1497c\" (UID: \"70324ded-62d7-47fc-a67a-110ecde1497c\") " Oct 06 10:02:51 crc kubenswrapper[4824]: I1006 10:02:51.210501 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/70324ded-62d7-47fc-a67a-110ecde1497c-utilities" (OuterVolumeSpecName: "utilities") pod "70324ded-62d7-47fc-a67a-110ecde1497c" (UID: "70324ded-62d7-47fc-a67a-110ecde1497c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:02:51 crc kubenswrapper[4824]: I1006 10:02:51.215333 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/70324ded-62d7-47fc-a67a-110ecde1497c-kube-api-access-v2pk8" (OuterVolumeSpecName: "kube-api-access-v2pk8") pod "70324ded-62d7-47fc-a67a-110ecde1497c" (UID: "70324ded-62d7-47fc-a67a-110ecde1497c"). InnerVolumeSpecName "kube-api-access-v2pk8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:02:51 crc kubenswrapper[4824]: I1006 10:02:51.221644 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/70324ded-62d7-47fc-a67a-110ecde1497c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "70324ded-62d7-47fc-a67a-110ecde1497c" (UID: "70324ded-62d7-47fc-a67a-110ecde1497c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:02:51 crc kubenswrapper[4824]: I1006 10:02:51.282738 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="030ddedc-a25b-47f6-88b6-c1286dc1c910" path="/var/lib/kubelet/pods/030ddedc-a25b-47f6-88b6-c1286dc1c910/volumes" Oct 06 10:02:51 crc kubenswrapper[4824]: I1006 10:02:51.283500 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="06d3772b-e467-41a1-bf9b-9d6520ab68ef" path="/var/lib/kubelet/pods/06d3772b-e467-41a1-bf9b-9d6520ab68ef/volumes" Oct 06 10:02:51 crc kubenswrapper[4824]: I1006 10:02:51.284248 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c79214ab-ae6c-444b-bd4b-072c9a8bd37c" path="/var/lib/kubelet/pods/c79214ab-ae6c-444b-bd4b-072c9a8bd37c/volumes" Oct 06 10:02:51 crc kubenswrapper[4824]: I1006 10:02:51.310236 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/70324ded-62d7-47fc-a67a-110ecde1497c-utilities\") on node \"crc\" DevicePath \"\"" Oct 06 10:02:51 crc kubenswrapper[4824]: I1006 10:02:51.310271 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v2pk8\" (UniqueName: \"kubernetes.io/projected/70324ded-62d7-47fc-a67a-110ecde1497c-kube-api-access-v2pk8\") on node \"crc\" DevicePath \"\"" Oct 06 10:02:51 crc kubenswrapper[4824]: I1006 10:02:51.310283 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/70324ded-62d7-47fc-a67a-110ecde1497c-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 06 10:02:51 crc kubenswrapper[4824]: I1006 10:02:51.709145 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-db6vn" event={"ID":"a3a514b8-f221-465e-bbb2-3c88bcb4f622","Type":"ContainerStarted","Data":"5acb00bd82e1644c375bc7fcada8a8a928ca5e60f149465888bfdb96769c249d"} Oct 06 10:02:51 crc kubenswrapper[4824]: I1006 10:02:51.709844 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-db6vn" event={"ID":"a3a514b8-f221-465e-bbb2-3c88bcb4f622","Type":"ContainerStarted","Data":"8bd4d2d603ddfefcb8d182cf2e1ae41833e736b86d88d155485806ffccfaca6d"} Oct 06 10:02:51 crc kubenswrapper[4824]: I1006 10:02:51.709935 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-db6vn" Oct 06 10:02:51 crc kubenswrapper[4824]: I1006 10:02:51.711462 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-db6vn" Oct 06 10:02:51 crc kubenswrapper[4824]: I1006 10:02:51.717426 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-qq4zz" event={"ID":"d2595750-aae8-4f17-8782-b7eccb0c2948","Type":"ContainerDied","Data":"fde24d5c6f8274bd8d527ae7747c3941c580a57dc5640c813a74782ac1f9d4b5"} Oct 06 10:02:51 crc kubenswrapper[4824]: I1006 10:02:51.717503 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-qq4zz" Oct 06 10:02:51 crc kubenswrapper[4824]: I1006 10:02:51.717621 4824 scope.go:117] "RemoveContainer" containerID="d27b1d97570bdd6569bbf1dc7b40419c3ce223f7aac865ed7a1b5c65b7acfcaa" Oct 06 10:02:51 crc kubenswrapper[4824]: I1006 10:02:51.721938 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bgbpb" event={"ID":"70324ded-62d7-47fc-a67a-110ecde1497c","Type":"ContainerDied","Data":"51ee4f7e0be054dee3cb8b7bcc662959901a5f2cf8f67435de0ef0d484a89694"} Oct 06 10:02:51 crc kubenswrapper[4824]: I1006 10:02:51.722069 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bgbpb" Oct 06 10:02:51 crc kubenswrapper[4824]: I1006 10:02:51.734486 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-db6vn" podStartSLOduration=1.734468404 podStartE2EDuration="1.734468404s" podCreationTimestamp="2025-10-06 10:02:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:02:51.7324233 +0000 UTC m=+281.096846161" watchObservedRunningTime="2025-10-06 10:02:51.734468404 +0000 UTC m=+281.098891265" Oct 06 10:02:51 crc kubenswrapper[4824]: I1006 10:02:51.740345 4824 scope.go:117] "RemoveContainer" containerID="e4db6b7d450dc54a789cbfe6f13124becd9116d29f9d07848dbeb547dab40be5" Oct 06 10:02:51 crc kubenswrapper[4824]: I1006 10:02:51.746726 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-qq4zz"] Oct 06 10:02:51 crc kubenswrapper[4824]: I1006 10:02:51.752039 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-qq4zz"] Oct 06 10:02:51 crc kubenswrapper[4824]: I1006 10:02:51.764541 4824 scope.go:117] "RemoveContainer" containerID="f5d533e67f48f603c4640f3ea8a4d5b9e0e4fa5f3750daf25c5acd92a566d758" Oct 06 10:02:51 crc kubenswrapper[4824]: I1006 10:02:51.793348 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-bgbpb"] Oct 06 10:02:51 crc kubenswrapper[4824]: I1006 10:02:51.794037 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-bgbpb"] Oct 06 10:02:51 crc kubenswrapper[4824]: I1006 10:02:51.797629 4824 scope.go:117] "RemoveContainer" containerID="a535c6214daf2963716066e8472758cee9b58fdecb11e0420742fd13986ddc34" Oct 06 10:02:52 crc kubenswrapper[4824]: I1006 10:02:52.352839 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-xdszq"] Oct 06 10:02:52 crc kubenswrapper[4824]: E1006 10:02:52.353113 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c79214ab-ae6c-444b-bd4b-072c9a8bd37c" containerName="extract-content" Oct 06 10:02:52 crc kubenswrapper[4824]: I1006 10:02:52.353130 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="c79214ab-ae6c-444b-bd4b-072c9a8bd37c" containerName="extract-content" Oct 06 10:02:52 crc kubenswrapper[4824]: E1006 10:02:52.353144 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="030ddedc-a25b-47f6-88b6-c1286dc1c910" containerName="extract-content" Oct 06 10:02:52 crc kubenswrapper[4824]: I1006 10:02:52.353152 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="030ddedc-a25b-47f6-88b6-c1286dc1c910" containerName="extract-content" Oct 06 10:02:52 crc kubenswrapper[4824]: E1006 10:02:52.353164 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2595750-aae8-4f17-8782-b7eccb0c2948" containerName="marketplace-operator" Oct 06 10:02:52 crc kubenswrapper[4824]: I1006 10:02:52.353174 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2595750-aae8-4f17-8782-b7eccb0c2948" containerName="marketplace-operator" Oct 06 10:02:52 crc kubenswrapper[4824]: E1006 10:02:52.353186 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06d3772b-e467-41a1-bf9b-9d6520ab68ef" containerName="extract-utilities" Oct 06 10:02:52 crc kubenswrapper[4824]: I1006 10:02:52.353193 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="06d3772b-e467-41a1-bf9b-9d6520ab68ef" containerName="extract-utilities" Oct 06 10:02:52 crc kubenswrapper[4824]: E1006 10:02:52.353212 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="030ddedc-a25b-47f6-88b6-c1286dc1c910" containerName="registry-server" Oct 06 10:02:52 crc kubenswrapper[4824]: I1006 10:02:52.353247 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="030ddedc-a25b-47f6-88b6-c1286dc1c910" containerName="registry-server" Oct 06 10:02:52 crc kubenswrapper[4824]: E1006 10:02:52.353259 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70324ded-62d7-47fc-a67a-110ecde1497c" containerName="registry-server" Oct 06 10:02:52 crc kubenswrapper[4824]: I1006 10:02:52.353266 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="70324ded-62d7-47fc-a67a-110ecde1497c" containerName="registry-server" Oct 06 10:02:52 crc kubenswrapper[4824]: E1006 10:02:52.353279 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06d3772b-e467-41a1-bf9b-9d6520ab68ef" containerName="extract-content" Oct 06 10:02:52 crc kubenswrapper[4824]: I1006 10:02:52.353291 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="06d3772b-e467-41a1-bf9b-9d6520ab68ef" containerName="extract-content" Oct 06 10:02:52 crc kubenswrapper[4824]: E1006 10:02:52.353303 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70324ded-62d7-47fc-a67a-110ecde1497c" containerName="extract-utilities" Oct 06 10:02:52 crc kubenswrapper[4824]: I1006 10:02:52.353310 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="70324ded-62d7-47fc-a67a-110ecde1497c" containerName="extract-utilities" Oct 06 10:02:52 crc kubenswrapper[4824]: E1006 10:02:52.353321 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06d3772b-e467-41a1-bf9b-9d6520ab68ef" containerName="registry-server" Oct 06 10:02:52 crc kubenswrapper[4824]: I1006 10:02:52.353330 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="06d3772b-e467-41a1-bf9b-9d6520ab68ef" containerName="registry-server" Oct 06 10:02:52 crc kubenswrapper[4824]: E1006 10:02:52.353340 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="030ddedc-a25b-47f6-88b6-c1286dc1c910" containerName="extract-utilities" Oct 06 10:02:52 crc kubenswrapper[4824]: I1006 10:02:52.353350 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="030ddedc-a25b-47f6-88b6-c1286dc1c910" containerName="extract-utilities" Oct 06 10:02:52 crc kubenswrapper[4824]: E1006 10:02:52.353359 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c79214ab-ae6c-444b-bd4b-072c9a8bd37c" containerName="registry-server" Oct 06 10:02:52 crc kubenswrapper[4824]: I1006 10:02:52.353367 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="c79214ab-ae6c-444b-bd4b-072c9a8bd37c" containerName="registry-server" Oct 06 10:02:52 crc kubenswrapper[4824]: E1006 10:02:52.353374 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70324ded-62d7-47fc-a67a-110ecde1497c" containerName="extract-content" Oct 06 10:02:52 crc kubenswrapper[4824]: I1006 10:02:52.353382 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="70324ded-62d7-47fc-a67a-110ecde1497c" containerName="extract-content" Oct 06 10:02:52 crc kubenswrapper[4824]: E1006 10:02:52.353390 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c79214ab-ae6c-444b-bd4b-072c9a8bd37c" containerName="extract-utilities" Oct 06 10:02:52 crc kubenswrapper[4824]: I1006 10:02:52.353400 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="c79214ab-ae6c-444b-bd4b-072c9a8bd37c" containerName="extract-utilities" Oct 06 10:02:52 crc kubenswrapper[4824]: I1006 10:02:52.353500 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="c79214ab-ae6c-444b-bd4b-072c9a8bd37c" containerName="registry-server" Oct 06 10:02:52 crc kubenswrapper[4824]: I1006 10:02:52.353513 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="030ddedc-a25b-47f6-88b6-c1286dc1c910" containerName="registry-server" Oct 06 10:02:52 crc kubenswrapper[4824]: I1006 10:02:52.353529 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="d2595750-aae8-4f17-8782-b7eccb0c2948" containerName="marketplace-operator" Oct 06 10:02:52 crc kubenswrapper[4824]: I1006 10:02:52.353542 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="70324ded-62d7-47fc-a67a-110ecde1497c" containerName="registry-server" Oct 06 10:02:52 crc kubenswrapper[4824]: I1006 10:02:52.353552 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="06d3772b-e467-41a1-bf9b-9d6520ab68ef" containerName="registry-server" Oct 06 10:02:52 crc kubenswrapper[4824]: I1006 10:02:52.354277 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xdszq" Oct 06 10:02:52 crc kubenswrapper[4824]: I1006 10:02:52.356655 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Oct 06 10:02:52 crc kubenswrapper[4824]: I1006 10:02:52.364942 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xdszq"] Oct 06 10:02:52 crc kubenswrapper[4824]: I1006 10:02:52.423066 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/504bf3d4-9589-4a5e-ad25-e16f783e0115-utilities\") pod \"redhat-marketplace-xdszq\" (UID: \"504bf3d4-9589-4a5e-ad25-e16f783e0115\") " pod="openshift-marketplace/redhat-marketplace-xdszq" Oct 06 10:02:52 crc kubenswrapper[4824]: I1006 10:02:52.423156 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j88v5\" (UniqueName: \"kubernetes.io/projected/504bf3d4-9589-4a5e-ad25-e16f783e0115-kube-api-access-j88v5\") pod \"redhat-marketplace-xdszq\" (UID: \"504bf3d4-9589-4a5e-ad25-e16f783e0115\") " pod="openshift-marketplace/redhat-marketplace-xdszq" Oct 06 10:02:52 crc kubenswrapper[4824]: I1006 10:02:52.423190 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/504bf3d4-9589-4a5e-ad25-e16f783e0115-catalog-content\") pod \"redhat-marketplace-xdszq\" (UID: \"504bf3d4-9589-4a5e-ad25-e16f783e0115\") " pod="openshift-marketplace/redhat-marketplace-xdszq" Oct 06 10:02:52 crc kubenswrapper[4824]: I1006 10:02:52.525188 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/504bf3d4-9589-4a5e-ad25-e16f783e0115-catalog-content\") pod \"redhat-marketplace-xdszq\" (UID: \"504bf3d4-9589-4a5e-ad25-e16f783e0115\") " pod="openshift-marketplace/redhat-marketplace-xdszq" Oct 06 10:02:52 crc kubenswrapper[4824]: I1006 10:02:52.525280 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/504bf3d4-9589-4a5e-ad25-e16f783e0115-utilities\") pod \"redhat-marketplace-xdszq\" (UID: \"504bf3d4-9589-4a5e-ad25-e16f783e0115\") " pod="openshift-marketplace/redhat-marketplace-xdszq" Oct 06 10:02:52 crc kubenswrapper[4824]: I1006 10:02:52.525400 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j88v5\" (UniqueName: \"kubernetes.io/projected/504bf3d4-9589-4a5e-ad25-e16f783e0115-kube-api-access-j88v5\") pod \"redhat-marketplace-xdszq\" (UID: \"504bf3d4-9589-4a5e-ad25-e16f783e0115\") " pod="openshift-marketplace/redhat-marketplace-xdszq" Oct 06 10:02:52 crc kubenswrapper[4824]: I1006 10:02:52.526353 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/504bf3d4-9589-4a5e-ad25-e16f783e0115-catalog-content\") pod \"redhat-marketplace-xdszq\" (UID: \"504bf3d4-9589-4a5e-ad25-e16f783e0115\") " pod="openshift-marketplace/redhat-marketplace-xdszq" Oct 06 10:02:52 crc kubenswrapper[4824]: I1006 10:02:52.526547 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/504bf3d4-9589-4a5e-ad25-e16f783e0115-utilities\") pod \"redhat-marketplace-xdszq\" (UID: \"504bf3d4-9589-4a5e-ad25-e16f783e0115\") " pod="openshift-marketplace/redhat-marketplace-xdszq" Oct 06 10:02:52 crc kubenswrapper[4824]: I1006 10:02:52.530895 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-txrmb"] Oct 06 10:02:52 crc kubenswrapper[4824]: I1006 10:02:52.531854 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-txrmb" Oct 06 10:02:52 crc kubenswrapper[4824]: I1006 10:02:52.534577 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Oct 06 10:02:52 crc kubenswrapper[4824]: I1006 10:02:52.549703 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-txrmb"] Oct 06 10:02:52 crc kubenswrapper[4824]: I1006 10:02:52.572792 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j88v5\" (UniqueName: \"kubernetes.io/projected/504bf3d4-9589-4a5e-ad25-e16f783e0115-kube-api-access-j88v5\") pod \"redhat-marketplace-xdszq\" (UID: \"504bf3d4-9589-4a5e-ad25-e16f783e0115\") " pod="openshift-marketplace/redhat-marketplace-xdszq" Oct 06 10:02:52 crc kubenswrapper[4824]: I1006 10:02:52.626447 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5fe63731-833a-45c3-b4f0-73ddc05383e1-catalog-content\") pod \"redhat-operators-txrmb\" (UID: \"5fe63731-833a-45c3-b4f0-73ddc05383e1\") " pod="openshift-marketplace/redhat-operators-txrmb" Oct 06 10:02:52 crc kubenswrapper[4824]: I1006 10:02:52.626612 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5fe63731-833a-45c3-b4f0-73ddc05383e1-utilities\") pod \"redhat-operators-txrmb\" (UID: \"5fe63731-833a-45c3-b4f0-73ddc05383e1\") " pod="openshift-marketplace/redhat-operators-txrmb" Oct 06 10:02:52 crc kubenswrapper[4824]: I1006 10:02:52.626650 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w2lkd\" (UniqueName: \"kubernetes.io/projected/5fe63731-833a-45c3-b4f0-73ddc05383e1-kube-api-access-w2lkd\") pod \"redhat-operators-txrmb\" (UID: \"5fe63731-833a-45c3-b4f0-73ddc05383e1\") " pod="openshift-marketplace/redhat-operators-txrmb" Oct 06 10:02:52 crc kubenswrapper[4824]: I1006 10:02:52.713365 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xdszq" Oct 06 10:02:52 crc kubenswrapper[4824]: I1006 10:02:52.728784 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5fe63731-833a-45c3-b4f0-73ddc05383e1-utilities\") pod \"redhat-operators-txrmb\" (UID: \"5fe63731-833a-45c3-b4f0-73ddc05383e1\") " pod="openshift-marketplace/redhat-operators-txrmb" Oct 06 10:02:52 crc kubenswrapper[4824]: I1006 10:02:52.728828 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w2lkd\" (UniqueName: \"kubernetes.io/projected/5fe63731-833a-45c3-b4f0-73ddc05383e1-kube-api-access-w2lkd\") pod \"redhat-operators-txrmb\" (UID: \"5fe63731-833a-45c3-b4f0-73ddc05383e1\") " pod="openshift-marketplace/redhat-operators-txrmb" Oct 06 10:02:52 crc kubenswrapper[4824]: I1006 10:02:52.728859 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5fe63731-833a-45c3-b4f0-73ddc05383e1-catalog-content\") pod \"redhat-operators-txrmb\" (UID: \"5fe63731-833a-45c3-b4f0-73ddc05383e1\") " pod="openshift-marketplace/redhat-operators-txrmb" Oct 06 10:02:52 crc kubenswrapper[4824]: I1006 10:02:52.729359 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5fe63731-833a-45c3-b4f0-73ddc05383e1-catalog-content\") pod \"redhat-operators-txrmb\" (UID: \"5fe63731-833a-45c3-b4f0-73ddc05383e1\") " pod="openshift-marketplace/redhat-operators-txrmb" Oct 06 10:02:52 crc kubenswrapper[4824]: I1006 10:02:52.729438 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5fe63731-833a-45c3-b4f0-73ddc05383e1-utilities\") pod \"redhat-operators-txrmb\" (UID: \"5fe63731-833a-45c3-b4f0-73ddc05383e1\") " pod="openshift-marketplace/redhat-operators-txrmb" Oct 06 10:02:52 crc kubenswrapper[4824]: I1006 10:02:52.771938 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w2lkd\" (UniqueName: \"kubernetes.io/projected/5fe63731-833a-45c3-b4f0-73ddc05383e1-kube-api-access-w2lkd\") pod \"redhat-operators-txrmb\" (UID: \"5fe63731-833a-45c3-b4f0-73ddc05383e1\") " pod="openshift-marketplace/redhat-operators-txrmb" Oct 06 10:02:52 crc kubenswrapper[4824]: I1006 10:02:52.858838 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-txrmb" Oct 06 10:02:52 crc kubenswrapper[4824]: I1006 10:02:52.958778 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xdszq"] Oct 06 10:02:53 crc kubenswrapper[4824]: I1006 10:02:53.089545 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-txrmb"] Oct 06 10:02:53 crc kubenswrapper[4824]: W1006 10:02:53.097737 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5fe63731_833a_45c3_b4f0_73ddc05383e1.slice/crio-93099c5b8c0ba07ad990ae7d135501f077a0493d7559f9bdc215e4e5e4172787 WatchSource:0}: Error finding container 93099c5b8c0ba07ad990ae7d135501f077a0493d7559f9bdc215e4e5e4172787: Status 404 returned error can't find the container with id 93099c5b8c0ba07ad990ae7d135501f077a0493d7559f9bdc215e4e5e4172787 Oct 06 10:02:53 crc kubenswrapper[4824]: I1006 10:02:53.283720 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="70324ded-62d7-47fc-a67a-110ecde1497c" path="/var/lib/kubelet/pods/70324ded-62d7-47fc-a67a-110ecde1497c/volumes" Oct 06 10:02:53 crc kubenswrapper[4824]: I1006 10:02:53.284584 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d2595750-aae8-4f17-8782-b7eccb0c2948" path="/var/lib/kubelet/pods/d2595750-aae8-4f17-8782-b7eccb0c2948/volumes" Oct 06 10:02:53 crc kubenswrapper[4824]: I1006 10:02:53.741737 4824 generic.go:334] "Generic (PLEG): container finished" podID="5fe63731-833a-45c3-b4f0-73ddc05383e1" containerID="a5fad09e3dcaedd867384b2a523e9663c29845ea78983ea35cfcd6dce8583e2e" exitCode=0 Oct 06 10:02:53 crc kubenswrapper[4824]: I1006 10:02:53.741854 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-txrmb" event={"ID":"5fe63731-833a-45c3-b4f0-73ddc05383e1","Type":"ContainerDied","Data":"a5fad09e3dcaedd867384b2a523e9663c29845ea78983ea35cfcd6dce8583e2e"} Oct 06 10:02:53 crc kubenswrapper[4824]: I1006 10:02:53.742384 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-txrmb" event={"ID":"5fe63731-833a-45c3-b4f0-73ddc05383e1","Type":"ContainerStarted","Data":"93099c5b8c0ba07ad990ae7d135501f077a0493d7559f9bdc215e4e5e4172787"} Oct 06 10:02:53 crc kubenswrapper[4824]: I1006 10:02:53.744968 4824 generic.go:334] "Generic (PLEG): container finished" podID="504bf3d4-9589-4a5e-ad25-e16f783e0115" containerID="430535c810fec91d2326afcdda68e9a1a6e963256d3bb94c55eaef18444d471a" exitCode=0 Oct 06 10:02:53 crc kubenswrapper[4824]: I1006 10:02:53.746236 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xdszq" event={"ID":"504bf3d4-9589-4a5e-ad25-e16f783e0115","Type":"ContainerDied","Data":"430535c810fec91d2326afcdda68e9a1a6e963256d3bb94c55eaef18444d471a"} Oct 06 10:02:53 crc kubenswrapper[4824]: I1006 10:02:53.746289 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xdszq" event={"ID":"504bf3d4-9589-4a5e-ad25-e16f783e0115","Type":"ContainerStarted","Data":"8be436911a881bf31bf66b371203d8769bd8b0d0e79ae7888a40c09b9a783d79"} Oct 06 10:02:54 crc kubenswrapper[4824]: I1006 10:02:54.748689 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-mblh7"] Oct 06 10:02:54 crc kubenswrapper[4824]: I1006 10:02:54.752431 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mblh7" Oct 06 10:02:54 crc kubenswrapper[4824]: I1006 10:02:54.755303 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Oct 06 10:02:54 crc kubenswrapper[4824]: I1006 10:02:54.767728 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mblh7"] Oct 06 10:02:54 crc kubenswrapper[4824]: I1006 10:02:54.862895 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2dc6af8a-35a7-4fa8-9e03-eb6df841f127-catalog-content\") pod \"certified-operators-mblh7\" (UID: \"2dc6af8a-35a7-4fa8-9e03-eb6df841f127\") " pod="openshift-marketplace/certified-operators-mblh7" Oct 06 10:02:54 crc kubenswrapper[4824]: I1006 10:02:54.863392 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2dc6af8a-35a7-4fa8-9e03-eb6df841f127-utilities\") pod \"certified-operators-mblh7\" (UID: \"2dc6af8a-35a7-4fa8-9e03-eb6df841f127\") " pod="openshift-marketplace/certified-operators-mblh7" Oct 06 10:02:54 crc kubenswrapper[4824]: I1006 10:02:54.863449 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ll4bb\" (UniqueName: \"kubernetes.io/projected/2dc6af8a-35a7-4fa8-9e03-eb6df841f127-kube-api-access-ll4bb\") pod \"certified-operators-mblh7\" (UID: \"2dc6af8a-35a7-4fa8-9e03-eb6df841f127\") " pod="openshift-marketplace/certified-operators-mblh7" Oct 06 10:02:54 crc kubenswrapper[4824]: I1006 10:02:54.934349 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-s2nbh"] Oct 06 10:02:54 crc kubenswrapper[4824]: I1006 10:02:54.936290 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-s2nbh" Oct 06 10:02:54 crc kubenswrapper[4824]: I1006 10:02:54.938572 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Oct 06 10:02:54 crc kubenswrapper[4824]: I1006 10:02:54.952062 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-s2nbh"] Oct 06 10:02:54 crc kubenswrapper[4824]: I1006 10:02:54.965517 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2dc6af8a-35a7-4fa8-9e03-eb6df841f127-catalog-content\") pod \"certified-operators-mblh7\" (UID: \"2dc6af8a-35a7-4fa8-9e03-eb6df841f127\") " pod="openshift-marketplace/certified-operators-mblh7" Oct 06 10:02:54 crc kubenswrapper[4824]: I1006 10:02:54.965596 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2dc6af8a-35a7-4fa8-9e03-eb6df841f127-utilities\") pod \"certified-operators-mblh7\" (UID: \"2dc6af8a-35a7-4fa8-9e03-eb6df841f127\") " pod="openshift-marketplace/certified-operators-mblh7" Oct 06 10:02:54 crc kubenswrapper[4824]: I1006 10:02:54.965642 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ll4bb\" (UniqueName: \"kubernetes.io/projected/2dc6af8a-35a7-4fa8-9e03-eb6df841f127-kube-api-access-ll4bb\") pod \"certified-operators-mblh7\" (UID: \"2dc6af8a-35a7-4fa8-9e03-eb6df841f127\") " pod="openshift-marketplace/certified-operators-mblh7" Oct 06 10:02:54 crc kubenswrapper[4824]: I1006 10:02:54.966638 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2dc6af8a-35a7-4fa8-9e03-eb6df841f127-catalog-content\") pod \"certified-operators-mblh7\" (UID: \"2dc6af8a-35a7-4fa8-9e03-eb6df841f127\") " pod="openshift-marketplace/certified-operators-mblh7" Oct 06 10:02:54 crc kubenswrapper[4824]: I1006 10:02:54.966727 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2dc6af8a-35a7-4fa8-9e03-eb6df841f127-utilities\") pod \"certified-operators-mblh7\" (UID: \"2dc6af8a-35a7-4fa8-9e03-eb6df841f127\") " pod="openshift-marketplace/certified-operators-mblh7" Oct 06 10:02:54 crc kubenswrapper[4824]: I1006 10:02:54.991320 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ll4bb\" (UniqueName: \"kubernetes.io/projected/2dc6af8a-35a7-4fa8-9e03-eb6df841f127-kube-api-access-ll4bb\") pod \"certified-operators-mblh7\" (UID: \"2dc6af8a-35a7-4fa8-9e03-eb6df841f127\") " pod="openshift-marketplace/certified-operators-mblh7" Oct 06 10:02:55 crc kubenswrapper[4824]: I1006 10:02:55.066570 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nk77w\" (UniqueName: \"kubernetes.io/projected/4519644a-379b-451a-9dd9-fb528ced61bf-kube-api-access-nk77w\") pod \"community-operators-s2nbh\" (UID: \"4519644a-379b-451a-9dd9-fb528ced61bf\") " pod="openshift-marketplace/community-operators-s2nbh" Oct 06 10:02:55 crc kubenswrapper[4824]: I1006 10:02:55.067288 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4519644a-379b-451a-9dd9-fb528ced61bf-utilities\") pod \"community-operators-s2nbh\" (UID: \"4519644a-379b-451a-9dd9-fb528ced61bf\") " pod="openshift-marketplace/community-operators-s2nbh" Oct 06 10:02:55 crc kubenswrapper[4824]: I1006 10:02:55.067321 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4519644a-379b-451a-9dd9-fb528ced61bf-catalog-content\") pod \"community-operators-s2nbh\" (UID: \"4519644a-379b-451a-9dd9-fb528ced61bf\") " pod="openshift-marketplace/community-operators-s2nbh" Oct 06 10:02:55 crc kubenswrapper[4824]: I1006 10:02:55.121670 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mblh7" Oct 06 10:02:55 crc kubenswrapper[4824]: I1006 10:02:55.168920 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4519644a-379b-451a-9dd9-fb528ced61bf-utilities\") pod \"community-operators-s2nbh\" (UID: \"4519644a-379b-451a-9dd9-fb528ced61bf\") " pod="openshift-marketplace/community-operators-s2nbh" Oct 06 10:02:55 crc kubenswrapper[4824]: I1006 10:02:55.169003 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4519644a-379b-451a-9dd9-fb528ced61bf-catalog-content\") pod \"community-operators-s2nbh\" (UID: \"4519644a-379b-451a-9dd9-fb528ced61bf\") " pod="openshift-marketplace/community-operators-s2nbh" Oct 06 10:02:55 crc kubenswrapper[4824]: I1006 10:02:55.169087 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nk77w\" (UniqueName: \"kubernetes.io/projected/4519644a-379b-451a-9dd9-fb528ced61bf-kube-api-access-nk77w\") pod \"community-operators-s2nbh\" (UID: \"4519644a-379b-451a-9dd9-fb528ced61bf\") " pod="openshift-marketplace/community-operators-s2nbh" Oct 06 10:02:55 crc kubenswrapper[4824]: I1006 10:02:55.170302 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4519644a-379b-451a-9dd9-fb528ced61bf-utilities\") pod \"community-operators-s2nbh\" (UID: \"4519644a-379b-451a-9dd9-fb528ced61bf\") " pod="openshift-marketplace/community-operators-s2nbh" Oct 06 10:02:55 crc kubenswrapper[4824]: I1006 10:02:55.170577 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4519644a-379b-451a-9dd9-fb528ced61bf-catalog-content\") pod \"community-operators-s2nbh\" (UID: \"4519644a-379b-451a-9dd9-fb528ced61bf\") " pod="openshift-marketplace/community-operators-s2nbh" Oct 06 10:02:55 crc kubenswrapper[4824]: I1006 10:02:55.189675 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nk77w\" (UniqueName: \"kubernetes.io/projected/4519644a-379b-451a-9dd9-fb528ced61bf-kube-api-access-nk77w\") pod \"community-operators-s2nbh\" (UID: \"4519644a-379b-451a-9dd9-fb528ced61bf\") " pod="openshift-marketplace/community-operators-s2nbh" Oct 06 10:02:55 crc kubenswrapper[4824]: I1006 10:02:55.270021 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-s2nbh" Oct 06 10:02:55 crc kubenswrapper[4824]: I1006 10:02:55.558499 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mblh7"] Oct 06 10:02:55 crc kubenswrapper[4824]: W1006 10:02:55.581421 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2dc6af8a_35a7_4fa8_9e03_eb6df841f127.slice/crio-37f2e50b237316a41f34f968e5d1a9348c9a7bc081cbb1ddf6ad4ed0722243c8 WatchSource:0}: Error finding container 37f2e50b237316a41f34f968e5d1a9348c9a7bc081cbb1ddf6ad4ed0722243c8: Status 404 returned error can't find the container with id 37f2e50b237316a41f34f968e5d1a9348c9a7bc081cbb1ddf6ad4ed0722243c8 Oct 06 10:02:55 crc kubenswrapper[4824]: I1006 10:02:55.676620 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-s2nbh"] Oct 06 10:02:55 crc kubenswrapper[4824]: W1006 10:02:55.705793 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4519644a_379b_451a_9dd9_fb528ced61bf.slice/crio-8a7b43a456d2aba71e4351ac018c321ef916292a97750dc975e5800493344113 WatchSource:0}: Error finding container 8a7b43a456d2aba71e4351ac018c321ef916292a97750dc975e5800493344113: Status 404 returned error can't find the container with id 8a7b43a456d2aba71e4351ac018c321ef916292a97750dc975e5800493344113 Oct 06 10:02:55 crc kubenswrapper[4824]: I1006 10:02:55.784917 4824 generic.go:334] "Generic (PLEG): container finished" podID="5fe63731-833a-45c3-b4f0-73ddc05383e1" containerID="4608802294a820026e0deae051a3356124536fdfe9af562cd07233f5f8546bc7" exitCode=0 Oct 06 10:02:55 crc kubenswrapper[4824]: I1006 10:02:55.784989 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-txrmb" event={"ID":"5fe63731-833a-45c3-b4f0-73ddc05383e1","Type":"ContainerDied","Data":"4608802294a820026e0deae051a3356124536fdfe9af562cd07233f5f8546bc7"} Oct 06 10:02:55 crc kubenswrapper[4824]: I1006 10:02:55.787692 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mblh7" event={"ID":"2dc6af8a-35a7-4fa8-9e03-eb6df841f127","Type":"ContainerStarted","Data":"37f2e50b237316a41f34f968e5d1a9348c9a7bc081cbb1ddf6ad4ed0722243c8"} Oct 06 10:02:55 crc kubenswrapper[4824]: I1006 10:02:55.790117 4824 generic.go:334] "Generic (PLEG): container finished" podID="504bf3d4-9589-4a5e-ad25-e16f783e0115" containerID="7af1c406247f195684dc014e7b841f231a2af27f6938250a06b0bce6f29be813" exitCode=0 Oct 06 10:02:55 crc kubenswrapper[4824]: I1006 10:02:55.790151 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xdszq" event={"ID":"504bf3d4-9589-4a5e-ad25-e16f783e0115","Type":"ContainerDied","Data":"7af1c406247f195684dc014e7b841f231a2af27f6938250a06b0bce6f29be813"} Oct 06 10:02:55 crc kubenswrapper[4824]: I1006 10:02:55.791836 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-s2nbh" event={"ID":"4519644a-379b-451a-9dd9-fb528ced61bf","Type":"ContainerStarted","Data":"8a7b43a456d2aba71e4351ac018c321ef916292a97750dc975e5800493344113"} Oct 06 10:02:56 crc kubenswrapper[4824]: I1006 10:02:56.798773 4824 generic.go:334] "Generic (PLEG): container finished" podID="4519644a-379b-451a-9dd9-fb528ced61bf" containerID="ac1b729047f224db28da183cfa20adca0f759766a9ea0204532acd58b5e3f184" exitCode=0 Oct 06 10:02:56 crc kubenswrapper[4824]: I1006 10:02:56.798882 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-s2nbh" event={"ID":"4519644a-379b-451a-9dd9-fb528ced61bf","Type":"ContainerDied","Data":"ac1b729047f224db28da183cfa20adca0f759766a9ea0204532acd58b5e3f184"} Oct 06 10:02:56 crc kubenswrapper[4824]: I1006 10:02:56.807874 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-txrmb" event={"ID":"5fe63731-833a-45c3-b4f0-73ddc05383e1","Type":"ContainerStarted","Data":"363a47a824f4f4cf7625979b822234ca2e0ecdcc86a4192924e3b14ce374e3e1"} Oct 06 10:02:56 crc kubenswrapper[4824]: I1006 10:02:56.811240 4824 generic.go:334] "Generic (PLEG): container finished" podID="2dc6af8a-35a7-4fa8-9e03-eb6df841f127" containerID="02a73e92f7d16e4aa0cb1e98a9588538f1d54bc5019897923572105db4ab5faa" exitCode=0 Oct 06 10:02:56 crc kubenswrapper[4824]: I1006 10:02:56.811297 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mblh7" event={"ID":"2dc6af8a-35a7-4fa8-9e03-eb6df841f127","Type":"ContainerDied","Data":"02a73e92f7d16e4aa0cb1e98a9588538f1d54bc5019897923572105db4ab5faa"} Oct 06 10:02:56 crc kubenswrapper[4824]: I1006 10:02:56.824231 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xdszq" event={"ID":"504bf3d4-9589-4a5e-ad25-e16f783e0115","Type":"ContainerStarted","Data":"8e72577e764d83b2a4ce72144a72fc9ba51fa10a35b088b06d6c4186b8c604a8"} Oct 06 10:02:56 crc kubenswrapper[4824]: I1006 10:02:56.855589 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-txrmb" podStartSLOduration=2.113561247 podStartE2EDuration="4.855572156s" podCreationTimestamp="2025-10-06 10:02:52 +0000 UTC" firstStartedPulling="2025-10-06 10:02:53.747222944 +0000 UTC m=+283.111645815" lastFinishedPulling="2025-10-06 10:02:56.489233823 +0000 UTC m=+285.853656724" observedRunningTime="2025-10-06 10:02:56.853848191 +0000 UTC m=+286.218271052" watchObservedRunningTime="2025-10-06 10:02:56.855572156 +0000 UTC m=+286.219995017" Oct 06 10:02:56 crc kubenswrapper[4824]: I1006 10:02:56.874966 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-xdszq" podStartSLOduration=2.101683979 podStartE2EDuration="4.874940919s" podCreationTimestamp="2025-10-06 10:02:52 +0000 UTC" firstStartedPulling="2025-10-06 10:02:53.749151764 +0000 UTC m=+283.113574665" lastFinishedPulling="2025-10-06 10:02:56.522408734 +0000 UTC m=+285.886831605" observedRunningTime="2025-10-06 10:02:56.872719912 +0000 UTC m=+286.237142773" watchObservedRunningTime="2025-10-06 10:02:56.874940919 +0000 UTC m=+286.239363780" Oct 06 10:02:57 crc kubenswrapper[4824]: I1006 10:02:57.834317 4824 generic.go:334] "Generic (PLEG): container finished" podID="4519644a-379b-451a-9dd9-fb528ced61bf" containerID="784ce105bcc2af26312c7fdd37c27911658788af04773f7aca6a06e2dd8894ec" exitCode=0 Oct 06 10:02:57 crc kubenswrapper[4824]: I1006 10:02:57.834377 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-s2nbh" event={"ID":"4519644a-379b-451a-9dd9-fb528ced61bf","Type":"ContainerDied","Data":"784ce105bcc2af26312c7fdd37c27911658788af04773f7aca6a06e2dd8894ec"} Oct 06 10:02:59 crc kubenswrapper[4824]: I1006 10:02:59.854780 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-s2nbh" event={"ID":"4519644a-379b-451a-9dd9-fb528ced61bf","Type":"ContainerStarted","Data":"7461b687445bbe5bc305f24d78140a751c9150c4e0f11b23736f5409dc1f7a84"} Oct 06 10:02:59 crc kubenswrapper[4824]: I1006 10:02:59.859286 4824 generic.go:334] "Generic (PLEG): container finished" podID="2dc6af8a-35a7-4fa8-9e03-eb6df841f127" containerID="a40bb9c79b960cea486e6069930b8ba3138ef1280e11ef7ce75852d7318e04f5" exitCode=0 Oct 06 10:02:59 crc kubenswrapper[4824]: I1006 10:02:59.859348 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mblh7" event={"ID":"2dc6af8a-35a7-4fa8-9e03-eb6df841f127","Type":"ContainerDied","Data":"a40bb9c79b960cea486e6069930b8ba3138ef1280e11ef7ce75852d7318e04f5"} Oct 06 10:02:59 crc kubenswrapper[4824]: I1006 10:02:59.883529 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-s2nbh" podStartSLOduration=4.064276685 podStartE2EDuration="5.88351203s" podCreationTimestamp="2025-10-06 10:02:54 +0000 UTC" firstStartedPulling="2025-10-06 10:02:56.801668076 +0000 UTC m=+286.166090977" lastFinishedPulling="2025-10-06 10:02:58.620903461 +0000 UTC m=+287.985326322" observedRunningTime="2025-10-06 10:02:59.879930817 +0000 UTC m=+289.244353678" watchObservedRunningTime="2025-10-06 10:02:59.88351203 +0000 UTC m=+289.247934891" Oct 06 10:03:00 crc kubenswrapper[4824]: I1006 10:03:00.868692 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mblh7" event={"ID":"2dc6af8a-35a7-4fa8-9e03-eb6df841f127","Type":"ContainerStarted","Data":"c8396ec36cf27e34346218350a4b3de10f354e9d4518ed7e6c76246bacbd2ab8"} Oct 06 10:03:00 crc kubenswrapper[4824]: I1006 10:03:00.891619 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-mblh7" podStartSLOduration=3.07937289 podStartE2EDuration="6.89159812s" podCreationTimestamp="2025-10-06 10:02:54 +0000 UTC" firstStartedPulling="2025-10-06 10:02:56.812807956 +0000 UTC m=+286.177230817" lastFinishedPulling="2025-10-06 10:03:00.625033186 +0000 UTC m=+289.989456047" observedRunningTime="2025-10-06 10:03:00.888655773 +0000 UTC m=+290.253078644" watchObservedRunningTime="2025-10-06 10:03:00.89159812 +0000 UTC m=+290.256020981" Oct 06 10:03:02 crc kubenswrapper[4824]: I1006 10:03:02.715262 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-xdszq" Oct 06 10:03:02 crc kubenswrapper[4824]: I1006 10:03:02.715764 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-xdszq" Oct 06 10:03:02 crc kubenswrapper[4824]: I1006 10:03:02.761160 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-xdszq" Oct 06 10:03:02 crc kubenswrapper[4824]: I1006 10:03:02.859966 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-txrmb" Oct 06 10:03:02 crc kubenswrapper[4824]: I1006 10:03:02.860033 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-txrmb" Oct 06 10:03:02 crc kubenswrapper[4824]: I1006 10:03:02.904265 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-txrmb" Oct 06 10:03:02 crc kubenswrapper[4824]: I1006 10:03:02.923556 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-xdszq" Oct 06 10:03:02 crc kubenswrapper[4824]: I1006 10:03:02.959458 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-txrmb" Oct 06 10:03:05 crc kubenswrapper[4824]: I1006 10:03:05.122580 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-mblh7" Oct 06 10:03:05 crc kubenswrapper[4824]: I1006 10:03:05.123031 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-mblh7" Oct 06 10:03:05 crc kubenswrapper[4824]: I1006 10:03:05.171169 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-mblh7" Oct 06 10:03:05 crc kubenswrapper[4824]: I1006 10:03:05.271129 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-s2nbh" Oct 06 10:03:05 crc kubenswrapper[4824]: I1006 10:03:05.271270 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-s2nbh" Oct 06 10:03:05 crc kubenswrapper[4824]: I1006 10:03:05.322175 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-s2nbh" Oct 06 10:03:05 crc kubenswrapper[4824]: I1006 10:03:05.931711 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-s2nbh" Oct 06 10:03:05 crc kubenswrapper[4824]: I1006 10:03:05.938714 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-mblh7" Oct 06 10:04:13 crc kubenswrapper[4824]: I1006 10:04:13.915475 4824 patch_prober.go:28] interesting pod/machine-config-daemon-khgzw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 10:04:13 crc kubenswrapper[4824]: I1006 10:04:13.916362 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 10:04:43 crc kubenswrapper[4824]: I1006 10:04:43.915952 4824 patch_prober.go:28] interesting pod/machine-config-daemon-khgzw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 10:04:43 crc kubenswrapper[4824]: I1006 10:04:43.917204 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 10:05:13 crc kubenswrapper[4824]: I1006 10:05:13.915706 4824 patch_prober.go:28] interesting pod/machine-config-daemon-khgzw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 10:05:13 crc kubenswrapper[4824]: I1006 10:05:13.916619 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 10:05:13 crc kubenswrapper[4824]: I1006 10:05:13.916715 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" Oct 06 10:05:13 crc kubenswrapper[4824]: I1006 10:05:13.917940 4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"849cbe616fe0e1fc3c60c65e540b59cf8cac7989da24b5c2d88ead6d16645186"} pod="openshift-machine-config-operator/machine-config-daemon-khgzw" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 06 10:05:13 crc kubenswrapper[4824]: I1006 10:05:13.918121 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" containerName="machine-config-daemon" containerID="cri-o://849cbe616fe0e1fc3c60c65e540b59cf8cac7989da24b5c2d88ead6d16645186" gracePeriod=600 Oct 06 10:05:14 crc kubenswrapper[4824]: I1006 10:05:14.913055 4824 generic.go:334] "Generic (PLEG): container finished" podID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" containerID="849cbe616fe0e1fc3c60c65e540b59cf8cac7989da24b5c2d88ead6d16645186" exitCode=0 Oct 06 10:05:14 crc kubenswrapper[4824]: I1006 10:05:14.913153 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" event={"ID":"e1140c8c-93f1-4fce-8c06-a69561ff8a6d","Type":"ContainerDied","Data":"849cbe616fe0e1fc3c60c65e540b59cf8cac7989da24b5c2d88ead6d16645186"} Oct 06 10:05:14 crc kubenswrapper[4824]: I1006 10:05:14.914044 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" event={"ID":"e1140c8c-93f1-4fce-8c06-a69561ff8a6d","Type":"ContainerStarted","Data":"d6acecda0f23f881c285040e9b7cdcc459c602934016ff58bff3f9c6edb5dbaf"} Oct 06 10:05:14 crc kubenswrapper[4824]: I1006 10:05:14.914094 4824 scope.go:117] "RemoveContainer" containerID="5512be36013baeca611ee92645a1d339da237b70d63aac43334beafe24513d44" Oct 06 10:05:51 crc kubenswrapper[4824]: I1006 10:05:51.616674 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-lm66v"] Oct 06 10:05:51 crc kubenswrapper[4824]: I1006 10:05:51.618774 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-lm66v" Oct 06 10:05:51 crc kubenswrapper[4824]: I1006 10:05:51.636317 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-lm66v"] Oct 06 10:05:51 crc kubenswrapper[4824]: I1006 10:05:51.688803 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/cadaa7f4-8302-478e-986f-a1db0380fcf3-ca-trust-extracted\") pod \"image-registry-66df7c8f76-lm66v\" (UID: \"cadaa7f4-8302-478e-986f-a1db0380fcf3\") " pod="openshift-image-registry/image-registry-66df7c8f76-lm66v" Oct 06 10:05:51 crc kubenswrapper[4824]: I1006 10:05:51.688863 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/cadaa7f4-8302-478e-986f-a1db0380fcf3-registry-certificates\") pod \"image-registry-66df7c8f76-lm66v\" (UID: \"cadaa7f4-8302-478e-986f-a1db0380fcf3\") " pod="openshift-image-registry/image-registry-66df7c8f76-lm66v" Oct 06 10:05:51 crc kubenswrapper[4824]: I1006 10:05:51.688905 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/cadaa7f4-8302-478e-986f-a1db0380fcf3-bound-sa-token\") pod \"image-registry-66df7c8f76-lm66v\" (UID: \"cadaa7f4-8302-478e-986f-a1db0380fcf3\") " pod="openshift-image-registry/image-registry-66df7c8f76-lm66v" Oct 06 10:05:51 crc kubenswrapper[4824]: I1006 10:05:51.688933 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bfvhn\" (UniqueName: \"kubernetes.io/projected/cadaa7f4-8302-478e-986f-a1db0380fcf3-kube-api-access-bfvhn\") pod \"image-registry-66df7c8f76-lm66v\" (UID: \"cadaa7f4-8302-478e-986f-a1db0380fcf3\") " pod="openshift-image-registry/image-registry-66df7c8f76-lm66v" Oct 06 10:05:51 crc kubenswrapper[4824]: I1006 10:05:51.688963 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/cadaa7f4-8302-478e-986f-a1db0380fcf3-installation-pull-secrets\") pod \"image-registry-66df7c8f76-lm66v\" (UID: \"cadaa7f4-8302-478e-986f-a1db0380fcf3\") " pod="openshift-image-registry/image-registry-66df7c8f76-lm66v" Oct 06 10:05:51 crc kubenswrapper[4824]: I1006 10:05:51.689100 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-lm66v\" (UID: \"cadaa7f4-8302-478e-986f-a1db0380fcf3\") " pod="openshift-image-registry/image-registry-66df7c8f76-lm66v" Oct 06 10:05:51 crc kubenswrapper[4824]: I1006 10:05:51.689365 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/cadaa7f4-8302-478e-986f-a1db0380fcf3-trusted-ca\") pod \"image-registry-66df7c8f76-lm66v\" (UID: \"cadaa7f4-8302-478e-986f-a1db0380fcf3\") " pod="openshift-image-registry/image-registry-66df7c8f76-lm66v" Oct 06 10:05:51 crc kubenswrapper[4824]: I1006 10:05:51.689445 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/cadaa7f4-8302-478e-986f-a1db0380fcf3-registry-tls\") pod \"image-registry-66df7c8f76-lm66v\" (UID: \"cadaa7f4-8302-478e-986f-a1db0380fcf3\") " pod="openshift-image-registry/image-registry-66df7c8f76-lm66v" Oct 06 10:05:51 crc kubenswrapper[4824]: I1006 10:05:51.720648 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-lm66v\" (UID: \"cadaa7f4-8302-478e-986f-a1db0380fcf3\") " pod="openshift-image-registry/image-registry-66df7c8f76-lm66v" Oct 06 10:05:51 crc kubenswrapper[4824]: I1006 10:05:51.791546 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bfvhn\" (UniqueName: \"kubernetes.io/projected/cadaa7f4-8302-478e-986f-a1db0380fcf3-kube-api-access-bfvhn\") pod \"image-registry-66df7c8f76-lm66v\" (UID: \"cadaa7f4-8302-478e-986f-a1db0380fcf3\") " pod="openshift-image-registry/image-registry-66df7c8f76-lm66v" Oct 06 10:05:51 crc kubenswrapper[4824]: I1006 10:05:51.792146 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/cadaa7f4-8302-478e-986f-a1db0380fcf3-installation-pull-secrets\") pod \"image-registry-66df7c8f76-lm66v\" (UID: \"cadaa7f4-8302-478e-986f-a1db0380fcf3\") " pod="openshift-image-registry/image-registry-66df7c8f76-lm66v" Oct 06 10:05:51 crc kubenswrapper[4824]: I1006 10:05:51.792263 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/cadaa7f4-8302-478e-986f-a1db0380fcf3-trusted-ca\") pod \"image-registry-66df7c8f76-lm66v\" (UID: \"cadaa7f4-8302-478e-986f-a1db0380fcf3\") " pod="openshift-image-registry/image-registry-66df7c8f76-lm66v" Oct 06 10:05:51 crc kubenswrapper[4824]: I1006 10:05:51.792294 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/cadaa7f4-8302-478e-986f-a1db0380fcf3-registry-tls\") pod \"image-registry-66df7c8f76-lm66v\" (UID: \"cadaa7f4-8302-478e-986f-a1db0380fcf3\") " pod="openshift-image-registry/image-registry-66df7c8f76-lm66v" Oct 06 10:05:51 crc kubenswrapper[4824]: I1006 10:05:51.792328 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/cadaa7f4-8302-478e-986f-a1db0380fcf3-ca-trust-extracted\") pod \"image-registry-66df7c8f76-lm66v\" (UID: \"cadaa7f4-8302-478e-986f-a1db0380fcf3\") " pod="openshift-image-registry/image-registry-66df7c8f76-lm66v" Oct 06 10:05:51 crc kubenswrapper[4824]: I1006 10:05:51.792362 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/cadaa7f4-8302-478e-986f-a1db0380fcf3-registry-certificates\") pod \"image-registry-66df7c8f76-lm66v\" (UID: \"cadaa7f4-8302-478e-986f-a1db0380fcf3\") " pod="openshift-image-registry/image-registry-66df7c8f76-lm66v" Oct 06 10:05:51 crc kubenswrapper[4824]: I1006 10:05:51.792410 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/cadaa7f4-8302-478e-986f-a1db0380fcf3-bound-sa-token\") pod \"image-registry-66df7c8f76-lm66v\" (UID: \"cadaa7f4-8302-478e-986f-a1db0380fcf3\") " pod="openshift-image-registry/image-registry-66df7c8f76-lm66v" Oct 06 10:05:51 crc kubenswrapper[4824]: I1006 10:05:51.793180 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/cadaa7f4-8302-478e-986f-a1db0380fcf3-ca-trust-extracted\") pod \"image-registry-66df7c8f76-lm66v\" (UID: \"cadaa7f4-8302-478e-986f-a1db0380fcf3\") " pod="openshift-image-registry/image-registry-66df7c8f76-lm66v" Oct 06 10:05:51 crc kubenswrapper[4824]: I1006 10:05:51.795518 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/cadaa7f4-8302-478e-986f-a1db0380fcf3-registry-certificates\") pod \"image-registry-66df7c8f76-lm66v\" (UID: \"cadaa7f4-8302-478e-986f-a1db0380fcf3\") " pod="openshift-image-registry/image-registry-66df7c8f76-lm66v" Oct 06 10:05:51 crc kubenswrapper[4824]: I1006 10:05:51.795892 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/cadaa7f4-8302-478e-986f-a1db0380fcf3-trusted-ca\") pod \"image-registry-66df7c8f76-lm66v\" (UID: \"cadaa7f4-8302-478e-986f-a1db0380fcf3\") " pod="openshift-image-registry/image-registry-66df7c8f76-lm66v" Oct 06 10:05:51 crc kubenswrapper[4824]: I1006 10:05:51.801768 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/cadaa7f4-8302-478e-986f-a1db0380fcf3-registry-tls\") pod \"image-registry-66df7c8f76-lm66v\" (UID: \"cadaa7f4-8302-478e-986f-a1db0380fcf3\") " pod="openshift-image-registry/image-registry-66df7c8f76-lm66v" Oct 06 10:05:51 crc kubenswrapper[4824]: I1006 10:05:51.806733 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/cadaa7f4-8302-478e-986f-a1db0380fcf3-installation-pull-secrets\") pod \"image-registry-66df7c8f76-lm66v\" (UID: \"cadaa7f4-8302-478e-986f-a1db0380fcf3\") " pod="openshift-image-registry/image-registry-66df7c8f76-lm66v" Oct 06 10:05:51 crc kubenswrapper[4824]: I1006 10:05:51.819190 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/cadaa7f4-8302-478e-986f-a1db0380fcf3-bound-sa-token\") pod \"image-registry-66df7c8f76-lm66v\" (UID: \"cadaa7f4-8302-478e-986f-a1db0380fcf3\") " pod="openshift-image-registry/image-registry-66df7c8f76-lm66v" Oct 06 10:05:51 crc kubenswrapper[4824]: I1006 10:05:51.823210 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bfvhn\" (UniqueName: \"kubernetes.io/projected/cadaa7f4-8302-478e-986f-a1db0380fcf3-kube-api-access-bfvhn\") pod \"image-registry-66df7c8f76-lm66v\" (UID: \"cadaa7f4-8302-478e-986f-a1db0380fcf3\") " pod="openshift-image-registry/image-registry-66df7c8f76-lm66v" Oct 06 10:05:51 crc kubenswrapper[4824]: I1006 10:05:51.938623 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-lm66v" Oct 06 10:05:52 crc kubenswrapper[4824]: I1006 10:05:52.231742 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-lm66v"] Oct 06 10:05:53 crc kubenswrapper[4824]: I1006 10:05:53.211433 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-lm66v" event={"ID":"cadaa7f4-8302-478e-986f-a1db0380fcf3","Type":"ContainerStarted","Data":"b62a719ede1d2f4c5eeb60e7cae154c1b94201babc82da57a20b5ec8b8cbdd8b"} Oct 06 10:05:53 crc kubenswrapper[4824]: I1006 10:05:53.212046 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-lm66v" Oct 06 10:05:53 crc kubenswrapper[4824]: I1006 10:05:53.212071 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-lm66v" event={"ID":"cadaa7f4-8302-478e-986f-a1db0380fcf3","Type":"ContainerStarted","Data":"16a0bfb44527ea3b0294554dc1eb6832b2aa5e9c0df0bbfcb1fa810fe911a9b3"} Oct 06 10:05:53 crc kubenswrapper[4824]: I1006 10:05:53.242436 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-lm66v" podStartSLOduration=2.242410309 podStartE2EDuration="2.242410309s" podCreationTimestamp="2025-10-06 10:05:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:05:53.237885996 +0000 UTC m=+462.602308897" watchObservedRunningTime="2025-10-06 10:05:53.242410309 +0000 UTC m=+462.606833200" Oct 06 10:06:11 crc kubenswrapper[4824]: I1006 10:06:11.947894 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-lm66v" Oct 06 10:06:12 crc kubenswrapper[4824]: I1006 10:06:12.025102 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-8l8gd"] Oct 06 10:06:37 crc kubenswrapper[4824]: I1006 10:06:37.099924 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-8l8gd" podUID="c2f6e84b-a28c-4ead-97f5-978b0639e5bf" containerName="registry" containerID="cri-o://f379e7ee906ebe5d001d226f8d5629b56c3d3313684aac7bd4ea8a156413ffb5" gracePeriod=30 Oct 06 10:06:37 crc kubenswrapper[4824]: I1006 10:06:37.540549 4824 generic.go:334] "Generic (PLEG): container finished" podID="c2f6e84b-a28c-4ead-97f5-978b0639e5bf" containerID="f379e7ee906ebe5d001d226f8d5629b56c3d3313684aac7bd4ea8a156413ffb5" exitCode=0 Oct 06 10:06:37 crc kubenswrapper[4824]: I1006 10:06:37.541506 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-8l8gd" event={"ID":"c2f6e84b-a28c-4ead-97f5-978b0639e5bf","Type":"ContainerDied","Data":"f379e7ee906ebe5d001d226f8d5629b56c3d3313684aac7bd4ea8a156413ffb5"} Oct 06 10:06:37 crc kubenswrapper[4824]: I1006 10:06:37.601203 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-8l8gd" Oct 06 10:06:37 crc kubenswrapper[4824]: I1006 10:06:37.730611 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/c2f6e84b-a28c-4ead-97f5-978b0639e5bf-registry-certificates\") pod \"c2f6e84b-a28c-4ead-97f5-978b0639e5bf\" (UID: \"c2f6e84b-a28c-4ead-97f5-978b0639e5bf\") " Oct 06 10:06:37 crc kubenswrapper[4824]: I1006 10:06:37.730745 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c2f6e84b-a28c-4ead-97f5-978b0639e5bf-bound-sa-token\") pod \"c2f6e84b-a28c-4ead-97f5-978b0639e5bf\" (UID: \"c2f6e84b-a28c-4ead-97f5-978b0639e5bf\") " Oct 06 10:06:37 crc kubenswrapper[4824]: I1006 10:06:37.730888 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-66bhb\" (UniqueName: \"kubernetes.io/projected/c2f6e84b-a28c-4ead-97f5-978b0639e5bf-kube-api-access-66bhb\") pod \"c2f6e84b-a28c-4ead-97f5-978b0639e5bf\" (UID: \"c2f6e84b-a28c-4ead-97f5-978b0639e5bf\") " Oct 06 10:06:37 crc kubenswrapper[4824]: I1006 10:06:37.731134 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"c2f6e84b-a28c-4ead-97f5-978b0639e5bf\" (UID: \"c2f6e84b-a28c-4ead-97f5-978b0639e5bf\") " Oct 06 10:06:37 crc kubenswrapper[4824]: I1006 10:06:37.731238 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c2f6e84b-a28c-4ead-97f5-978b0639e5bf-trusted-ca\") pod \"c2f6e84b-a28c-4ead-97f5-978b0639e5bf\" (UID: \"c2f6e84b-a28c-4ead-97f5-978b0639e5bf\") " Oct 06 10:06:37 crc kubenswrapper[4824]: I1006 10:06:37.731306 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/c2f6e84b-a28c-4ead-97f5-978b0639e5bf-ca-trust-extracted\") pod \"c2f6e84b-a28c-4ead-97f5-978b0639e5bf\" (UID: \"c2f6e84b-a28c-4ead-97f5-978b0639e5bf\") " Oct 06 10:06:37 crc kubenswrapper[4824]: I1006 10:06:37.731360 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/c2f6e84b-a28c-4ead-97f5-978b0639e5bf-installation-pull-secrets\") pod \"c2f6e84b-a28c-4ead-97f5-978b0639e5bf\" (UID: \"c2f6e84b-a28c-4ead-97f5-978b0639e5bf\") " Oct 06 10:06:37 crc kubenswrapper[4824]: I1006 10:06:37.731450 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/c2f6e84b-a28c-4ead-97f5-978b0639e5bf-registry-tls\") pod \"c2f6e84b-a28c-4ead-97f5-978b0639e5bf\" (UID: \"c2f6e84b-a28c-4ead-97f5-978b0639e5bf\") " Oct 06 10:06:37 crc kubenswrapper[4824]: I1006 10:06:37.732151 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c2f6e84b-a28c-4ead-97f5-978b0639e5bf-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "c2f6e84b-a28c-4ead-97f5-978b0639e5bf" (UID: "c2f6e84b-a28c-4ead-97f5-978b0639e5bf"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:06:37 crc kubenswrapper[4824]: I1006 10:06:37.733292 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c2f6e84b-a28c-4ead-97f5-978b0639e5bf-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "c2f6e84b-a28c-4ead-97f5-978b0639e5bf" (UID: "c2f6e84b-a28c-4ead-97f5-978b0639e5bf"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:06:37 crc kubenswrapper[4824]: I1006 10:06:37.743406 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c2f6e84b-a28c-4ead-97f5-978b0639e5bf-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "c2f6e84b-a28c-4ead-97f5-978b0639e5bf" (UID: "c2f6e84b-a28c-4ead-97f5-978b0639e5bf"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:06:37 crc kubenswrapper[4824]: I1006 10:06:37.744013 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c2f6e84b-a28c-4ead-97f5-978b0639e5bf-kube-api-access-66bhb" (OuterVolumeSpecName: "kube-api-access-66bhb") pod "c2f6e84b-a28c-4ead-97f5-978b0639e5bf" (UID: "c2f6e84b-a28c-4ead-97f5-978b0639e5bf"). InnerVolumeSpecName "kube-api-access-66bhb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:06:37 crc kubenswrapper[4824]: I1006 10:06:37.744247 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2f6e84b-a28c-4ead-97f5-978b0639e5bf-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "c2f6e84b-a28c-4ead-97f5-978b0639e5bf" (UID: "c2f6e84b-a28c-4ead-97f5-978b0639e5bf"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:06:37 crc kubenswrapper[4824]: I1006 10:06:37.744910 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c2f6e84b-a28c-4ead-97f5-978b0639e5bf-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "c2f6e84b-a28c-4ead-97f5-978b0639e5bf" (UID: "c2f6e84b-a28c-4ead-97f5-978b0639e5bf"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:06:37 crc kubenswrapper[4824]: I1006 10:06:37.763087 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "c2f6e84b-a28c-4ead-97f5-978b0639e5bf" (UID: "c2f6e84b-a28c-4ead-97f5-978b0639e5bf"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 06 10:06:37 crc kubenswrapper[4824]: I1006 10:06:37.774730 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c2f6e84b-a28c-4ead-97f5-978b0639e5bf-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "c2f6e84b-a28c-4ead-97f5-978b0639e5bf" (UID: "c2f6e84b-a28c-4ead-97f5-978b0639e5bf"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:06:37 crc kubenswrapper[4824]: I1006 10:06:37.840133 4824 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/c2f6e84b-a28c-4ead-97f5-978b0639e5bf-registry-tls\") on node \"crc\" DevicePath \"\"" Oct 06 10:06:37 crc kubenswrapper[4824]: I1006 10:06:37.840230 4824 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/c2f6e84b-a28c-4ead-97f5-978b0639e5bf-registry-certificates\") on node \"crc\" DevicePath \"\"" Oct 06 10:06:37 crc kubenswrapper[4824]: I1006 10:06:37.840260 4824 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c2f6e84b-a28c-4ead-97f5-978b0639e5bf-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 06 10:06:37 crc kubenswrapper[4824]: I1006 10:06:37.840286 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-66bhb\" (UniqueName: \"kubernetes.io/projected/c2f6e84b-a28c-4ead-97f5-978b0639e5bf-kube-api-access-66bhb\") on node \"crc\" DevicePath \"\"" Oct 06 10:06:37 crc kubenswrapper[4824]: I1006 10:06:37.840315 4824 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c2f6e84b-a28c-4ead-97f5-978b0639e5bf-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 06 10:06:37 crc kubenswrapper[4824]: I1006 10:06:37.840335 4824 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/c2f6e84b-a28c-4ead-97f5-978b0639e5bf-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Oct 06 10:06:37 crc kubenswrapper[4824]: I1006 10:06:37.840356 4824 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/c2f6e84b-a28c-4ead-97f5-978b0639e5bf-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Oct 06 10:06:38 crc kubenswrapper[4824]: I1006 10:06:38.553167 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-8l8gd" event={"ID":"c2f6e84b-a28c-4ead-97f5-978b0639e5bf","Type":"ContainerDied","Data":"bc9f21429286d2f02ba3385cf69c2113f5f39234e15a2aab5f3143545966c9ab"} Oct 06 10:06:38 crc kubenswrapper[4824]: I1006 10:06:38.554213 4824 scope.go:117] "RemoveContainer" containerID="f379e7ee906ebe5d001d226f8d5629b56c3d3313684aac7bd4ea8a156413ffb5" Oct 06 10:06:38 crc kubenswrapper[4824]: I1006 10:06:38.553286 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-8l8gd" Oct 06 10:06:38 crc kubenswrapper[4824]: I1006 10:06:38.605779 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-8l8gd"] Oct 06 10:06:38 crc kubenswrapper[4824]: I1006 10:06:38.610411 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-8l8gd"] Oct 06 10:06:39 crc kubenswrapper[4824]: I1006 10:06:39.286956 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c2f6e84b-a28c-4ead-97f5-978b0639e5bf" path="/var/lib/kubelet/pods/c2f6e84b-a28c-4ead-97f5-978b0639e5bf/volumes" Oct 06 10:07:43 crc kubenswrapper[4824]: I1006 10:07:43.915729 4824 patch_prober.go:28] interesting pod/machine-config-daemon-khgzw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 10:07:43 crc kubenswrapper[4824]: I1006 10:07:43.916285 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 10:08:13 crc kubenswrapper[4824]: I1006 10:08:13.211641 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-bbqpq"] Oct 06 10:08:13 crc kubenswrapper[4824]: E1006 10:08:13.212555 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2f6e84b-a28c-4ead-97f5-978b0639e5bf" containerName="registry" Oct 06 10:08:13 crc kubenswrapper[4824]: I1006 10:08:13.212572 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2f6e84b-a28c-4ead-97f5-978b0639e5bf" containerName="registry" Oct 06 10:08:13 crc kubenswrapper[4824]: I1006 10:08:13.212753 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="c2f6e84b-a28c-4ead-97f5-978b0639e5bf" containerName="registry" Oct 06 10:08:13 crc kubenswrapper[4824]: I1006 10:08:13.213241 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-bbqpq" Oct 06 10:08:13 crc kubenswrapper[4824]: I1006 10:08:13.216182 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Oct 06 10:08:13 crc kubenswrapper[4824]: I1006 10:08:13.216561 4824 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-kkpl4" Oct 06 10:08:13 crc kubenswrapper[4824]: I1006 10:08:13.218604 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Oct 06 10:08:13 crc kubenswrapper[4824]: I1006 10:08:13.228792 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-5b446d88c5-bdn2f"] Oct 06 10:08:13 crc kubenswrapper[4824]: I1006 10:08:13.231351 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-bdn2f" Oct 06 10:08:13 crc kubenswrapper[4824]: I1006 10:08:13.233923 4824 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-zhfzh" Oct 06 10:08:13 crc kubenswrapper[4824]: I1006 10:08:13.235554 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-bbqpq"] Oct 06 10:08:13 crc kubenswrapper[4824]: I1006 10:08:13.241218 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-htx98"] Oct 06 10:08:13 crc kubenswrapper[4824]: I1006 10:08:13.242260 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-htx98" Oct 06 10:08:13 crc kubenswrapper[4824]: I1006 10:08:13.244588 4824 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-xck78" Oct 06 10:08:13 crc kubenswrapper[4824]: I1006 10:08:13.263879 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-bdn2f"] Oct 06 10:08:13 crc kubenswrapper[4824]: I1006 10:08:13.306942 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-htx98"] Oct 06 10:08:13 crc kubenswrapper[4824]: I1006 10:08:13.339846 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-756zv\" (UniqueName: \"kubernetes.io/projected/b5759913-9cb4-4012-bffb-fe2f8085543c-kube-api-access-756zv\") pod \"cert-manager-cainjector-7f985d654d-bbqpq\" (UID: \"b5759913-9cb4-4012-bffb-fe2f8085543c\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-bbqpq" Oct 06 10:08:13 crc kubenswrapper[4824]: I1006 10:08:13.339888 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6fk7s\" (UniqueName: \"kubernetes.io/projected/bde1f651-2eda-4d2d-8ab5-5fcdcd4580a4-kube-api-access-6fk7s\") pod \"cert-manager-5b446d88c5-bdn2f\" (UID: \"bde1f651-2eda-4d2d-8ab5-5fcdcd4580a4\") " pod="cert-manager/cert-manager-5b446d88c5-bdn2f" Oct 06 10:08:13 crc kubenswrapper[4824]: I1006 10:08:13.440860 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-756zv\" (UniqueName: \"kubernetes.io/projected/b5759913-9cb4-4012-bffb-fe2f8085543c-kube-api-access-756zv\") pod \"cert-manager-cainjector-7f985d654d-bbqpq\" (UID: \"b5759913-9cb4-4012-bffb-fe2f8085543c\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-bbqpq" Oct 06 10:08:13 crc kubenswrapper[4824]: I1006 10:08:13.440930 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6fk7s\" (UniqueName: \"kubernetes.io/projected/bde1f651-2eda-4d2d-8ab5-5fcdcd4580a4-kube-api-access-6fk7s\") pod \"cert-manager-5b446d88c5-bdn2f\" (UID: \"bde1f651-2eda-4d2d-8ab5-5fcdcd4580a4\") " pod="cert-manager/cert-manager-5b446d88c5-bdn2f" Oct 06 10:08:13 crc kubenswrapper[4824]: I1006 10:08:13.441034 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8d4mq\" (UniqueName: \"kubernetes.io/projected/d6f420d5-ed6d-4201-b78f-0bfa304e3b6b-kube-api-access-8d4mq\") pod \"cert-manager-webhook-5655c58dd6-htx98\" (UID: \"d6f420d5-ed6d-4201-b78f-0bfa304e3b6b\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-htx98" Oct 06 10:08:13 crc kubenswrapper[4824]: I1006 10:08:13.461619 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6fk7s\" (UniqueName: \"kubernetes.io/projected/bde1f651-2eda-4d2d-8ab5-5fcdcd4580a4-kube-api-access-6fk7s\") pod \"cert-manager-5b446d88c5-bdn2f\" (UID: \"bde1f651-2eda-4d2d-8ab5-5fcdcd4580a4\") " pod="cert-manager/cert-manager-5b446d88c5-bdn2f" Oct 06 10:08:13 crc kubenswrapper[4824]: I1006 10:08:13.461664 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-756zv\" (UniqueName: \"kubernetes.io/projected/b5759913-9cb4-4012-bffb-fe2f8085543c-kube-api-access-756zv\") pod \"cert-manager-cainjector-7f985d654d-bbqpq\" (UID: \"b5759913-9cb4-4012-bffb-fe2f8085543c\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-bbqpq" Oct 06 10:08:13 crc kubenswrapper[4824]: I1006 10:08:13.534447 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-bbqpq" Oct 06 10:08:13 crc kubenswrapper[4824]: I1006 10:08:13.542835 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8d4mq\" (UniqueName: \"kubernetes.io/projected/d6f420d5-ed6d-4201-b78f-0bfa304e3b6b-kube-api-access-8d4mq\") pod \"cert-manager-webhook-5655c58dd6-htx98\" (UID: \"d6f420d5-ed6d-4201-b78f-0bfa304e3b6b\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-htx98" Oct 06 10:08:13 crc kubenswrapper[4824]: I1006 10:08:13.553843 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-bdn2f" Oct 06 10:08:13 crc kubenswrapper[4824]: I1006 10:08:13.563054 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8d4mq\" (UniqueName: \"kubernetes.io/projected/d6f420d5-ed6d-4201-b78f-0bfa304e3b6b-kube-api-access-8d4mq\") pod \"cert-manager-webhook-5655c58dd6-htx98\" (UID: \"d6f420d5-ed6d-4201-b78f-0bfa304e3b6b\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-htx98" Oct 06 10:08:13 crc kubenswrapper[4824]: I1006 10:08:13.569515 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-htx98" Oct 06 10:08:13 crc kubenswrapper[4824]: I1006 10:08:13.899192 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-bdn2f"] Oct 06 10:08:13 crc kubenswrapper[4824]: I1006 10:08:13.902470 4824 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 06 10:08:13 crc kubenswrapper[4824]: I1006 10:08:13.915676 4824 patch_prober.go:28] interesting pod/machine-config-daemon-khgzw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 10:08:13 crc kubenswrapper[4824]: I1006 10:08:13.916176 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 10:08:14 crc kubenswrapper[4824]: I1006 10:08:14.037529 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-htx98"] Oct 06 10:08:14 crc kubenswrapper[4824]: I1006 10:08:14.046376 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-bbqpq"] Oct 06 10:08:14 crc kubenswrapper[4824]: W1006 10:08:14.050735 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb5759913_9cb4_4012_bffb_fe2f8085543c.slice/crio-9bda057c6189df9bc99d83244af879292d61dcc5351aff689f905e6c34c027e6 WatchSource:0}: Error finding container 9bda057c6189df9bc99d83244af879292d61dcc5351aff689f905e6c34c027e6: Status 404 returned error can't find the container with id 9bda057c6189df9bc99d83244af879292d61dcc5351aff689f905e6c34c027e6 Oct 06 10:08:14 crc kubenswrapper[4824]: I1006 10:08:14.286897 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-bbqpq" event={"ID":"b5759913-9cb4-4012-bffb-fe2f8085543c","Type":"ContainerStarted","Data":"9bda057c6189df9bc99d83244af879292d61dcc5351aff689f905e6c34c027e6"} Oct 06 10:08:14 crc kubenswrapper[4824]: I1006 10:08:14.289055 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-htx98" event={"ID":"d6f420d5-ed6d-4201-b78f-0bfa304e3b6b","Type":"ContainerStarted","Data":"9e3e72af50ebe71c56641f16e4be02b32463c27b717e9531e76992edd1822e30"} Oct 06 10:08:14 crc kubenswrapper[4824]: I1006 10:08:14.290740 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-bdn2f" event={"ID":"bde1f651-2eda-4d2d-8ab5-5fcdcd4580a4","Type":"ContainerStarted","Data":"d716fb82dcec1e23a1595f57331a13b5ddbed9c957d29f8e3337ead3038ad2d3"} Oct 06 10:08:22 crc kubenswrapper[4824]: I1006 10:08:22.951138 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-4vwms"] Oct 06 10:08:22 crc kubenswrapper[4824]: I1006 10:08:22.952915 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" podUID="2f511ece-25eb-465f-b85c-9e5f3f886c21" containerName="ovn-controller" containerID="cri-o://09638fbc9d163ad0b63fa57d24ea5298207ad1f55bab8671be311960be2627a8" gracePeriod=30 Oct 06 10:08:22 crc kubenswrapper[4824]: I1006 10:08:22.953384 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" podUID="2f511ece-25eb-465f-b85c-9e5f3f886c21" containerName="northd" containerID="cri-o://6213164d366243c2139d5971041d6f076cf38fd0cce4c0ca5495111963487ec8" gracePeriod=30 Oct 06 10:08:22 crc kubenswrapper[4824]: I1006 10:08:22.953662 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" podUID="2f511ece-25eb-465f-b85c-9e5f3f886c21" containerName="nbdb" containerID="cri-o://9d6064ad3bcdafebb4c66e7d91f183cc787e9ce13660545c71a722b3ff04ce75" gracePeriod=30 Oct 06 10:08:22 crc kubenswrapper[4824]: I1006 10:08:22.953551 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" podUID="2f511ece-25eb-465f-b85c-9e5f3f886c21" containerName="sbdb" containerID="cri-o://0644c7db28c7d26aa8d69cd5b9fbcff4d4c06b96df113137dea22a72d9fecd85" gracePeriod=30 Oct 06 10:08:22 crc kubenswrapper[4824]: I1006 10:08:22.953850 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" podUID="2f511ece-25eb-465f-b85c-9e5f3f886c21" containerName="kube-rbac-proxy-node" containerID="cri-o://10bfed3e523b7422ce1e981f98166eca78bd12dd4a7cf0634915e1f4ecc3de2b" gracePeriod=30 Oct 06 10:08:22 crc kubenswrapper[4824]: I1006 10:08:22.953911 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" podUID="2f511ece-25eb-465f-b85c-9e5f3f886c21" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://43199f86ad4a1f02bc57a926cd96b2c3c0e2ea46fb36a3f495668706be4c8041" gracePeriod=30 Oct 06 10:08:22 crc kubenswrapper[4824]: I1006 10:08:22.953898 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" podUID="2f511ece-25eb-465f-b85c-9e5f3f886c21" containerName="ovn-acl-logging" containerID="cri-o://aa6d59fac987f60a4340ab35d84aa059f2b7085a06362e7207a79702d17eee6d" gracePeriod=30 Oct 06 10:08:23 crc kubenswrapper[4824]: I1006 10:08:23.037477 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" podUID="2f511ece-25eb-465f-b85c-9e5f3f886c21" containerName="ovnkube-controller" containerID="cri-o://3dffdccf2aac654a5a0f629b8c5b2233c4561cb036262068f2289f6e95b46575" gracePeriod=30 Oct 06 10:08:23 crc kubenswrapper[4824]: I1006 10:08:23.358326 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-szn8q_b65abf9f-4e88-4571-960d-3ca997d9c344/kube-multus/2.log" Oct 06 10:08:23 crc kubenswrapper[4824]: I1006 10:08:23.359543 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-szn8q_b65abf9f-4e88-4571-960d-3ca997d9c344/kube-multus/1.log" Oct 06 10:08:23 crc kubenswrapper[4824]: I1006 10:08:23.359605 4824 generic.go:334] "Generic (PLEG): container finished" podID="b65abf9f-4e88-4571-960d-3ca997d9c344" containerID="238ede7256cac751995c87da2f05e25727cc9496ddf838e9d57f365f50acf4ba" exitCode=2 Oct 06 10:08:23 crc kubenswrapper[4824]: I1006 10:08:23.359706 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-szn8q" event={"ID":"b65abf9f-4e88-4571-960d-3ca997d9c344","Type":"ContainerDied","Data":"238ede7256cac751995c87da2f05e25727cc9496ddf838e9d57f365f50acf4ba"} Oct 06 10:08:23 crc kubenswrapper[4824]: I1006 10:08:23.359797 4824 scope.go:117] "RemoveContainer" containerID="c80f7c0d85c4c76aa3c4ff4e078900ccc3be12fa5bb75f595ead6eeacf39e3f4" Oct 06 10:08:23 crc kubenswrapper[4824]: I1006 10:08:23.361479 4824 scope.go:117] "RemoveContainer" containerID="238ede7256cac751995c87da2f05e25727cc9496ddf838e9d57f365f50acf4ba" Oct 06 10:08:23 crc kubenswrapper[4824]: E1006 10:08:23.361788 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-szn8q_openshift-multus(b65abf9f-4e88-4571-960d-3ca997d9c344)\"" pod="openshift-multus/multus-szn8q" podUID="b65abf9f-4e88-4571-960d-3ca997d9c344" Oct 06 10:08:23 crc kubenswrapper[4824]: I1006 10:08:23.368227 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-4vwms_2f511ece-25eb-465f-b85c-9e5f3f886c21/ovnkube-controller/3.log" Oct 06 10:08:23 crc kubenswrapper[4824]: I1006 10:08:23.370727 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-4vwms_2f511ece-25eb-465f-b85c-9e5f3f886c21/ovn-acl-logging/0.log" Oct 06 10:08:23 crc kubenswrapper[4824]: I1006 10:08:23.371882 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-4vwms_2f511ece-25eb-465f-b85c-9e5f3f886c21/ovn-controller/0.log" Oct 06 10:08:23 crc kubenswrapper[4824]: I1006 10:08:23.372432 4824 generic.go:334] "Generic (PLEG): container finished" podID="2f511ece-25eb-465f-b85c-9e5f3f886c21" containerID="3dffdccf2aac654a5a0f629b8c5b2233c4561cb036262068f2289f6e95b46575" exitCode=0 Oct 06 10:08:23 crc kubenswrapper[4824]: I1006 10:08:23.372466 4824 generic.go:334] "Generic (PLEG): container finished" podID="2f511ece-25eb-465f-b85c-9e5f3f886c21" containerID="0644c7db28c7d26aa8d69cd5b9fbcff4d4c06b96df113137dea22a72d9fecd85" exitCode=0 Oct 06 10:08:23 crc kubenswrapper[4824]: I1006 10:08:23.372475 4824 generic.go:334] "Generic (PLEG): container finished" podID="2f511ece-25eb-465f-b85c-9e5f3f886c21" containerID="9d6064ad3bcdafebb4c66e7d91f183cc787e9ce13660545c71a722b3ff04ce75" exitCode=0 Oct 06 10:08:23 crc kubenswrapper[4824]: I1006 10:08:23.372484 4824 generic.go:334] "Generic (PLEG): container finished" podID="2f511ece-25eb-465f-b85c-9e5f3f886c21" containerID="6213164d366243c2139d5971041d6f076cf38fd0cce4c0ca5495111963487ec8" exitCode=0 Oct 06 10:08:23 crc kubenswrapper[4824]: I1006 10:08:23.372494 4824 generic.go:334] "Generic (PLEG): container finished" podID="2f511ece-25eb-465f-b85c-9e5f3f886c21" containerID="43199f86ad4a1f02bc57a926cd96b2c3c0e2ea46fb36a3f495668706be4c8041" exitCode=0 Oct 06 10:08:23 crc kubenswrapper[4824]: I1006 10:08:23.372504 4824 generic.go:334] "Generic (PLEG): container finished" podID="2f511ece-25eb-465f-b85c-9e5f3f886c21" containerID="10bfed3e523b7422ce1e981f98166eca78bd12dd4a7cf0634915e1f4ecc3de2b" exitCode=0 Oct 06 10:08:23 crc kubenswrapper[4824]: I1006 10:08:23.372514 4824 generic.go:334] "Generic (PLEG): container finished" podID="2f511ece-25eb-465f-b85c-9e5f3f886c21" containerID="aa6d59fac987f60a4340ab35d84aa059f2b7085a06362e7207a79702d17eee6d" exitCode=143 Oct 06 10:08:23 crc kubenswrapper[4824]: I1006 10:08:23.372524 4824 generic.go:334] "Generic (PLEG): container finished" podID="2f511ece-25eb-465f-b85c-9e5f3f886c21" containerID="09638fbc9d163ad0b63fa57d24ea5298207ad1f55bab8671be311960be2627a8" exitCode=143 Oct 06 10:08:23 crc kubenswrapper[4824]: I1006 10:08:23.372520 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" event={"ID":"2f511ece-25eb-465f-b85c-9e5f3f886c21","Type":"ContainerDied","Data":"3dffdccf2aac654a5a0f629b8c5b2233c4561cb036262068f2289f6e95b46575"} Oct 06 10:08:23 crc kubenswrapper[4824]: I1006 10:08:23.372580 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" event={"ID":"2f511ece-25eb-465f-b85c-9e5f3f886c21","Type":"ContainerDied","Data":"0644c7db28c7d26aa8d69cd5b9fbcff4d4c06b96df113137dea22a72d9fecd85"} Oct 06 10:08:23 crc kubenswrapper[4824]: I1006 10:08:23.372595 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" event={"ID":"2f511ece-25eb-465f-b85c-9e5f3f886c21","Type":"ContainerDied","Data":"9d6064ad3bcdafebb4c66e7d91f183cc787e9ce13660545c71a722b3ff04ce75"} Oct 06 10:08:23 crc kubenswrapper[4824]: I1006 10:08:23.372610 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" event={"ID":"2f511ece-25eb-465f-b85c-9e5f3f886c21","Type":"ContainerDied","Data":"6213164d366243c2139d5971041d6f076cf38fd0cce4c0ca5495111963487ec8"} Oct 06 10:08:23 crc kubenswrapper[4824]: I1006 10:08:23.372623 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" event={"ID":"2f511ece-25eb-465f-b85c-9e5f3f886c21","Type":"ContainerDied","Data":"43199f86ad4a1f02bc57a926cd96b2c3c0e2ea46fb36a3f495668706be4c8041"} Oct 06 10:08:23 crc kubenswrapper[4824]: I1006 10:08:23.372636 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" event={"ID":"2f511ece-25eb-465f-b85c-9e5f3f886c21","Type":"ContainerDied","Data":"10bfed3e523b7422ce1e981f98166eca78bd12dd4a7cf0634915e1f4ecc3de2b"} Oct 06 10:08:23 crc kubenswrapper[4824]: I1006 10:08:23.372649 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" event={"ID":"2f511ece-25eb-465f-b85c-9e5f3f886c21","Type":"ContainerDied","Data":"aa6d59fac987f60a4340ab35d84aa059f2b7085a06362e7207a79702d17eee6d"} Oct 06 10:08:23 crc kubenswrapper[4824]: I1006 10:08:23.372662 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" event={"ID":"2f511ece-25eb-465f-b85c-9e5f3f886c21","Type":"ContainerDied","Data":"09638fbc9d163ad0b63fa57d24ea5298207ad1f55bab8671be311960be2627a8"} Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.675708 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-4vwms_2f511ece-25eb-465f-b85c-9e5f3f886c21/ovnkube-controller/3.log" Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.678464 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-4vwms_2f511ece-25eb-465f-b85c-9e5f3f886c21/ovn-acl-logging/0.log" Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.679026 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-4vwms_2f511ece-25eb-465f-b85c-9e5f3f886c21/ovn-controller/0.log" Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.679472 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.741559 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-p4qwn"] Oct 06 10:08:25 crc kubenswrapper[4824]: E1006 10:08:25.741824 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f511ece-25eb-465f-b85c-9e5f3f886c21" containerName="ovn-acl-logging" Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.741840 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f511ece-25eb-465f-b85c-9e5f3f886c21" containerName="ovn-acl-logging" Oct 06 10:08:25 crc kubenswrapper[4824]: E1006 10:08:25.741851 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f511ece-25eb-465f-b85c-9e5f3f886c21" containerName="northd" Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.741858 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f511ece-25eb-465f-b85c-9e5f3f886c21" containerName="northd" Oct 06 10:08:25 crc kubenswrapper[4824]: E1006 10:08:25.741866 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f511ece-25eb-465f-b85c-9e5f3f886c21" containerName="kubecfg-setup" Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.741876 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f511ece-25eb-465f-b85c-9e5f3f886c21" containerName="kubecfg-setup" Oct 06 10:08:25 crc kubenswrapper[4824]: E1006 10:08:25.741912 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f511ece-25eb-465f-b85c-9e5f3f886c21" containerName="ovnkube-controller" Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.741922 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f511ece-25eb-465f-b85c-9e5f3f886c21" containerName="ovnkube-controller" Oct 06 10:08:25 crc kubenswrapper[4824]: E1006 10:08:25.741929 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f511ece-25eb-465f-b85c-9e5f3f886c21" containerName="ovnkube-controller" Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.741937 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f511ece-25eb-465f-b85c-9e5f3f886c21" containerName="ovnkube-controller" Oct 06 10:08:25 crc kubenswrapper[4824]: E1006 10:08:25.741947 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f511ece-25eb-465f-b85c-9e5f3f886c21" containerName="nbdb" Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.741954 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f511ece-25eb-465f-b85c-9e5f3f886c21" containerName="nbdb" Oct 06 10:08:25 crc kubenswrapper[4824]: E1006 10:08:25.741963 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f511ece-25eb-465f-b85c-9e5f3f886c21" containerName="ovn-controller" Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.741970 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f511ece-25eb-465f-b85c-9e5f3f886c21" containerName="ovn-controller" Oct 06 10:08:25 crc kubenswrapper[4824]: E1006 10:08:25.741999 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f511ece-25eb-465f-b85c-9e5f3f886c21" containerName="sbdb" Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.742006 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f511ece-25eb-465f-b85c-9e5f3f886c21" containerName="sbdb" Oct 06 10:08:25 crc kubenswrapper[4824]: E1006 10:08:25.742019 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f511ece-25eb-465f-b85c-9e5f3f886c21" containerName="kube-rbac-proxy-ovn-metrics" Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.742028 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f511ece-25eb-465f-b85c-9e5f3f886c21" containerName="kube-rbac-proxy-ovn-metrics" Oct 06 10:08:25 crc kubenswrapper[4824]: E1006 10:08:25.742040 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f511ece-25eb-465f-b85c-9e5f3f886c21" containerName="kube-rbac-proxy-node" Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.742047 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f511ece-25eb-465f-b85c-9e5f3f886c21" containerName="kube-rbac-proxy-node" Oct 06 10:08:25 crc kubenswrapper[4824]: E1006 10:08:25.742059 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f511ece-25eb-465f-b85c-9e5f3f886c21" containerName="ovnkube-controller" Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.742066 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f511ece-25eb-465f-b85c-9e5f3f886c21" containerName="ovnkube-controller" Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.742175 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f511ece-25eb-465f-b85c-9e5f3f886c21" containerName="ovnkube-controller" Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.742185 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f511ece-25eb-465f-b85c-9e5f3f886c21" containerName="kube-rbac-proxy-node" Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.742191 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f511ece-25eb-465f-b85c-9e5f3f886c21" containerName="ovnkube-controller" Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.742197 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f511ece-25eb-465f-b85c-9e5f3f886c21" containerName="ovnkube-controller" Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.742204 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f511ece-25eb-465f-b85c-9e5f3f886c21" containerName="kube-rbac-proxy-ovn-metrics" Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.742214 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f511ece-25eb-465f-b85c-9e5f3f886c21" containerName="nbdb" Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.742222 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f511ece-25eb-465f-b85c-9e5f3f886c21" containerName="ovn-acl-logging" Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.742234 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f511ece-25eb-465f-b85c-9e5f3f886c21" containerName="ovn-controller" Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.742242 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f511ece-25eb-465f-b85c-9e5f3f886c21" containerName="northd" Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.742250 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f511ece-25eb-465f-b85c-9e5f3f886c21" containerName="sbdb" Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.742257 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f511ece-25eb-465f-b85c-9e5f3f886c21" containerName="ovnkube-controller" Oct 06 10:08:25 crc kubenswrapper[4824]: E1006 10:08:25.742346 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f511ece-25eb-465f-b85c-9e5f3f886c21" containerName="ovnkube-controller" Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.742353 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f511ece-25eb-465f-b85c-9e5f3f886c21" containerName="ovnkube-controller" Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.742443 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f511ece-25eb-465f-b85c-9e5f3f886c21" containerName="ovnkube-controller" Oct 06 10:08:25 crc kubenswrapper[4824]: E1006 10:08:25.742531 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f511ece-25eb-465f-b85c-9e5f3f886c21" containerName="ovnkube-controller" Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.742540 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f511ece-25eb-465f-b85c-9e5f3f886c21" containerName="ovnkube-controller" Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.764829 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-p4qwn" Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.780969 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/3e9fdef6-a3f1-4adc-b541-402f8b2e6fdf-env-overrides\") pod \"ovnkube-node-p4qwn\" (UID: \"3e9fdef6-a3f1-4adc-b541-402f8b2e6fdf\") " pod="openshift-ovn-kubernetes/ovnkube-node-p4qwn" Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.781117 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/3e9fdef6-a3f1-4adc-b541-402f8b2e6fdf-node-log\") pod \"ovnkube-node-p4qwn\" (UID: \"3e9fdef6-a3f1-4adc-b541-402f8b2e6fdf\") " pod="openshift-ovn-kubernetes/ovnkube-node-p4qwn" Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.781153 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/3e9fdef6-a3f1-4adc-b541-402f8b2e6fdf-host-run-netns\") pod \"ovnkube-node-p4qwn\" (UID: \"3e9fdef6-a3f1-4adc-b541-402f8b2e6fdf\") " pod="openshift-ovn-kubernetes/ovnkube-node-p4qwn" Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.781186 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/3e9fdef6-a3f1-4adc-b541-402f8b2e6fdf-host-cni-netd\") pod \"ovnkube-node-p4qwn\" (UID: \"3e9fdef6-a3f1-4adc-b541-402f8b2e6fdf\") " pod="openshift-ovn-kubernetes/ovnkube-node-p4qwn" Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.781233 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kpv2t\" (UniqueName: \"kubernetes.io/projected/3e9fdef6-a3f1-4adc-b541-402f8b2e6fdf-kube-api-access-kpv2t\") pod \"ovnkube-node-p4qwn\" (UID: \"3e9fdef6-a3f1-4adc-b541-402f8b2e6fdf\") " pod="openshift-ovn-kubernetes/ovnkube-node-p4qwn" Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.781273 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3e9fdef6-a3f1-4adc-b541-402f8b2e6fdf-host-run-ovn-kubernetes\") pod \"ovnkube-node-p4qwn\" (UID: \"3e9fdef6-a3f1-4adc-b541-402f8b2e6fdf\") " pod="openshift-ovn-kubernetes/ovnkube-node-p4qwn" Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.781313 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3e9fdef6-a3f1-4adc-b541-402f8b2e6fdf-etc-openvswitch\") pod \"ovnkube-node-p4qwn\" (UID: \"3e9fdef6-a3f1-4adc-b541-402f8b2e6fdf\") " pod="openshift-ovn-kubernetes/ovnkube-node-p4qwn" Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.781435 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/3e9fdef6-a3f1-4adc-b541-402f8b2e6fdf-ovn-node-metrics-cert\") pod \"ovnkube-node-p4qwn\" (UID: \"3e9fdef6-a3f1-4adc-b541-402f8b2e6fdf\") " pod="openshift-ovn-kubernetes/ovnkube-node-p4qwn" Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.781512 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/3e9fdef6-a3f1-4adc-b541-402f8b2e6fdf-ovnkube-script-lib\") pod \"ovnkube-node-p4qwn\" (UID: \"3e9fdef6-a3f1-4adc-b541-402f8b2e6fdf\") " pod="openshift-ovn-kubernetes/ovnkube-node-p4qwn" Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.781549 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/3e9fdef6-a3f1-4adc-b541-402f8b2e6fdf-host-kubelet\") pod \"ovnkube-node-p4qwn\" (UID: \"3e9fdef6-a3f1-4adc-b541-402f8b2e6fdf\") " pod="openshift-ovn-kubernetes/ovnkube-node-p4qwn" Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.781576 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/3e9fdef6-a3f1-4adc-b541-402f8b2e6fdf-host-cni-bin\") pod \"ovnkube-node-p4qwn\" (UID: \"3e9fdef6-a3f1-4adc-b541-402f8b2e6fdf\") " pod="openshift-ovn-kubernetes/ovnkube-node-p4qwn" Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.781606 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/3e9fdef6-a3f1-4adc-b541-402f8b2e6fdf-run-systemd\") pod \"ovnkube-node-p4qwn\" (UID: \"3e9fdef6-a3f1-4adc-b541-402f8b2e6fdf\") " pod="openshift-ovn-kubernetes/ovnkube-node-p4qwn" Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.781640 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3e9fdef6-a3f1-4adc-b541-402f8b2e6fdf-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-p4qwn\" (UID: \"3e9fdef6-a3f1-4adc-b541-402f8b2e6fdf\") " pod="openshift-ovn-kubernetes/ovnkube-node-p4qwn" Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.781685 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/3e9fdef6-a3f1-4adc-b541-402f8b2e6fdf-log-socket\") pod \"ovnkube-node-p4qwn\" (UID: \"3e9fdef6-a3f1-4adc-b541-402f8b2e6fdf\") " pod="openshift-ovn-kubernetes/ovnkube-node-p4qwn" Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.781740 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3e9fdef6-a3f1-4adc-b541-402f8b2e6fdf-run-openvswitch\") pod \"ovnkube-node-p4qwn\" (UID: \"3e9fdef6-a3f1-4adc-b541-402f8b2e6fdf\") " pod="openshift-ovn-kubernetes/ovnkube-node-p4qwn" Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.781770 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3e9fdef6-a3f1-4adc-b541-402f8b2e6fdf-var-lib-openvswitch\") pod \"ovnkube-node-p4qwn\" (UID: \"3e9fdef6-a3f1-4adc-b541-402f8b2e6fdf\") " pod="openshift-ovn-kubernetes/ovnkube-node-p4qwn" Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.782033 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/3e9fdef6-a3f1-4adc-b541-402f8b2e6fdf-ovnkube-config\") pod \"ovnkube-node-p4qwn\" (UID: \"3e9fdef6-a3f1-4adc-b541-402f8b2e6fdf\") " pod="openshift-ovn-kubernetes/ovnkube-node-p4qwn" Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.782075 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/3e9fdef6-a3f1-4adc-b541-402f8b2e6fdf-host-slash\") pod \"ovnkube-node-p4qwn\" (UID: \"3e9fdef6-a3f1-4adc-b541-402f8b2e6fdf\") " pod="openshift-ovn-kubernetes/ovnkube-node-p4qwn" Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.782110 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/3e9fdef6-a3f1-4adc-b541-402f8b2e6fdf-run-ovn\") pod \"ovnkube-node-p4qwn\" (UID: \"3e9fdef6-a3f1-4adc-b541-402f8b2e6fdf\") " pod="openshift-ovn-kubernetes/ovnkube-node-p4qwn" Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.782132 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/3e9fdef6-a3f1-4adc-b541-402f8b2e6fdf-systemd-units\") pod \"ovnkube-node-p4qwn\" (UID: \"3e9fdef6-a3f1-4adc-b541-402f8b2e6fdf\") " pod="openshift-ovn-kubernetes/ovnkube-node-p4qwn" Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.883104 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/2f511ece-25eb-465f-b85c-9e5f3f886c21-log-socket\") pod \"2f511ece-25eb-465f-b85c-9e5f3f886c21\" (UID: \"2f511ece-25eb-465f-b85c-9e5f3f886c21\") " Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.883225 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/2f511ece-25eb-465f-b85c-9e5f3f886c21-env-overrides\") pod \"2f511ece-25eb-465f-b85c-9e5f3f886c21\" (UID: \"2f511ece-25eb-465f-b85c-9e5f3f886c21\") " Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.883248 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2f511ece-25eb-465f-b85c-9e5f3f886c21-run-openvswitch\") pod \"2f511ece-25eb-465f-b85c-9e5f3f886c21\" (UID: \"2f511ece-25eb-465f-b85c-9e5f3f886c21\") " Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.883270 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2f511ece-25eb-465f-b85c-9e5f3f886c21-log-socket" (OuterVolumeSpecName: "log-socket") pod "2f511ece-25eb-465f-b85c-9e5f3f886c21" (UID: "2f511ece-25eb-465f-b85c-9e5f3f886c21"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.883444 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2f511ece-25eb-465f-b85c-9e5f3f886c21-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "2f511ece-25eb-465f-b85c-9e5f3f886c21" (UID: "2f511ece-25eb-465f-b85c-9e5f3f886c21"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.883821 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2f511ece-25eb-465f-b85c-9e5f3f886c21-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "2f511ece-25eb-465f-b85c-9e5f3f886c21" (UID: "2f511ece-25eb-465f-b85c-9e5f3f886c21"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.884182 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/2f511ece-25eb-465f-b85c-9e5f3f886c21-run-systemd\") pod \"2f511ece-25eb-465f-b85c-9e5f3f886c21\" (UID: \"2f511ece-25eb-465f-b85c-9e5f3f886c21\") " Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.884208 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/2f511ece-25eb-465f-b85c-9e5f3f886c21-host-run-netns\") pod \"2f511ece-25eb-465f-b85c-9e5f3f886c21\" (UID: \"2f511ece-25eb-465f-b85c-9e5f3f886c21\") " Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.884231 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/2f511ece-25eb-465f-b85c-9e5f3f886c21-host-slash\") pod \"2f511ece-25eb-465f-b85c-9e5f3f886c21\" (UID: \"2f511ece-25eb-465f-b85c-9e5f3f886c21\") " Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.884259 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mc4k6\" (UniqueName: \"kubernetes.io/projected/2f511ece-25eb-465f-b85c-9e5f3f886c21-kube-api-access-mc4k6\") pod \"2f511ece-25eb-465f-b85c-9e5f3f886c21\" (UID: \"2f511ece-25eb-465f-b85c-9e5f3f886c21\") " Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.884284 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2f511ece-25eb-465f-b85c-9e5f3f886c21-var-lib-openvswitch\") pod \"2f511ece-25eb-465f-b85c-9e5f3f886c21\" (UID: \"2f511ece-25eb-465f-b85c-9e5f3f886c21\") " Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.884308 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/2f511ece-25eb-465f-b85c-9e5f3f886c21-systemd-units\") pod \"2f511ece-25eb-465f-b85c-9e5f3f886c21\" (UID: \"2f511ece-25eb-465f-b85c-9e5f3f886c21\") " Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.884329 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/2f511ece-25eb-465f-b85c-9e5f3f886c21-node-log\") pod \"2f511ece-25eb-465f-b85c-9e5f3f886c21\" (UID: \"2f511ece-25eb-465f-b85c-9e5f3f886c21\") " Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.884329 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2f511ece-25eb-465f-b85c-9e5f3f886c21-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "2f511ece-25eb-465f-b85c-9e5f3f886c21" (UID: "2f511ece-25eb-465f-b85c-9e5f3f886c21"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.884350 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/2f511ece-25eb-465f-b85c-9e5f3f886c21-host-cni-bin\") pod \"2f511ece-25eb-465f-b85c-9e5f3f886c21\" (UID: \"2f511ece-25eb-465f-b85c-9e5f3f886c21\") " Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.884371 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2f511ece-25eb-465f-b85c-9e5f3f886c21-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "2f511ece-25eb-465f-b85c-9e5f3f886c21" (UID: "2f511ece-25eb-465f-b85c-9e5f3f886c21"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.884383 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2f511ece-25eb-465f-b85c-9e5f3f886c21-host-var-lib-cni-networks-ovn-kubernetes\") pod \"2f511ece-25eb-465f-b85c-9e5f3f886c21\" (UID: \"2f511ece-25eb-465f-b85c-9e5f3f886c21\") " Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.884419 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/2f511ece-25eb-465f-b85c-9e5f3f886c21-host-cni-netd\") pod \"2f511ece-25eb-465f-b85c-9e5f3f886c21\" (UID: \"2f511ece-25eb-465f-b85c-9e5f3f886c21\") " Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.884372 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2f511ece-25eb-465f-b85c-9e5f3f886c21-host-slash" (OuterVolumeSpecName: "host-slash") pod "2f511ece-25eb-465f-b85c-9e5f3f886c21" (UID: "2f511ece-25eb-465f-b85c-9e5f3f886c21"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.884427 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2f511ece-25eb-465f-b85c-9e5f3f886c21-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "2f511ece-25eb-465f-b85c-9e5f3f886c21" (UID: "2f511ece-25eb-465f-b85c-9e5f3f886c21"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.884397 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2f511ece-25eb-465f-b85c-9e5f3f886c21-node-log" (OuterVolumeSpecName: "node-log") pod "2f511ece-25eb-465f-b85c-9e5f3f886c21" (UID: "2f511ece-25eb-465f-b85c-9e5f3f886c21"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.884403 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2f511ece-25eb-465f-b85c-9e5f3f886c21-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "2f511ece-25eb-465f-b85c-9e5f3f886c21" (UID: "2f511ece-25eb-465f-b85c-9e5f3f886c21"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.884417 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2f511ece-25eb-465f-b85c-9e5f3f886c21-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "2f511ece-25eb-465f-b85c-9e5f3f886c21" (UID: "2f511ece-25eb-465f-b85c-9e5f3f886c21"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.884454 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2f511ece-25eb-465f-b85c-9e5f3f886c21-etc-openvswitch\") pod \"2f511ece-25eb-465f-b85c-9e5f3f886c21\" (UID: \"2f511ece-25eb-465f-b85c-9e5f3f886c21\") " Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.884468 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2f511ece-25eb-465f-b85c-9e5f3f886c21-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "2f511ece-25eb-465f-b85c-9e5f3f886c21" (UID: "2f511ece-25eb-465f-b85c-9e5f3f886c21"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.884475 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/2f511ece-25eb-465f-b85c-9e5f3f886c21-run-ovn\") pod \"2f511ece-25eb-465f-b85c-9e5f3f886c21\" (UID: \"2f511ece-25eb-465f-b85c-9e5f3f886c21\") " Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.884491 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2f511ece-25eb-465f-b85c-9e5f3f886c21-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "2f511ece-25eb-465f-b85c-9e5f3f886c21" (UID: "2f511ece-25eb-465f-b85c-9e5f3f886c21"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.884505 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2f511ece-25eb-465f-b85c-9e5f3f886c21-host-run-ovn-kubernetes\") pod \"2f511ece-25eb-465f-b85c-9e5f3f886c21\" (UID: \"2f511ece-25eb-465f-b85c-9e5f3f886c21\") " Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.884527 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2f511ece-25eb-465f-b85c-9e5f3f886c21-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "2f511ece-25eb-465f-b85c-9e5f3f886c21" (UID: "2f511ece-25eb-465f-b85c-9e5f3f886c21"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.884537 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/2f511ece-25eb-465f-b85c-9e5f3f886c21-ovnkube-script-lib\") pod \"2f511ece-25eb-465f-b85c-9e5f3f886c21\" (UID: \"2f511ece-25eb-465f-b85c-9e5f3f886c21\") " Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.884516 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2f511ece-25eb-465f-b85c-9e5f3f886c21-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "2f511ece-25eb-465f-b85c-9e5f3f886c21" (UID: "2f511ece-25eb-465f-b85c-9e5f3f886c21"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.884562 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/2f511ece-25eb-465f-b85c-9e5f3f886c21-host-kubelet\") pod \"2f511ece-25eb-465f-b85c-9e5f3f886c21\" (UID: \"2f511ece-25eb-465f-b85c-9e5f3f886c21\") " Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.884593 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/2f511ece-25eb-465f-b85c-9e5f3f886c21-ovn-node-metrics-cert\") pod \"2f511ece-25eb-465f-b85c-9e5f3f886c21\" (UID: \"2f511ece-25eb-465f-b85c-9e5f3f886c21\") " Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.884624 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/2f511ece-25eb-465f-b85c-9e5f3f886c21-ovnkube-config\") pod \"2f511ece-25eb-465f-b85c-9e5f3f886c21\" (UID: \"2f511ece-25eb-465f-b85c-9e5f3f886c21\") " Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.884651 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2f511ece-25eb-465f-b85c-9e5f3f886c21-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "2f511ece-25eb-465f-b85c-9e5f3f886c21" (UID: "2f511ece-25eb-465f-b85c-9e5f3f886c21"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.884745 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kpv2t\" (UniqueName: \"kubernetes.io/projected/3e9fdef6-a3f1-4adc-b541-402f8b2e6fdf-kube-api-access-kpv2t\") pod \"ovnkube-node-p4qwn\" (UID: \"3e9fdef6-a3f1-4adc-b541-402f8b2e6fdf\") " pod="openshift-ovn-kubernetes/ovnkube-node-p4qwn" Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.884776 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3e9fdef6-a3f1-4adc-b541-402f8b2e6fdf-host-run-ovn-kubernetes\") pod \"ovnkube-node-p4qwn\" (UID: \"3e9fdef6-a3f1-4adc-b541-402f8b2e6fdf\") " pod="openshift-ovn-kubernetes/ovnkube-node-p4qwn" Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.884807 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3e9fdef6-a3f1-4adc-b541-402f8b2e6fdf-etc-openvswitch\") pod \"ovnkube-node-p4qwn\" (UID: \"3e9fdef6-a3f1-4adc-b541-402f8b2e6fdf\") " pod="openshift-ovn-kubernetes/ovnkube-node-p4qwn" Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.884939 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/3e9fdef6-a3f1-4adc-b541-402f8b2e6fdf-ovn-node-metrics-cert\") pod \"ovnkube-node-p4qwn\" (UID: \"3e9fdef6-a3f1-4adc-b541-402f8b2e6fdf\") " pod="openshift-ovn-kubernetes/ovnkube-node-p4qwn" Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.884967 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/3e9fdef6-a3f1-4adc-b541-402f8b2e6fdf-ovnkube-script-lib\") pod \"ovnkube-node-p4qwn\" (UID: \"3e9fdef6-a3f1-4adc-b541-402f8b2e6fdf\") " pod="openshift-ovn-kubernetes/ovnkube-node-p4qwn" Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.885010 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/3e9fdef6-a3f1-4adc-b541-402f8b2e6fdf-host-kubelet\") pod \"ovnkube-node-p4qwn\" (UID: \"3e9fdef6-a3f1-4adc-b541-402f8b2e6fdf\") " pod="openshift-ovn-kubernetes/ovnkube-node-p4qwn" Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.885024 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2f511ece-25eb-465f-b85c-9e5f3f886c21-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "2f511ece-25eb-465f-b85c-9e5f3f886c21" (UID: "2f511ece-25eb-465f-b85c-9e5f3f886c21"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.885030 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/3e9fdef6-a3f1-4adc-b541-402f8b2e6fdf-host-cni-bin\") pod \"ovnkube-node-p4qwn\" (UID: \"3e9fdef6-a3f1-4adc-b541-402f8b2e6fdf\") " pod="openshift-ovn-kubernetes/ovnkube-node-p4qwn" Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.885085 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/3e9fdef6-a3f1-4adc-b541-402f8b2e6fdf-run-systemd\") pod \"ovnkube-node-p4qwn\" (UID: \"3e9fdef6-a3f1-4adc-b541-402f8b2e6fdf\") " pod="openshift-ovn-kubernetes/ovnkube-node-p4qwn" Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.885105 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3e9fdef6-a3f1-4adc-b541-402f8b2e6fdf-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-p4qwn\" (UID: \"3e9fdef6-a3f1-4adc-b541-402f8b2e6fdf\") " pod="openshift-ovn-kubernetes/ovnkube-node-p4qwn" Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.885111 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3e9fdef6-a3f1-4adc-b541-402f8b2e6fdf-etc-openvswitch\") pod \"ovnkube-node-p4qwn\" (UID: \"3e9fdef6-a3f1-4adc-b541-402f8b2e6fdf\") " pod="openshift-ovn-kubernetes/ovnkube-node-p4qwn" Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.885182 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/3e9fdef6-a3f1-4adc-b541-402f8b2e6fdf-log-socket\") pod \"ovnkube-node-p4qwn\" (UID: \"3e9fdef6-a3f1-4adc-b541-402f8b2e6fdf\") " pod="openshift-ovn-kubernetes/ovnkube-node-p4qwn" Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.885473 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2f511ece-25eb-465f-b85c-9e5f3f886c21-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "2f511ece-25eb-465f-b85c-9e5f3f886c21" (UID: "2f511ece-25eb-465f-b85c-9e5f3f886c21"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.885455 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/3e9fdef6-a3f1-4adc-b541-402f8b2e6fdf-host-cni-bin\") pod \"ovnkube-node-p4qwn\" (UID: \"3e9fdef6-a3f1-4adc-b541-402f8b2e6fdf\") " pod="openshift-ovn-kubernetes/ovnkube-node-p4qwn" Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.885130 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/3e9fdef6-a3f1-4adc-b541-402f8b2e6fdf-log-socket\") pod \"ovnkube-node-p4qwn\" (UID: \"3e9fdef6-a3f1-4adc-b541-402f8b2e6fdf\") " pod="openshift-ovn-kubernetes/ovnkube-node-p4qwn" Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.885844 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3e9fdef6-a3f1-4adc-b541-402f8b2e6fdf-run-openvswitch\") pod \"ovnkube-node-p4qwn\" (UID: \"3e9fdef6-a3f1-4adc-b541-402f8b2e6fdf\") " pod="openshift-ovn-kubernetes/ovnkube-node-p4qwn" Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.885920 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3e9fdef6-a3f1-4adc-b541-402f8b2e6fdf-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-p4qwn\" (UID: \"3e9fdef6-a3f1-4adc-b541-402f8b2e6fdf\") " pod="openshift-ovn-kubernetes/ovnkube-node-p4qwn" Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.885893 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3e9fdef6-a3f1-4adc-b541-402f8b2e6fdf-host-run-ovn-kubernetes\") pod \"ovnkube-node-p4qwn\" (UID: \"3e9fdef6-a3f1-4adc-b541-402f8b2e6fdf\") " pod="openshift-ovn-kubernetes/ovnkube-node-p4qwn" Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.885925 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3e9fdef6-a3f1-4adc-b541-402f8b2e6fdf-run-openvswitch\") pod \"ovnkube-node-p4qwn\" (UID: \"3e9fdef6-a3f1-4adc-b541-402f8b2e6fdf\") " pod="openshift-ovn-kubernetes/ovnkube-node-p4qwn" Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.885887 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/3e9fdef6-a3f1-4adc-b541-402f8b2e6fdf-host-kubelet\") pod \"ovnkube-node-p4qwn\" (UID: \"3e9fdef6-a3f1-4adc-b541-402f8b2e6fdf\") " pod="openshift-ovn-kubernetes/ovnkube-node-p4qwn" Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.886006 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/3e9fdef6-a3f1-4adc-b541-402f8b2e6fdf-run-systemd\") pod \"ovnkube-node-p4qwn\" (UID: \"3e9fdef6-a3f1-4adc-b541-402f8b2e6fdf\") " pod="openshift-ovn-kubernetes/ovnkube-node-p4qwn" Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.886056 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3e9fdef6-a3f1-4adc-b541-402f8b2e6fdf-var-lib-openvswitch\") pod \"ovnkube-node-p4qwn\" (UID: \"3e9fdef6-a3f1-4adc-b541-402f8b2e6fdf\") " pod="openshift-ovn-kubernetes/ovnkube-node-p4qwn" Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.885934 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3e9fdef6-a3f1-4adc-b541-402f8b2e6fdf-var-lib-openvswitch\") pod \"ovnkube-node-p4qwn\" (UID: \"3e9fdef6-a3f1-4adc-b541-402f8b2e6fdf\") " pod="openshift-ovn-kubernetes/ovnkube-node-p4qwn" Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.886313 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/3e9fdef6-a3f1-4adc-b541-402f8b2e6fdf-ovnkube-config\") pod \"ovnkube-node-p4qwn\" (UID: \"3e9fdef6-a3f1-4adc-b541-402f8b2e6fdf\") " pod="openshift-ovn-kubernetes/ovnkube-node-p4qwn" Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.886392 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/3e9fdef6-a3f1-4adc-b541-402f8b2e6fdf-host-slash\") pod \"ovnkube-node-p4qwn\" (UID: \"3e9fdef6-a3f1-4adc-b541-402f8b2e6fdf\") " pod="openshift-ovn-kubernetes/ovnkube-node-p4qwn" Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.886482 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/3e9fdef6-a3f1-4adc-b541-402f8b2e6fdf-run-ovn\") pod \"ovnkube-node-p4qwn\" (UID: \"3e9fdef6-a3f1-4adc-b541-402f8b2e6fdf\") " pod="openshift-ovn-kubernetes/ovnkube-node-p4qwn" Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.886554 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/3e9fdef6-a3f1-4adc-b541-402f8b2e6fdf-systemd-units\") pod \"ovnkube-node-p4qwn\" (UID: \"3e9fdef6-a3f1-4adc-b541-402f8b2e6fdf\") " pod="openshift-ovn-kubernetes/ovnkube-node-p4qwn" Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.886649 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/3e9fdef6-a3f1-4adc-b541-402f8b2e6fdf-env-overrides\") pod \"ovnkube-node-p4qwn\" (UID: \"3e9fdef6-a3f1-4adc-b541-402f8b2e6fdf\") " pod="openshift-ovn-kubernetes/ovnkube-node-p4qwn" Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.886747 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/3e9fdef6-a3f1-4adc-b541-402f8b2e6fdf-node-log\") pod \"ovnkube-node-p4qwn\" (UID: \"3e9fdef6-a3f1-4adc-b541-402f8b2e6fdf\") " pod="openshift-ovn-kubernetes/ovnkube-node-p4qwn" Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.886820 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/3e9fdef6-a3f1-4adc-b541-402f8b2e6fdf-host-run-netns\") pod \"ovnkube-node-p4qwn\" (UID: \"3e9fdef6-a3f1-4adc-b541-402f8b2e6fdf\") " pod="openshift-ovn-kubernetes/ovnkube-node-p4qwn" Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.886914 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/3e9fdef6-a3f1-4adc-b541-402f8b2e6fdf-host-cni-netd\") pod \"ovnkube-node-p4qwn\" (UID: \"3e9fdef6-a3f1-4adc-b541-402f8b2e6fdf\") " pod="openshift-ovn-kubernetes/ovnkube-node-p4qwn" Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.887001 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/3e9fdef6-a3f1-4adc-b541-402f8b2e6fdf-systemd-units\") pod \"ovnkube-node-p4qwn\" (UID: \"3e9fdef6-a3f1-4adc-b541-402f8b2e6fdf\") " pod="openshift-ovn-kubernetes/ovnkube-node-p4qwn" Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.886490 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/3e9fdef6-a3f1-4adc-b541-402f8b2e6fdf-ovnkube-script-lib\") pod \"ovnkube-node-p4qwn\" (UID: \"3e9fdef6-a3f1-4adc-b541-402f8b2e6fdf\") " pod="openshift-ovn-kubernetes/ovnkube-node-p4qwn" Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.886953 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/3e9fdef6-a3f1-4adc-b541-402f8b2e6fdf-ovnkube-config\") pod \"ovnkube-node-p4qwn\" (UID: \"3e9fdef6-a3f1-4adc-b541-402f8b2e6fdf\") " pod="openshift-ovn-kubernetes/ovnkube-node-p4qwn" Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.887264 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/3e9fdef6-a3f1-4adc-b541-402f8b2e6fdf-host-run-netns\") pod \"ovnkube-node-p4qwn\" (UID: \"3e9fdef6-a3f1-4adc-b541-402f8b2e6fdf\") " pod="openshift-ovn-kubernetes/ovnkube-node-p4qwn" Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.887345 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/3e9fdef6-a3f1-4adc-b541-402f8b2e6fdf-host-cni-netd\") pod \"ovnkube-node-p4qwn\" (UID: \"3e9fdef6-a3f1-4adc-b541-402f8b2e6fdf\") " pod="openshift-ovn-kubernetes/ovnkube-node-p4qwn" Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.887523 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/3e9fdef6-a3f1-4adc-b541-402f8b2e6fdf-run-ovn\") pod \"ovnkube-node-p4qwn\" (UID: \"3e9fdef6-a3f1-4adc-b541-402f8b2e6fdf\") " pod="openshift-ovn-kubernetes/ovnkube-node-p4qwn" Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.887542 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/3e9fdef6-a3f1-4adc-b541-402f8b2e6fdf-env-overrides\") pod \"ovnkube-node-p4qwn\" (UID: \"3e9fdef6-a3f1-4adc-b541-402f8b2e6fdf\") " pod="openshift-ovn-kubernetes/ovnkube-node-p4qwn" Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.887563 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/3e9fdef6-a3f1-4adc-b541-402f8b2e6fdf-node-log\") pod \"ovnkube-node-p4qwn\" (UID: \"3e9fdef6-a3f1-4adc-b541-402f8b2e6fdf\") " pod="openshift-ovn-kubernetes/ovnkube-node-p4qwn" Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.887578 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/3e9fdef6-a3f1-4adc-b541-402f8b2e6fdf-host-slash\") pod \"ovnkube-node-p4qwn\" (UID: \"3e9fdef6-a3f1-4adc-b541-402f8b2e6fdf\") " pod="openshift-ovn-kubernetes/ovnkube-node-p4qwn" Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.887599 4824 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/2f511ece-25eb-465f-b85c-9e5f3f886c21-host-cni-netd\") on node \"crc\" DevicePath \"\"" Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.887939 4824 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2f511ece-25eb-465f-b85c-9e5f3f886c21-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.888213 4824 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/2f511ece-25eb-465f-b85c-9e5f3f886c21-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.888277 4824 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2f511ece-25eb-465f-b85c-9e5f3f886c21-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.888344 4824 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/2f511ece-25eb-465f-b85c-9e5f3f886c21-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.888405 4824 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/2f511ece-25eb-465f-b85c-9e5f3f886c21-host-kubelet\") on node \"crc\" DevicePath \"\"" Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.888468 4824 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/2f511ece-25eb-465f-b85c-9e5f3f886c21-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.888526 4824 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/2f511ece-25eb-465f-b85c-9e5f3f886c21-log-socket\") on node \"crc\" DevicePath \"\"" Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.888591 4824 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/2f511ece-25eb-465f-b85c-9e5f3f886c21-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.888654 4824 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2f511ece-25eb-465f-b85c-9e5f3f886c21-run-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.888717 4824 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/2f511ece-25eb-465f-b85c-9e5f3f886c21-host-run-netns\") on node \"crc\" DevicePath \"\"" Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.891276 4824 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/2f511ece-25eb-465f-b85c-9e5f3f886c21-host-slash\") on node \"crc\" DevicePath \"\"" Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.891628 4824 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2f511ece-25eb-465f-b85c-9e5f3f886c21-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.891714 4824 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/2f511ece-25eb-465f-b85c-9e5f3f886c21-systemd-units\") on node \"crc\" DevicePath \"\"" Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.891801 4824 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/2f511ece-25eb-465f-b85c-9e5f3f886c21-node-log\") on node \"crc\" DevicePath \"\"" Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.891871 4824 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/2f511ece-25eb-465f-b85c-9e5f3f886c21-host-cni-bin\") on node \"crc\" DevicePath \"\"" Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.891939 4824 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2f511ece-25eb-465f-b85c-9e5f3f886c21-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.898841 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/3e9fdef6-a3f1-4adc-b541-402f8b2e6fdf-ovn-node-metrics-cert\") pod \"ovnkube-node-p4qwn\" (UID: \"3e9fdef6-a3f1-4adc-b541-402f8b2e6fdf\") " pod="openshift-ovn-kubernetes/ovnkube-node-p4qwn" Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.902305 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f511ece-25eb-465f-b85c-9e5f3f886c21-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "2f511ece-25eb-465f-b85c-9e5f3f886c21" (UID: "2f511ece-25eb-465f-b85c-9e5f3f886c21"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.903506 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2f511ece-25eb-465f-b85c-9e5f3f886c21-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "2f511ece-25eb-465f-b85c-9e5f3f886c21" (UID: "2f511ece-25eb-465f-b85c-9e5f3f886c21"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.905061 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2f511ece-25eb-465f-b85c-9e5f3f886c21-kube-api-access-mc4k6" (OuterVolumeSpecName: "kube-api-access-mc4k6") pod "2f511ece-25eb-465f-b85c-9e5f3f886c21" (UID: "2f511ece-25eb-465f-b85c-9e5f3f886c21"). InnerVolumeSpecName "kube-api-access-mc4k6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.912490 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kpv2t\" (UniqueName: \"kubernetes.io/projected/3e9fdef6-a3f1-4adc-b541-402f8b2e6fdf-kube-api-access-kpv2t\") pod \"ovnkube-node-p4qwn\" (UID: \"3e9fdef6-a3f1-4adc-b541-402f8b2e6fdf\") " pod="openshift-ovn-kubernetes/ovnkube-node-p4qwn" Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.993441 4824 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/2f511ece-25eb-465f-b85c-9e5f3f886c21-run-systemd\") on node \"crc\" DevicePath \"\"" Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.993503 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mc4k6\" (UniqueName: \"kubernetes.io/projected/2f511ece-25eb-465f-b85c-9e5f3f886c21-kube-api-access-mc4k6\") on node \"crc\" DevicePath \"\"" Oct 06 10:08:25 crc kubenswrapper[4824]: I1006 10:08:25.993528 4824 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/2f511ece-25eb-465f-b85c-9e5f3f886c21-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 06 10:08:26 crc kubenswrapper[4824]: I1006 10:08:26.057736 4824 scope.go:117] "RemoveContainer" containerID="9c88940bd9c33c0ee0feef797ad549f874cad42942352072dba2b4eaf1e8b803" Oct 06 10:08:26 crc kubenswrapper[4824]: I1006 10:08:26.090103 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-p4qwn" Oct 06 10:08:26 crc kubenswrapper[4824]: I1006 10:08:26.396039 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-bdn2f" event={"ID":"bde1f651-2eda-4d2d-8ab5-5fcdcd4580a4","Type":"ContainerStarted","Data":"23bd7078572657e46c156b6d3415d9ee8dba6f8ade13ba1ab7f6a8198212069d"} Oct 06 10:08:26 crc kubenswrapper[4824]: I1006 10:08:26.397695 4824 generic.go:334] "Generic (PLEG): container finished" podID="3e9fdef6-a3f1-4adc-b541-402f8b2e6fdf" containerID="9409f0b3ecbfc1725d79a58e70226f332c62b58194c35e300e58e6f8023c09a3" exitCode=0 Oct 06 10:08:26 crc kubenswrapper[4824]: I1006 10:08:26.397733 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p4qwn" event={"ID":"3e9fdef6-a3f1-4adc-b541-402f8b2e6fdf","Type":"ContainerDied","Data":"9409f0b3ecbfc1725d79a58e70226f332c62b58194c35e300e58e6f8023c09a3"} Oct 06 10:08:26 crc kubenswrapper[4824]: I1006 10:08:26.397781 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p4qwn" event={"ID":"3e9fdef6-a3f1-4adc-b541-402f8b2e6fdf","Type":"ContainerStarted","Data":"f1133d29aecc44b7ebdfe23983b94aaa4aff9a46b79e39f1016614e4de3971ef"} Oct 06 10:08:26 crc kubenswrapper[4824]: I1006 10:08:26.399632 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-szn8q_b65abf9f-4e88-4571-960d-3ca997d9c344/kube-multus/2.log" Oct 06 10:08:26 crc kubenswrapper[4824]: I1006 10:08:26.406458 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-4vwms_2f511ece-25eb-465f-b85c-9e5f3f886c21/ovn-acl-logging/0.log" Oct 06 10:08:26 crc kubenswrapper[4824]: I1006 10:08:26.407189 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-4vwms_2f511ece-25eb-465f-b85c-9e5f3f886c21/ovn-controller/0.log" Oct 06 10:08:26 crc kubenswrapper[4824]: I1006 10:08:26.408108 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" event={"ID":"2f511ece-25eb-465f-b85c-9e5f3f886c21","Type":"ContainerDied","Data":"b7982e1a18ea2e6f50d59ed4fcc3bb8dd2242517d06b86feab7d9b463fca89cb"} Oct 06 10:08:26 crc kubenswrapper[4824]: I1006 10:08:26.408168 4824 scope.go:117] "RemoveContainer" containerID="3dffdccf2aac654a5a0f629b8c5b2233c4561cb036262068f2289f6e95b46575" Oct 06 10:08:26 crc kubenswrapper[4824]: I1006 10:08:26.408416 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-4vwms" Oct 06 10:08:26 crc kubenswrapper[4824]: I1006 10:08:26.410456 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-htx98" event={"ID":"d6f420d5-ed6d-4201-b78f-0bfa304e3b6b","Type":"ContainerStarted","Data":"f55c874163ac2606f070cef5d3d4df340b0bdbf51f8388c069cfd558a4c1d3a5"} Oct 06 10:08:26 crc kubenswrapper[4824]: I1006 10:08:26.410699 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-5655c58dd6-htx98" Oct 06 10:08:26 crc kubenswrapper[4824]: I1006 10:08:26.441675 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-5b446d88c5-bdn2f" podStartSLOduration=1.282215353 podStartE2EDuration="13.441649227s" podCreationTimestamp="2025-10-06 10:08:13 +0000 UTC" firstStartedPulling="2025-10-06 10:08:13.902146365 +0000 UTC m=+603.266569226" lastFinishedPulling="2025-10-06 10:08:26.061580199 +0000 UTC m=+615.426003100" observedRunningTime="2025-10-06 10:08:26.439503923 +0000 UTC m=+615.803926824" watchObservedRunningTime="2025-10-06 10:08:26.441649227 +0000 UTC m=+615.806072098" Oct 06 10:08:26 crc kubenswrapper[4824]: I1006 10:08:26.500041 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-5655c58dd6-htx98" podStartSLOduration=1.481800272 podStartE2EDuration="13.499971758s" podCreationTimestamp="2025-10-06 10:08:13 +0000 UTC" firstStartedPulling="2025-10-06 10:08:14.042936331 +0000 UTC m=+603.407359232" lastFinishedPulling="2025-10-06 10:08:26.061107857 +0000 UTC m=+615.425530718" observedRunningTime="2025-10-06 10:08:26.461879913 +0000 UTC m=+615.826302784" watchObservedRunningTime="2025-10-06 10:08:26.499971758 +0000 UTC m=+615.864394619" Oct 06 10:08:26 crc kubenswrapper[4824]: I1006 10:08:26.551760 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-4vwms"] Oct 06 10:08:26 crc kubenswrapper[4824]: I1006 10:08:26.565604 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-4vwms"] Oct 06 10:08:26 crc kubenswrapper[4824]: I1006 10:08:26.818021 4824 scope.go:117] "RemoveContainer" containerID="0644c7db28c7d26aa8d69cd5b9fbcff4d4c06b96df113137dea22a72d9fecd85" Oct 06 10:08:27 crc kubenswrapper[4824]: I1006 10:08:27.058896 4824 scope.go:117] "RemoveContainer" containerID="9d6064ad3bcdafebb4c66e7d91f183cc787e9ce13660545c71a722b3ff04ce75" Oct 06 10:08:27 crc kubenswrapper[4824]: I1006 10:08:27.095163 4824 scope.go:117] "RemoveContainer" containerID="6213164d366243c2139d5971041d6f076cf38fd0cce4c0ca5495111963487ec8" Oct 06 10:08:27 crc kubenswrapper[4824]: I1006 10:08:27.143243 4824 scope.go:117] "RemoveContainer" containerID="43199f86ad4a1f02bc57a926cd96b2c3c0e2ea46fb36a3f495668706be4c8041" Oct 06 10:08:27 crc kubenswrapper[4824]: I1006 10:08:27.200294 4824 scope.go:117] "RemoveContainer" containerID="10bfed3e523b7422ce1e981f98166eca78bd12dd4a7cf0634915e1f4ecc3de2b" Oct 06 10:08:27 crc kubenswrapper[4824]: I1006 10:08:27.231117 4824 scope.go:117] "RemoveContainer" containerID="aa6d59fac987f60a4340ab35d84aa059f2b7085a06362e7207a79702d17eee6d" Oct 06 10:08:27 crc kubenswrapper[4824]: I1006 10:08:27.260506 4824 scope.go:117] "RemoveContainer" containerID="09638fbc9d163ad0b63fa57d24ea5298207ad1f55bab8671be311960be2627a8" Oct 06 10:08:27 crc kubenswrapper[4824]: I1006 10:08:27.283311 4824 scope.go:117] "RemoveContainer" containerID="6da4c07be4d507a6f20bd8b42be6e6ea4712a956f57ef164036cfded7142a87e" Oct 06 10:08:27 crc kubenswrapper[4824]: I1006 10:08:27.288046 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2f511ece-25eb-465f-b85c-9e5f3f886c21" path="/var/lib/kubelet/pods/2f511ece-25eb-465f-b85c-9e5f3f886c21/volumes" Oct 06 10:08:27 crc kubenswrapper[4824]: I1006 10:08:27.422404 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p4qwn" event={"ID":"3e9fdef6-a3f1-4adc-b541-402f8b2e6fdf","Type":"ContainerStarted","Data":"ef4c1c7c2dd2d0e47333b922bae1272fcc1efd3b40819c9c7e2cda27e7a334d2"} Oct 06 10:08:27 crc kubenswrapper[4824]: I1006 10:08:27.426597 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-bbqpq" event={"ID":"b5759913-9cb4-4012-bffb-fe2f8085543c","Type":"ContainerStarted","Data":"689fd60105cbba872adeb8730731c9f81c5608caf6cb32cd7bd64816f33dbf74"} Oct 06 10:08:27 crc kubenswrapper[4824]: I1006 10:08:27.448914 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7f985d654d-bbqpq" podStartSLOduration=1.43980428 podStartE2EDuration="14.44889007s" podCreationTimestamp="2025-10-06 10:08:13 +0000 UTC" firstStartedPulling="2025-10-06 10:08:14.053514446 +0000 UTC m=+603.417937307" lastFinishedPulling="2025-10-06 10:08:27.062600196 +0000 UTC m=+616.427023097" observedRunningTime="2025-10-06 10:08:27.445586457 +0000 UTC m=+616.810009328" watchObservedRunningTime="2025-10-06 10:08:27.44889007 +0000 UTC m=+616.813312931" Oct 06 10:08:28 crc kubenswrapper[4824]: I1006 10:08:28.443956 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p4qwn" event={"ID":"3e9fdef6-a3f1-4adc-b541-402f8b2e6fdf","Type":"ContainerStarted","Data":"77203333da1945ece80d675e9f6d7448d726e6321b847a3e2f9ae1ced28354a8"} Oct 06 10:08:28 crc kubenswrapper[4824]: I1006 10:08:28.446118 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p4qwn" event={"ID":"3e9fdef6-a3f1-4adc-b541-402f8b2e6fdf","Type":"ContainerStarted","Data":"c29d72e768bce6b4da4d7ef43bbe80b1e15edabf647c485341e21966e16e8a84"} Oct 06 10:08:29 crc kubenswrapper[4824]: I1006 10:08:29.463191 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p4qwn" event={"ID":"3e9fdef6-a3f1-4adc-b541-402f8b2e6fdf","Type":"ContainerStarted","Data":"3f8382dc1f64c0b459cd3cae2c792b6adaecea6397b246876415a00f24a26918"} Oct 06 10:08:29 crc kubenswrapper[4824]: I1006 10:08:29.463836 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p4qwn" event={"ID":"3e9fdef6-a3f1-4adc-b541-402f8b2e6fdf","Type":"ContainerStarted","Data":"ea927fed89266a6c851818c439707bf3605a7187b49be8a9de0693d31054758b"} Oct 06 10:08:29 crc kubenswrapper[4824]: I1006 10:08:29.463874 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p4qwn" event={"ID":"3e9fdef6-a3f1-4adc-b541-402f8b2e6fdf","Type":"ContainerStarted","Data":"3228d0efda9813159dfc414acde980082b3db013a3d50cdc4612938119c340aa"} Oct 06 10:08:32 crc kubenswrapper[4824]: I1006 10:08:32.487560 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p4qwn" event={"ID":"3e9fdef6-a3f1-4adc-b541-402f8b2e6fdf","Type":"ContainerStarted","Data":"6b933c5c4f3c0911ad80d595f55ee3828f4149104d49f0037fdf59ae7e069c47"} Oct 06 10:08:33 crc kubenswrapper[4824]: I1006 10:08:33.574203 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-5655c58dd6-htx98" Oct 06 10:08:34 crc kubenswrapper[4824]: I1006 10:08:34.505665 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p4qwn" event={"ID":"3e9fdef6-a3f1-4adc-b541-402f8b2e6fdf","Type":"ContainerStarted","Data":"b360c72446472b3dae47dd524e8c8a8307d643be9cbb4d6651eef7b8ce51cb64"} Oct 06 10:08:34 crc kubenswrapper[4824]: I1006 10:08:34.506286 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-p4qwn" Oct 06 10:08:34 crc kubenswrapper[4824]: I1006 10:08:34.506313 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-p4qwn" Oct 06 10:08:34 crc kubenswrapper[4824]: I1006 10:08:34.506329 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-p4qwn" Oct 06 10:08:34 crc kubenswrapper[4824]: I1006 10:08:34.536843 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-p4qwn" Oct 06 10:08:34 crc kubenswrapper[4824]: I1006 10:08:34.540428 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-p4qwn" Oct 06 10:08:34 crc kubenswrapper[4824]: I1006 10:08:34.548593 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-p4qwn" podStartSLOduration=9.548566428 podStartE2EDuration="9.548566428s" podCreationTimestamp="2025-10-06 10:08:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:08:34.544117277 +0000 UTC m=+623.908540158" watchObservedRunningTime="2025-10-06 10:08:34.548566428 +0000 UTC m=+623.912989299" Oct 06 10:08:36 crc kubenswrapper[4824]: I1006 10:08:36.274531 4824 scope.go:117] "RemoveContainer" containerID="238ede7256cac751995c87da2f05e25727cc9496ddf838e9d57f365f50acf4ba" Oct 06 10:08:36 crc kubenswrapper[4824]: E1006 10:08:36.275328 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-szn8q_openshift-multus(b65abf9f-4e88-4571-960d-3ca997d9c344)\"" pod="openshift-multus/multus-szn8q" podUID="b65abf9f-4e88-4571-960d-3ca997d9c344" Oct 06 10:08:43 crc kubenswrapper[4824]: I1006 10:08:43.915799 4824 patch_prober.go:28] interesting pod/machine-config-daemon-khgzw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 10:08:43 crc kubenswrapper[4824]: I1006 10:08:43.916655 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 10:08:43 crc kubenswrapper[4824]: I1006 10:08:43.916739 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" Oct 06 10:08:43 crc kubenswrapper[4824]: I1006 10:08:43.917917 4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d6acecda0f23f881c285040e9b7cdcc459c602934016ff58bff3f9c6edb5dbaf"} pod="openshift-machine-config-operator/machine-config-daemon-khgzw" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 06 10:08:43 crc kubenswrapper[4824]: I1006 10:08:43.918081 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" containerName="machine-config-daemon" containerID="cri-o://d6acecda0f23f881c285040e9b7cdcc459c602934016ff58bff3f9c6edb5dbaf" gracePeriod=600 Oct 06 10:08:44 crc kubenswrapper[4824]: I1006 10:08:44.580108 4824 generic.go:334] "Generic (PLEG): container finished" podID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" containerID="d6acecda0f23f881c285040e9b7cdcc459c602934016ff58bff3f9c6edb5dbaf" exitCode=0 Oct 06 10:08:44 crc kubenswrapper[4824]: I1006 10:08:44.580191 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" event={"ID":"e1140c8c-93f1-4fce-8c06-a69561ff8a6d","Type":"ContainerDied","Data":"d6acecda0f23f881c285040e9b7cdcc459c602934016ff58bff3f9c6edb5dbaf"} Oct 06 10:08:44 crc kubenswrapper[4824]: I1006 10:08:44.580617 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" event={"ID":"e1140c8c-93f1-4fce-8c06-a69561ff8a6d","Type":"ContainerStarted","Data":"8fa1ede4cec498bdfe66760d5e74fd58534149105068e959659c8b228527cc04"} Oct 06 10:08:44 crc kubenswrapper[4824]: I1006 10:08:44.580648 4824 scope.go:117] "RemoveContainer" containerID="849cbe616fe0e1fc3c60c65e540b59cf8cac7989da24b5c2d88ead6d16645186" Oct 06 10:08:47 crc kubenswrapper[4824]: I1006 10:08:47.274565 4824 scope.go:117] "RemoveContainer" containerID="238ede7256cac751995c87da2f05e25727cc9496ddf838e9d57f365f50acf4ba" Oct 06 10:08:47 crc kubenswrapper[4824]: I1006 10:08:47.605863 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-szn8q_b65abf9f-4e88-4571-960d-3ca997d9c344/kube-multus/2.log" Oct 06 10:08:47 crc kubenswrapper[4824]: I1006 10:08:47.605972 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-szn8q" event={"ID":"b65abf9f-4e88-4571-960d-3ca997d9c344","Type":"ContainerStarted","Data":"01c8e1910ed02fcf15bb0321cfe80a294821a10421285b5d2d72ab7e082cb4ca"} Oct 06 10:08:56 crc kubenswrapper[4824]: I1006 10:08:56.128308 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-p4qwn" Oct 06 10:09:16 crc kubenswrapper[4824]: I1006 10:09:16.359952 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cxdt8w"] Oct 06 10:09:16 crc kubenswrapper[4824]: I1006 10:09:16.363257 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cxdt8w" Oct 06 10:09:16 crc kubenswrapper[4824]: I1006 10:09:16.367157 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 06 10:09:16 crc kubenswrapper[4824]: I1006 10:09:16.374084 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cxdt8w"] Oct 06 10:09:16 crc kubenswrapper[4824]: I1006 10:09:16.516701 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5e47aa0e-c660-4f88-8fb1-8c873268879d-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cxdt8w\" (UID: \"5e47aa0e-c660-4f88-8fb1-8c873268879d\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cxdt8w" Oct 06 10:09:16 crc kubenswrapper[4824]: I1006 10:09:16.517161 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c7lkf\" (UniqueName: \"kubernetes.io/projected/5e47aa0e-c660-4f88-8fb1-8c873268879d-kube-api-access-c7lkf\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cxdt8w\" (UID: \"5e47aa0e-c660-4f88-8fb1-8c873268879d\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cxdt8w" Oct 06 10:09:16 crc kubenswrapper[4824]: I1006 10:09:16.517310 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5e47aa0e-c660-4f88-8fb1-8c873268879d-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cxdt8w\" (UID: \"5e47aa0e-c660-4f88-8fb1-8c873268879d\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cxdt8w" Oct 06 10:09:16 crc kubenswrapper[4824]: I1006 10:09:16.618760 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c7lkf\" (UniqueName: \"kubernetes.io/projected/5e47aa0e-c660-4f88-8fb1-8c873268879d-kube-api-access-c7lkf\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cxdt8w\" (UID: \"5e47aa0e-c660-4f88-8fb1-8c873268879d\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cxdt8w" Oct 06 10:09:16 crc kubenswrapper[4824]: I1006 10:09:16.619192 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5e47aa0e-c660-4f88-8fb1-8c873268879d-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cxdt8w\" (UID: \"5e47aa0e-c660-4f88-8fb1-8c873268879d\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cxdt8w" Oct 06 10:09:16 crc kubenswrapper[4824]: I1006 10:09:16.619341 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5e47aa0e-c660-4f88-8fb1-8c873268879d-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cxdt8w\" (UID: \"5e47aa0e-c660-4f88-8fb1-8c873268879d\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cxdt8w" Oct 06 10:09:16 crc kubenswrapper[4824]: I1006 10:09:16.619797 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5e47aa0e-c660-4f88-8fb1-8c873268879d-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cxdt8w\" (UID: \"5e47aa0e-c660-4f88-8fb1-8c873268879d\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cxdt8w" Oct 06 10:09:16 crc kubenswrapper[4824]: I1006 10:09:16.620015 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5e47aa0e-c660-4f88-8fb1-8c873268879d-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cxdt8w\" (UID: \"5e47aa0e-c660-4f88-8fb1-8c873268879d\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cxdt8w" Oct 06 10:09:16 crc kubenswrapper[4824]: I1006 10:09:16.650693 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c7lkf\" (UniqueName: \"kubernetes.io/projected/5e47aa0e-c660-4f88-8fb1-8c873268879d-kube-api-access-c7lkf\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cxdt8w\" (UID: \"5e47aa0e-c660-4f88-8fb1-8c873268879d\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cxdt8w" Oct 06 10:09:16 crc kubenswrapper[4824]: I1006 10:09:16.701603 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cxdt8w" Oct 06 10:09:17 crc kubenswrapper[4824]: I1006 10:09:17.021101 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cxdt8w"] Oct 06 10:09:17 crc kubenswrapper[4824]: I1006 10:09:17.872475 4824 generic.go:334] "Generic (PLEG): container finished" podID="5e47aa0e-c660-4f88-8fb1-8c873268879d" containerID="50718ce819172a32b427591995268885b7aa2232f3280242590c3ae562bd5052" exitCode=0 Oct 06 10:09:17 crc kubenswrapper[4824]: I1006 10:09:17.872680 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cxdt8w" event={"ID":"5e47aa0e-c660-4f88-8fb1-8c873268879d","Type":"ContainerDied","Data":"50718ce819172a32b427591995268885b7aa2232f3280242590c3ae562bd5052"} Oct 06 10:09:17 crc kubenswrapper[4824]: I1006 10:09:17.874257 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cxdt8w" event={"ID":"5e47aa0e-c660-4f88-8fb1-8c873268879d","Type":"ContainerStarted","Data":"6f0b93f016bbda1dc66944473f518c695267b7ef2c4049c1b38e14acb304a14c"} Oct 06 10:09:19 crc kubenswrapper[4824]: I1006 10:09:19.892290 4824 generic.go:334] "Generic (PLEG): container finished" podID="5e47aa0e-c660-4f88-8fb1-8c873268879d" containerID="33f1b58ffd98d9f1c82d12554641ddfde0213d47e739ff2ac668b7e03386c168" exitCode=0 Oct 06 10:09:19 crc kubenswrapper[4824]: I1006 10:09:19.892658 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cxdt8w" event={"ID":"5e47aa0e-c660-4f88-8fb1-8c873268879d","Type":"ContainerDied","Data":"33f1b58ffd98d9f1c82d12554641ddfde0213d47e739ff2ac668b7e03386c168"} Oct 06 10:09:20 crc kubenswrapper[4824]: I1006 10:09:20.907679 4824 generic.go:334] "Generic (PLEG): container finished" podID="5e47aa0e-c660-4f88-8fb1-8c873268879d" containerID="6c6e93531e0d206fd065c5493bfe46c39087c90e8006e023c99b58071e82b112" exitCode=0 Oct 06 10:09:20 crc kubenswrapper[4824]: I1006 10:09:20.907811 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cxdt8w" event={"ID":"5e47aa0e-c660-4f88-8fb1-8c873268879d","Type":"ContainerDied","Data":"6c6e93531e0d206fd065c5493bfe46c39087c90e8006e023c99b58071e82b112"} Oct 06 10:09:22 crc kubenswrapper[4824]: I1006 10:09:22.325747 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cxdt8w" Oct 06 10:09:22 crc kubenswrapper[4824]: I1006 10:09:22.418271 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5e47aa0e-c660-4f88-8fb1-8c873268879d-bundle\") pod \"5e47aa0e-c660-4f88-8fb1-8c873268879d\" (UID: \"5e47aa0e-c660-4f88-8fb1-8c873268879d\") " Oct 06 10:09:22 crc kubenswrapper[4824]: I1006 10:09:22.418545 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c7lkf\" (UniqueName: \"kubernetes.io/projected/5e47aa0e-c660-4f88-8fb1-8c873268879d-kube-api-access-c7lkf\") pod \"5e47aa0e-c660-4f88-8fb1-8c873268879d\" (UID: \"5e47aa0e-c660-4f88-8fb1-8c873268879d\") " Oct 06 10:09:22 crc kubenswrapper[4824]: I1006 10:09:22.418602 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5e47aa0e-c660-4f88-8fb1-8c873268879d-util\") pod \"5e47aa0e-c660-4f88-8fb1-8c873268879d\" (UID: \"5e47aa0e-c660-4f88-8fb1-8c873268879d\") " Oct 06 10:09:22 crc kubenswrapper[4824]: I1006 10:09:22.419596 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5e47aa0e-c660-4f88-8fb1-8c873268879d-bundle" (OuterVolumeSpecName: "bundle") pod "5e47aa0e-c660-4f88-8fb1-8c873268879d" (UID: "5e47aa0e-c660-4f88-8fb1-8c873268879d"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:09:22 crc kubenswrapper[4824]: I1006 10:09:22.425563 4824 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5e47aa0e-c660-4f88-8fb1-8c873268879d-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 10:09:22 crc kubenswrapper[4824]: I1006 10:09:22.427318 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5e47aa0e-c660-4f88-8fb1-8c873268879d-kube-api-access-c7lkf" (OuterVolumeSpecName: "kube-api-access-c7lkf") pod "5e47aa0e-c660-4f88-8fb1-8c873268879d" (UID: "5e47aa0e-c660-4f88-8fb1-8c873268879d"). InnerVolumeSpecName "kube-api-access-c7lkf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:09:22 crc kubenswrapper[4824]: I1006 10:09:22.527754 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c7lkf\" (UniqueName: \"kubernetes.io/projected/5e47aa0e-c660-4f88-8fb1-8c873268879d-kube-api-access-c7lkf\") on node \"crc\" DevicePath \"\"" Oct 06 10:09:22 crc kubenswrapper[4824]: I1006 10:09:22.630688 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5e47aa0e-c660-4f88-8fb1-8c873268879d-util" (OuterVolumeSpecName: "util") pod "5e47aa0e-c660-4f88-8fb1-8c873268879d" (UID: "5e47aa0e-c660-4f88-8fb1-8c873268879d"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:09:22 crc kubenswrapper[4824]: I1006 10:09:22.731163 4824 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5e47aa0e-c660-4f88-8fb1-8c873268879d-util\") on node \"crc\" DevicePath \"\"" Oct 06 10:09:22 crc kubenswrapper[4824]: I1006 10:09:22.927872 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cxdt8w" event={"ID":"5e47aa0e-c660-4f88-8fb1-8c873268879d","Type":"ContainerDied","Data":"6f0b93f016bbda1dc66944473f518c695267b7ef2c4049c1b38e14acb304a14c"} Oct 06 10:09:22 crc kubenswrapper[4824]: I1006 10:09:22.928440 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6f0b93f016bbda1dc66944473f518c695267b7ef2c4049c1b38e14acb304a14c" Oct 06 10:09:22 crc kubenswrapper[4824]: I1006 10:09:22.928005 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cxdt8w" Oct 06 10:09:25 crc kubenswrapper[4824]: I1006 10:09:25.013891 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-grzj6"] Oct 06 10:09:25 crc kubenswrapper[4824]: E1006 10:09:25.014249 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e47aa0e-c660-4f88-8fb1-8c873268879d" containerName="extract" Oct 06 10:09:25 crc kubenswrapper[4824]: I1006 10:09:25.014265 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e47aa0e-c660-4f88-8fb1-8c873268879d" containerName="extract" Oct 06 10:09:25 crc kubenswrapper[4824]: E1006 10:09:25.014274 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e47aa0e-c660-4f88-8fb1-8c873268879d" containerName="util" Oct 06 10:09:25 crc kubenswrapper[4824]: I1006 10:09:25.014281 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e47aa0e-c660-4f88-8fb1-8c873268879d" containerName="util" Oct 06 10:09:25 crc kubenswrapper[4824]: E1006 10:09:25.014299 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e47aa0e-c660-4f88-8fb1-8c873268879d" containerName="pull" Oct 06 10:09:25 crc kubenswrapper[4824]: I1006 10:09:25.014306 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e47aa0e-c660-4f88-8fb1-8c873268879d" containerName="pull" Oct 06 10:09:25 crc kubenswrapper[4824]: I1006 10:09:25.014413 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e47aa0e-c660-4f88-8fb1-8c873268879d" containerName="extract" Oct 06 10:09:25 crc kubenswrapper[4824]: I1006 10:09:25.014841 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-858ddd8f98-grzj6" Oct 06 10:09:25 crc kubenswrapper[4824]: I1006 10:09:25.018249 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-bmsdf" Oct 06 10:09:25 crc kubenswrapper[4824]: I1006 10:09:25.018328 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Oct 06 10:09:25 crc kubenswrapper[4824]: I1006 10:09:25.018716 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Oct 06 10:09:25 crc kubenswrapper[4824]: I1006 10:09:25.041702 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-grzj6"] Oct 06 10:09:25 crc kubenswrapper[4824]: I1006 10:09:25.168685 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ht8r2\" (UniqueName: \"kubernetes.io/projected/230d69e7-8f29-41a3-af2f-b1b93c58e8c2-kube-api-access-ht8r2\") pod \"nmstate-operator-858ddd8f98-grzj6\" (UID: \"230d69e7-8f29-41a3-af2f-b1b93c58e8c2\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-grzj6" Oct 06 10:09:25 crc kubenswrapper[4824]: I1006 10:09:25.270794 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ht8r2\" (UniqueName: \"kubernetes.io/projected/230d69e7-8f29-41a3-af2f-b1b93c58e8c2-kube-api-access-ht8r2\") pod \"nmstate-operator-858ddd8f98-grzj6\" (UID: \"230d69e7-8f29-41a3-af2f-b1b93c58e8c2\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-grzj6" Oct 06 10:09:25 crc kubenswrapper[4824]: I1006 10:09:25.296413 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ht8r2\" (UniqueName: \"kubernetes.io/projected/230d69e7-8f29-41a3-af2f-b1b93c58e8c2-kube-api-access-ht8r2\") pod \"nmstate-operator-858ddd8f98-grzj6\" (UID: \"230d69e7-8f29-41a3-af2f-b1b93c58e8c2\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-grzj6" Oct 06 10:09:25 crc kubenswrapper[4824]: I1006 10:09:25.331780 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-858ddd8f98-grzj6" Oct 06 10:09:25 crc kubenswrapper[4824]: I1006 10:09:25.574764 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-grzj6"] Oct 06 10:09:25 crc kubenswrapper[4824]: I1006 10:09:25.950369 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-858ddd8f98-grzj6" event={"ID":"230d69e7-8f29-41a3-af2f-b1b93c58e8c2","Type":"ContainerStarted","Data":"469e113ef455cda26527385ef6fb44d908d1344a4bf790790571faf6b3fb3b32"} Oct 06 10:09:28 crc kubenswrapper[4824]: I1006 10:09:28.973342 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-858ddd8f98-grzj6" event={"ID":"230d69e7-8f29-41a3-af2f-b1b93c58e8c2","Type":"ContainerStarted","Data":"8fd5ec1b24e637d57787741f09e5fccb761cbcb222d3c8710d01b7645ad338ad"} Oct 06 10:09:29 crc kubenswrapper[4824]: I1006 10:09:29.004686 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-858ddd8f98-grzj6" podStartSLOduration=2.69890264 podStartE2EDuration="5.004651831s" podCreationTimestamp="2025-10-06 10:09:24 +0000 UTC" firstStartedPulling="2025-10-06 10:09:25.588769889 +0000 UTC m=+674.953192750" lastFinishedPulling="2025-10-06 10:09:27.89451908 +0000 UTC m=+677.258941941" observedRunningTime="2025-10-06 10:09:29.001366579 +0000 UTC m=+678.365789480" watchObservedRunningTime="2025-10-06 10:09:29.004651831 +0000 UTC m=+678.369074732" Oct 06 10:09:30 crc kubenswrapper[4824]: I1006 10:09:30.051170 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-6vncz"] Oct 06 10:09:30 crc kubenswrapper[4824]: I1006 10:09:30.053198 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-6vncz" Oct 06 10:09:30 crc kubenswrapper[4824]: I1006 10:09:30.056354 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-q4w7v" Oct 06 10:09:30 crc kubenswrapper[4824]: I1006 10:09:30.062863 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-b7dd2"] Oct 06 10:09:30 crc kubenswrapper[4824]: I1006 10:09:30.064678 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-b7dd2" Oct 06 10:09:30 crc kubenswrapper[4824]: I1006 10:09:30.067913 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Oct 06 10:09:30 crc kubenswrapper[4824]: I1006 10:09:30.083560 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-6vncz"] Oct 06 10:09:30 crc kubenswrapper[4824]: I1006 10:09:30.095410 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-b7dd2"] Oct 06 10:09:30 crc kubenswrapper[4824]: I1006 10:09:30.102267 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-jbv5q"] Oct 06 10:09:30 crc kubenswrapper[4824]: I1006 10:09:30.103173 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-jbv5q" Oct 06 10:09:30 crc kubenswrapper[4824]: I1006 10:09:30.147447 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6pmsl\" (UniqueName: \"kubernetes.io/projected/355b7d6a-3ab4-4d16-8923-f15646b3af54-kube-api-access-6pmsl\") pod \"nmstate-metrics-fdff9cb8d-6vncz\" (UID: \"355b7d6a-3ab4-4d16-8923-f15646b3af54\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-6vncz" Oct 06 10:09:30 crc kubenswrapper[4824]: I1006 10:09:30.248650 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/3048d3eb-7f41-4b26-92b3-6d60f5a3fa18-dbus-socket\") pod \"nmstate-handler-jbv5q\" (UID: \"3048d3eb-7f41-4b26-92b3-6d60f5a3fa18\") " pod="openshift-nmstate/nmstate-handler-jbv5q" Oct 06 10:09:30 crc kubenswrapper[4824]: I1006 10:09:30.248702 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/b706331c-d771-4212-950f-6e3fd16faa43-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-b7dd2\" (UID: \"b706331c-d771-4212-950f-6e3fd16faa43\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-b7dd2" Oct 06 10:09:30 crc kubenswrapper[4824]: I1006 10:09:30.248733 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9j5zm\" (UniqueName: \"kubernetes.io/projected/b706331c-d771-4212-950f-6e3fd16faa43-kube-api-access-9j5zm\") pod \"nmstate-webhook-6cdbc54649-b7dd2\" (UID: \"b706331c-d771-4212-950f-6e3fd16faa43\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-b7dd2" Oct 06 10:09:30 crc kubenswrapper[4824]: I1006 10:09:30.248764 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/3048d3eb-7f41-4b26-92b3-6d60f5a3fa18-ovs-socket\") pod \"nmstate-handler-jbv5q\" (UID: \"3048d3eb-7f41-4b26-92b3-6d60f5a3fa18\") " pod="openshift-nmstate/nmstate-handler-jbv5q" Oct 06 10:09:30 crc kubenswrapper[4824]: I1006 10:09:30.248917 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m4gqr\" (UniqueName: \"kubernetes.io/projected/3048d3eb-7f41-4b26-92b3-6d60f5a3fa18-kube-api-access-m4gqr\") pod \"nmstate-handler-jbv5q\" (UID: \"3048d3eb-7f41-4b26-92b3-6d60f5a3fa18\") " pod="openshift-nmstate/nmstate-handler-jbv5q" Oct 06 10:09:30 crc kubenswrapper[4824]: I1006 10:09:30.249202 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/3048d3eb-7f41-4b26-92b3-6d60f5a3fa18-nmstate-lock\") pod \"nmstate-handler-jbv5q\" (UID: \"3048d3eb-7f41-4b26-92b3-6d60f5a3fa18\") " pod="openshift-nmstate/nmstate-handler-jbv5q" Oct 06 10:09:30 crc kubenswrapper[4824]: I1006 10:09:30.249267 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6pmsl\" (UniqueName: \"kubernetes.io/projected/355b7d6a-3ab4-4d16-8923-f15646b3af54-kube-api-access-6pmsl\") pod \"nmstate-metrics-fdff9cb8d-6vncz\" (UID: \"355b7d6a-3ab4-4d16-8923-f15646b3af54\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-6vncz" Oct 06 10:09:30 crc kubenswrapper[4824]: I1006 10:09:30.256547 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-7nwh2"] Oct 06 10:09:30 crc kubenswrapper[4824]: I1006 10:09:30.257424 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-7nwh2" Oct 06 10:09:30 crc kubenswrapper[4824]: I1006 10:09:30.258998 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Oct 06 10:09:30 crc kubenswrapper[4824]: I1006 10:09:30.259379 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Oct 06 10:09:30 crc kubenswrapper[4824]: I1006 10:09:30.259617 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-9fzgt" Oct 06 10:09:30 crc kubenswrapper[4824]: I1006 10:09:30.276313 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6pmsl\" (UniqueName: \"kubernetes.io/projected/355b7d6a-3ab4-4d16-8923-f15646b3af54-kube-api-access-6pmsl\") pod \"nmstate-metrics-fdff9cb8d-6vncz\" (UID: \"355b7d6a-3ab4-4d16-8923-f15646b3af54\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-6vncz" Oct 06 10:09:30 crc kubenswrapper[4824]: I1006 10:09:30.276530 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-7nwh2"] Oct 06 10:09:30 crc kubenswrapper[4824]: I1006 10:09:30.351194 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/3048d3eb-7f41-4b26-92b3-6d60f5a3fa18-ovs-socket\") pod \"nmstate-handler-jbv5q\" (UID: \"3048d3eb-7f41-4b26-92b3-6d60f5a3fa18\") " pod="openshift-nmstate/nmstate-handler-jbv5q" Oct 06 10:09:30 crc kubenswrapper[4824]: I1006 10:09:30.351918 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/3048d3eb-7f41-4b26-92b3-6d60f5a3fa18-ovs-socket\") pod \"nmstate-handler-jbv5q\" (UID: \"3048d3eb-7f41-4b26-92b3-6d60f5a3fa18\") " pod="openshift-nmstate/nmstate-handler-jbv5q" Oct 06 10:09:30 crc kubenswrapper[4824]: I1006 10:09:30.359336 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m4gqr\" (UniqueName: \"kubernetes.io/projected/3048d3eb-7f41-4b26-92b3-6d60f5a3fa18-kube-api-access-m4gqr\") pod \"nmstate-handler-jbv5q\" (UID: \"3048d3eb-7f41-4b26-92b3-6d60f5a3fa18\") " pod="openshift-nmstate/nmstate-handler-jbv5q" Oct 06 10:09:30 crc kubenswrapper[4824]: I1006 10:09:30.359668 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/3048d3eb-7f41-4b26-92b3-6d60f5a3fa18-nmstate-lock\") pod \"nmstate-handler-jbv5q\" (UID: \"3048d3eb-7f41-4b26-92b3-6d60f5a3fa18\") " pod="openshift-nmstate/nmstate-handler-jbv5q" Oct 06 10:09:30 crc kubenswrapper[4824]: I1006 10:09:30.359713 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/00271786-ab42-4507-a545-8c5bdf1b976d-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-7nwh2\" (UID: \"00271786-ab42-4507-a545-8c5bdf1b976d\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-7nwh2" Oct 06 10:09:30 crc kubenswrapper[4824]: I1006 10:09:30.359806 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/3048d3eb-7f41-4b26-92b3-6d60f5a3fa18-dbus-socket\") pod \"nmstate-handler-jbv5q\" (UID: \"3048d3eb-7f41-4b26-92b3-6d60f5a3fa18\") " pod="openshift-nmstate/nmstate-handler-jbv5q" Oct 06 10:09:30 crc kubenswrapper[4824]: I1006 10:09:30.359824 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/3048d3eb-7f41-4b26-92b3-6d60f5a3fa18-nmstate-lock\") pod \"nmstate-handler-jbv5q\" (UID: \"3048d3eb-7f41-4b26-92b3-6d60f5a3fa18\") " pod="openshift-nmstate/nmstate-handler-jbv5q" Oct 06 10:09:30 crc kubenswrapper[4824]: I1006 10:09:30.359839 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/b706331c-d771-4212-950f-6e3fd16faa43-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-b7dd2\" (UID: \"b706331c-d771-4212-950f-6e3fd16faa43\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-b7dd2" Oct 06 10:09:30 crc kubenswrapper[4824]: I1006 10:09:30.359871 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/00271786-ab42-4507-a545-8c5bdf1b976d-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-7nwh2\" (UID: \"00271786-ab42-4507-a545-8c5bdf1b976d\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-7nwh2" Oct 06 10:09:30 crc kubenswrapper[4824]: I1006 10:09:30.359917 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9j5zm\" (UniqueName: \"kubernetes.io/projected/b706331c-d771-4212-950f-6e3fd16faa43-kube-api-access-9j5zm\") pod \"nmstate-webhook-6cdbc54649-b7dd2\" (UID: \"b706331c-d771-4212-950f-6e3fd16faa43\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-b7dd2" Oct 06 10:09:30 crc kubenswrapper[4824]: I1006 10:09:30.360075 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dtvd5\" (UniqueName: \"kubernetes.io/projected/00271786-ab42-4507-a545-8c5bdf1b976d-kube-api-access-dtvd5\") pod \"nmstate-console-plugin-6b874cbd85-7nwh2\" (UID: \"00271786-ab42-4507-a545-8c5bdf1b976d\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-7nwh2" Oct 06 10:09:30 crc kubenswrapper[4824]: I1006 10:09:30.360277 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/3048d3eb-7f41-4b26-92b3-6d60f5a3fa18-dbus-socket\") pod \"nmstate-handler-jbv5q\" (UID: \"3048d3eb-7f41-4b26-92b3-6d60f5a3fa18\") " pod="openshift-nmstate/nmstate-handler-jbv5q" Oct 06 10:09:30 crc kubenswrapper[4824]: I1006 10:09:30.373398 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/b706331c-d771-4212-950f-6e3fd16faa43-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-b7dd2\" (UID: \"b706331c-d771-4212-950f-6e3fd16faa43\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-b7dd2" Oct 06 10:09:30 crc kubenswrapper[4824]: I1006 10:09:30.378748 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9j5zm\" (UniqueName: \"kubernetes.io/projected/b706331c-d771-4212-950f-6e3fd16faa43-kube-api-access-9j5zm\") pod \"nmstate-webhook-6cdbc54649-b7dd2\" (UID: \"b706331c-d771-4212-950f-6e3fd16faa43\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-b7dd2" Oct 06 10:09:30 crc kubenswrapper[4824]: I1006 10:09:30.382474 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-6vncz" Oct 06 10:09:30 crc kubenswrapper[4824]: I1006 10:09:30.382654 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m4gqr\" (UniqueName: \"kubernetes.io/projected/3048d3eb-7f41-4b26-92b3-6d60f5a3fa18-kube-api-access-m4gqr\") pod \"nmstate-handler-jbv5q\" (UID: \"3048d3eb-7f41-4b26-92b3-6d60f5a3fa18\") " pod="openshift-nmstate/nmstate-handler-jbv5q" Oct 06 10:09:30 crc kubenswrapper[4824]: I1006 10:09:30.401610 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-b7dd2" Oct 06 10:09:30 crc kubenswrapper[4824]: I1006 10:09:30.422317 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-jbv5q" Oct 06 10:09:30 crc kubenswrapper[4824]: I1006 10:09:30.461577 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/00271786-ab42-4507-a545-8c5bdf1b976d-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-7nwh2\" (UID: \"00271786-ab42-4507-a545-8c5bdf1b976d\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-7nwh2" Oct 06 10:09:30 crc kubenswrapper[4824]: I1006 10:09:30.462325 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/00271786-ab42-4507-a545-8c5bdf1b976d-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-7nwh2\" (UID: \"00271786-ab42-4507-a545-8c5bdf1b976d\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-7nwh2" Oct 06 10:09:30 crc kubenswrapper[4824]: I1006 10:09:30.462358 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dtvd5\" (UniqueName: \"kubernetes.io/projected/00271786-ab42-4507-a545-8c5bdf1b976d-kube-api-access-dtvd5\") pod \"nmstate-console-plugin-6b874cbd85-7nwh2\" (UID: \"00271786-ab42-4507-a545-8c5bdf1b976d\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-7nwh2" Oct 06 10:09:30 crc kubenswrapper[4824]: I1006 10:09:30.467054 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/00271786-ab42-4507-a545-8c5bdf1b976d-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-7nwh2\" (UID: \"00271786-ab42-4507-a545-8c5bdf1b976d\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-7nwh2" Oct 06 10:09:30 crc kubenswrapper[4824]: I1006 10:09:30.467330 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/00271786-ab42-4507-a545-8c5bdf1b976d-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-7nwh2\" (UID: \"00271786-ab42-4507-a545-8c5bdf1b976d\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-7nwh2" Oct 06 10:09:30 crc kubenswrapper[4824]: I1006 10:09:30.487662 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dtvd5\" (UniqueName: \"kubernetes.io/projected/00271786-ab42-4507-a545-8c5bdf1b976d-kube-api-access-dtvd5\") pod \"nmstate-console-plugin-6b874cbd85-7nwh2\" (UID: \"00271786-ab42-4507-a545-8c5bdf1b976d\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-7nwh2" Oct 06 10:09:30 crc kubenswrapper[4824]: I1006 10:09:30.517543 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-5976bc9cd4-ltdb2"] Oct 06 10:09:30 crc kubenswrapper[4824]: I1006 10:09:30.518380 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-5976bc9cd4-ltdb2" Oct 06 10:09:30 crc kubenswrapper[4824]: I1006 10:09:30.542087 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-5976bc9cd4-ltdb2"] Oct 06 10:09:30 crc kubenswrapper[4824]: I1006 10:09:30.576220 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-7nwh2" Oct 06 10:09:30 crc kubenswrapper[4824]: I1006 10:09:30.668109 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/691bfc14-d8bd-48fd-9557-ec9d85dd6618-console-config\") pod \"console-5976bc9cd4-ltdb2\" (UID: \"691bfc14-d8bd-48fd-9557-ec9d85dd6618\") " pod="openshift-console/console-5976bc9cd4-ltdb2" Oct 06 10:09:30 crc kubenswrapper[4824]: I1006 10:09:30.668292 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/691bfc14-d8bd-48fd-9557-ec9d85dd6618-console-serving-cert\") pod \"console-5976bc9cd4-ltdb2\" (UID: \"691bfc14-d8bd-48fd-9557-ec9d85dd6618\") " pod="openshift-console/console-5976bc9cd4-ltdb2" Oct 06 10:09:30 crc kubenswrapper[4824]: I1006 10:09:30.668407 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/691bfc14-d8bd-48fd-9557-ec9d85dd6618-trusted-ca-bundle\") pod \"console-5976bc9cd4-ltdb2\" (UID: \"691bfc14-d8bd-48fd-9557-ec9d85dd6618\") " pod="openshift-console/console-5976bc9cd4-ltdb2" Oct 06 10:09:30 crc kubenswrapper[4824]: I1006 10:09:30.668490 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/691bfc14-d8bd-48fd-9557-ec9d85dd6618-service-ca\") pod \"console-5976bc9cd4-ltdb2\" (UID: \"691bfc14-d8bd-48fd-9557-ec9d85dd6618\") " pod="openshift-console/console-5976bc9cd4-ltdb2" Oct 06 10:09:30 crc kubenswrapper[4824]: I1006 10:09:30.668529 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4w4sf\" (UniqueName: \"kubernetes.io/projected/691bfc14-d8bd-48fd-9557-ec9d85dd6618-kube-api-access-4w4sf\") pod \"console-5976bc9cd4-ltdb2\" (UID: \"691bfc14-d8bd-48fd-9557-ec9d85dd6618\") " pod="openshift-console/console-5976bc9cd4-ltdb2" Oct 06 10:09:30 crc kubenswrapper[4824]: I1006 10:09:30.668553 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/691bfc14-d8bd-48fd-9557-ec9d85dd6618-console-oauth-config\") pod \"console-5976bc9cd4-ltdb2\" (UID: \"691bfc14-d8bd-48fd-9557-ec9d85dd6618\") " pod="openshift-console/console-5976bc9cd4-ltdb2" Oct 06 10:09:30 crc kubenswrapper[4824]: I1006 10:09:30.668576 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/691bfc14-d8bd-48fd-9557-ec9d85dd6618-oauth-serving-cert\") pod \"console-5976bc9cd4-ltdb2\" (UID: \"691bfc14-d8bd-48fd-9557-ec9d85dd6618\") " pod="openshift-console/console-5976bc9cd4-ltdb2" Oct 06 10:09:30 crc kubenswrapper[4824]: I1006 10:09:30.769663 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/691bfc14-d8bd-48fd-9557-ec9d85dd6618-console-config\") pod \"console-5976bc9cd4-ltdb2\" (UID: \"691bfc14-d8bd-48fd-9557-ec9d85dd6618\") " pod="openshift-console/console-5976bc9cd4-ltdb2" Oct 06 10:09:30 crc kubenswrapper[4824]: I1006 10:09:30.770173 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/691bfc14-d8bd-48fd-9557-ec9d85dd6618-console-serving-cert\") pod \"console-5976bc9cd4-ltdb2\" (UID: \"691bfc14-d8bd-48fd-9557-ec9d85dd6618\") " pod="openshift-console/console-5976bc9cd4-ltdb2" Oct 06 10:09:30 crc kubenswrapper[4824]: I1006 10:09:30.770190 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/691bfc14-d8bd-48fd-9557-ec9d85dd6618-trusted-ca-bundle\") pod \"console-5976bc9cd4-ltdb2\" (UID: \"691bfc14-d8bd-48fd-9557-ec9d85dd6618\") " pod="openshift-console/console-5976bc9cd4-ltdb2" Oct 06 10:09:30 crc kubenswrapper[4824]: I1006 10:09:30.770229 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/691bfc14-d8bd-48fd-9557-ec9d85dd6618-service-ca\") pod \"console-5976bc9cd4-ltdb2\" (UID: \"691bfc14-d8bd-48fd-9557-ec9d85dd6618\") " pod="openshift-console/console-5976bc9cd4-ltdb2" Oct 06 10:09:30 crc kubenswrapper[4824]: I1006 10:09:30.770255 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4w4sf\" (UniqueName: \"kubernetes.io/projected/691bfc14-d8bd-48fd-9557-ec9d85dd6618-kube-api-access-4w4sf\") pod \"console-5976bc9cd4-ltdb2\" (UID: \"691bfc14-d8bd-48fd-9557-ec9d85dd6618\") " pod="openshift-console/console-5976bc9cd4-ltdb2" Oct 06 10:09:30 crc kubenswrapper[4824]: I1006 10:09:30.770273 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/691bfc14-d8bd-48fd-9557-ec9d85dd6618-console-oauth-config\") pod \"console-5976bc9cd4-ltdb2\" (UID: \"691bfc14-d8bd-48fd-9557-ec9d85dd6618\") " pod="openshift-console/console-5976bc9cd4-ltdb2" Oct 06 10:09:30 crc kubenswrapper[4824]: I1006 10:09:30.770291 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/691bfc14-d8bd-48fd-9557-ec9d85dd6618-oauth-serving-cert\") pod \"console-5976bc9cd4-ltdb2\" (UID: \"691bfc14-d8bd-48fd-9557-ec9d85dd6618\") " pod="openshift-console/console-5976bc9cd4-ltdb2" Oct 06 10:09:30 crc kubenswrapper[4824]: I1006 10:09:30.771029 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/691bfc14-d8bd-48fd-9557-ec9d85dd6618-console-config\") pod \"console-5976bc9cd4-ltdb2\" (UID: \"691bfc14-d8bd-48fd-9557-ec9d85dd6618\") " pod="openshift-console/console-5976bc9cd4-ltdb2" Oct 06 10:09:30 crc kubenswrapper[4824]: I1006 10:09:30.771195 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/691bfc14-d8bd-48fd-9557-ec9d85dd6618-oauth-serving-cert\") pod \"console-5976bc9cd4-ltdb2\" (UID: \"691bfc14-d8bd-48fd-9557-ec9d85dd6618\") " pod="openshift-console/console-5976bc9cd4-ltdb2" Oct 06 10:09:30 crc kubenswrapper[4824]: I1006 10:09:30.772444 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/691bfc14-d8bd-48fd-9557-ec9d85dd6618-service-ca\") pod \"console-5976bc9cd4-ltdb2\" (UID: \"691bfc14-d8bd-48fd-9557-ec9d85dd6618\") " pod="openshift-console/console-5976bc9cd4-ltdb2" Oct 06 10:09:30 crc kubenswrapper[4824]: I1006 10:09:30.776473 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/691bfc14-d8bd-48fd-9557-ec9d85dd6618-console-serving-cert\") pod \"console-5976bc9cd4-ltdb2\" (UID: \"691bfc14-d8bd-48fd-9557-ec9d85dd6618\") " pod="openshift-console/console-5976bc9cd4-ltdb2" Oct 06 10:09:30 crc kubenswrapper[4824]: I1006 10:09:30.776492 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/691bfc14-d8bd-48fd-9557-ec9d85dd6618-trusted-ca-bundle\") pod \"console-5976bc9cd4-ltdb2\" (UID: \"691bfc14-d8bd-48fd-9557-ec9d85dd6618\") " pod="openshift-console/console-5976bc9cd4-ltdb2" Oct 06 10:09:30 crc kubenswrapper[4824]: I1006 10:09:30.776836 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/691bfc14-d8bd-48fd-9557-ec9d85dd6618-console-oauth-config\") pod \"console-5976bc9cd4-ltdb2\" (UID: \"691bfc14-d8bd-48fd-9557-ec9d85dd6618\") " pod="openshift-console/console-5976bc9cd4-ltdb2" Oct 06 10:09:30 crc kubenswrapper[4824]: I1006 10:09:30.787544 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4w4sf\" (UniqueName: \"kubernetes.io/projected/691bfc14-d8bd-48fd-9557-ec9d85dd6618-kube-api-access-4w4sf\") pod \"console-5976bc9cd4-ltdb2\" (UID: \"691bfc14-d8bd-48fd-9557-ec9d85dd6618\") " pod="openshift-console/console-5976bc9cd4-ltdb2" Oct 06 10:09:30 crc kubenswrapper[4824]: I1006 10:09:30.851696 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-7nwh2"] Oct 06 10:09:30 crc kubenswrapper[4824]: I1006 10:09:30.858748 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-5976bc9cd4-ltdb2" Oct 06 10:09:30 crc kubenswrapper[4824]: I1006 10:09:30.898931 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-b7dd2"] Oct 06 10:09:30 crc kubenswrapper[4824]: W1006 10:09:30.902944 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb706331c_d771_4212_950f_6e3fd16faa43.slice/crio-ed4efc48932ed2b26138a1b5f347c93a1235c80a6003728694d72f1f4a2a8b30 WatchSource:0}: Error finding container ed4efc48932ed2b26138a1b5f347c93a1235c80a6003728694d72f1f4a2a8b30: Status 404 returned error can't find the container with id ed4efc48932ed2b26138a1b5f347c93a1235c80a6003728694d72f1f4a2a8b30 Oct 06 10:09:30 crc kubenswrapper[4824]: I1006 10:09:30.975084 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-6vncz"] Oct 06 10:09:30 crc kubenswrapper[4824]: W1006 10:09:30.976881 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod355b7d6a_3ab4_4d16_8923_f15646b3af54.slice/crio-ce885aa5b812e64066e56ade7657c4933690cf41abf21c96bace9a72c8e36b48 WatchSource:0}: Error finding container ce885aa5b812e64066e56ade7657c4933690cf41abf21c96bace9a72c8e36b48: Status 404 returned error can't find the container with id ce885aa5b812e64066e56ade7657c4933690cf41abf21c96bace9a72c8e36b48 Oct 06 10:09:31 crc kubenswrapper[4824]: I1006 10:09:30.993466 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-b7dd2" event={"ID":"b706331c-d771-4212-950f-6e3fd16faa43","Type":"ContainerStarted","Data":"ed4efc48932ed2b26138a1b5f347c93a1235c80a6003728694d72f1f4a2a8b30"} Oct 06 10:09:31 crc kubenswrapper[4824]: I1006 10:09:30.998023 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-6vncz" event={"ID":"355b7d6a-3ab4-4d16-8923-f15646b3af54","Type":"ContainerStarted","Data":"ce885aa5b812e64066e56ade7657c4933690cf41abf21c96bace9a72c8e36b48"} Oct 06 10:09:31 crc kubenswrapper[4824]: I1006 10:09:30.998919 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-7nwh2" event={"ID":"00271786-ab42-4507-a545-8c5bdf1b976d","Type":"ContainerStarted","Data":"b3d0c2a460396a173e63c2dbec1be698076736143c15ad7acd13c4407640ae44"} Oct 06 10:09:31 crc kubenswrapper[4824]: I1006 10:09:31.000614 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-jbv5q" event={"ID":"3048d3eb-7f41-4b26-92b3-6d60f5a3fa18","Type":"ContainerStarted","Data":"dfaa021a3bbfaebac3acb9707dc6fb17783e181d73efb740fb2b9bbdeb765099"} Oct 06 10:09:31 crc kubenswrapper[4824]: I1006 10:09:31.314556 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-5976bc9cd4-ltdb2"] Oct 06 10:09:31 crc kubenswrapper[4824]: W1006 10:09:31.323904 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod691bfc14_d8bd_48fd_9557_ec9d85dd6618.slice/crio-173ef9a2a8a6241ab2b977a9e6f47c6e58f303aacb1b7cc9064c2203ee5938ae WatchSource:0}: Error finding container 173ef9a2a8a6241ab2b977a9e6f47c6e58f303aacb1b7cc9064c2203ee5938ae: Status 404 returned error can't find the container with id 173ef9a2a8a6241ab2b977a9e6f47c6e58f303aacb1b7cc9064c2203ee5938ae Oct 06 10:09:32 crc kubenswrapper[4824]: I1006 10:09:32.014090 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-5976bc9cd4-ltdb2" event={"ID":"691bfc14-d8bd-48fd-9557-ec9d85dd6618","Type":"ContainerStarted","Data":"3e82739ff2120b2285c91a65c9f67a26b43c830d031b23f0b60aa7bdddad64ba"} Oct 06 10:09:32 crc kubenswrapper[4824]: I1006 10:09:32.014151 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-5976bc9cd4-ltdb2" event={"ID":"691bfc14-d8bd-48fd-9557-ec9d85dd6618","Type":"ContainerStarted","Data":"173ef9a2a8a6241ab2b977a9e6f47c6e58f303aacb1b7cc9064c2203ee5938ae"} Oct 06 10:09:32 crc kubenswrapper[4824]: I1006 10:09:32.063891 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-5976bc9cd4-ltdb2" podStartSLOduration=2.063858614 podStartE2EDuration="2.063858614s" podCreationTimestamp="2025-10-06 10:09:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:09:32.046198793 +0000 UTC m=+681.410621734" watchObservedRunningTime="2025-10-06 10:09:32.063858614 +0000 UTC m=+681.428281485" Oct 06 10:09:34 crc kubenswrapper[4824]: I1006 10:09:34.045097 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-6vncz" event={"ID":"355b7d6a-3ab4-4d16-8923-f15646b3af54","Type":"ContainerStarted","Data":"63b36106ad9213ba27eafed59f2270b263eda3ff6766ce22cf200725ed49139a"} Oct 06 10:09:34 crc kubenswrapper[4824]: I1006 10:09:34.048496 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-7nwh2" event={"ID":"00271786-ab42-4507-a545-8c5bdf1b976d","Type":"ContainerStarted","Data":"3b6679fdf9cb4232d70fc3306b5b54f3875d2bd37da0d121b0993ca13eed18f3"} Oct 06 10:09:34 crc kubenswrapper[4824]: I1006 10:09:34.052155 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-b7dd2" event={"ID":"b706331c-d771-4212-950f-6e3fd16faa43","Type":"ContainerStarted","Data":"7234c26b65988391f951489d5a870cee97c38e78f2d8995c3005d8de1fba3149"} Oct 06 10:09:34 crc kubenswrapper[4824]: I1006 10:09:34.052362 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-b7dd2" Oct 06 10:09:34 crc kubenswrapper[4824]: I1006 10:09:34.077844 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-7nwh2" podStartSLOduration=1.244273098 podStartE2EDuration="4.077810432s" podCreationTimestamp="2025-10-06 10:09:30 +0000 UTC" firstStartedPulling="2025-10-06 10:09:30.855376222 +0000 UTC m=+680.219799083" lastFinishedPulling="2025-10-06 10:09:33.688913526 +0000 UTC m=+683.053336417" observedRunningTime="2025-10-06 10:09:34.067639296 +0000 UTC m=+683.432062197" watchObservedRunningTime="2025-10-06 10:09:34.077810432 +0000 UTC m=+683.442233313" Oct 06 10:09:34 crc kubenswrapper[4824]: I1006 10:09:34.089061 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-b7dd2" podStartSLOduration=1.30586699 podStartE2EDuration="4.089036473s" podCreationTimestamp="2025-10-06 10:09:30 +0000 UTC" firstStartedPulling="2025-10-06 10:09:30.905993159 +0000 UTC m=+680.270416020" lastFinishedPulling="2025-10-06 10:09:33.689162632 +0000 UTC m=+683.053585503" observedRunningTime="2025-10-06 10:09:34.08696994 +0000 UTC m=+683.451392791" watchObservedRunningTime="2025-10-06 10:09:34.089036473 +0000 UTC m=+683.453459334" Oct 06 10:09:35 crc kubenswrapper[4824]: I1006 10:09:35.086675 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-jbv5q" event={"ID":"3048d3eb-7f41-4b26-92b3-6d60f5a3fa18","Type":"ContainerStarted","Data":"8d1c2ca5269dbf32a8ee82b43c8a1f7c219f5b3ac6eaf6a2f7499a282c20e8fe"} Oct 06 10:09:35 crc kubenswrapper[4824]: I1006 10:09:35.087117 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-jbv5q" Oct 06 10:09:35 crc kubenswrapper[4824]: I1006 10:09:35.113928 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-jbv5q" podStartSLOduration=1.866212907 podStartE2EDuration="5.113909209s" podCreationTimestamp="2025-10-06 10:09:30 +0000 UTC" firstStartedPulling="2025-10-06 10:09:30.471992464 +0000 UTC m=+679.836415325" lastFinishedPulling="2025-10-06 10:09:33.719688766 +0000 UTC m=+683.084111627" observedRunningTime="2025-10-06 10:09:35.108793261 +0000 UTC m=+684.473216122" watchObservedRunningTime="2025-10-06 10:09:35.113909209 +0000 UTC m=+684.478332070" Oct 06 10:09:36 crc kubenswrapper[4824]: I1006 10:09:36.097214 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-6vncz" event={"ID":"355b7d6a-3ab4-4d16-8923-f15646b3af54","Type":"ContainerStarted","Data":"fd3452802df218cd02d98c2772eba46ba33107fda90545000ff66f5ec72f3cc4"} Oct 06 10:09:37 crc kubenswrapper[4824]: I1006 10:09:37.154602 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-6vncz" podStartSLOduration=2.217387949 podStartE2EDuration="7.154558544s" podCreationTimestamp="2025-10-06 10:09:30 +0000 UTC" firstStartedPulling="2025-10-06 10:09:30.985558161 +0000 UTC m=+680.349981022" lastFinishedPulling="2025-10-06 10:09:35.922728716 +0000 UTC m=+685.287151617" observedRunningTime="2025-10-06 10:09:37.143335333 +0000 UTC m=+686.507758274" watchObservedRunningTime="2025-10-06 10:09:37.154558544 +0000 UTC m=+686.518981475" Oct 06 10:09:40 crc kubenswrapper[4824]: I1006 10:09:40.465523 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-jbv5q" Oct 06 10:09:40 crc kubenswrapper[4824]: I1006 10:09:40.859361 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-5976bc9cd4-ltdb2" Oct 06 10:09:40 crc kubenswrapper[4824]: I1006 10:09:40.859506 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-5976bc9cd4-ltdb2" Oct 06 10:09:40 crc kubenswrapper[4824]: I1006 10:09:40.868649 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-5976bc9cd4-ltdb2" Oct 06 10:09:41 crc kubenswrapper[4824]: I1006 10:09:41.157396 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-5976bc9cd4-ltdb2" Oct 06 10:09:41 crc kubenswrapper[4824]: I1006 10:09:41.228708 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-fpxws"] Oct 06 10:09:50 crc kubenswrapper[4824]: I1006 10:09:50.413233 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-b7dd2" Oct 06 10:10:06 crc kubenswrapper[4824]: I1006 10:10:06.297489 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-fpxws" podUID="722b7a46-52a1-49a7-a053-803334dc5964" containerName="console" containerID="cri-o://5a683afc848a67c0e8bf3bf69fd9a0c4ede0a48fc5f7c8ef2264f4232013651b" gracePeriod=15 Oct 06 10:10:06 crc kubenswrapper[4824]: I1006 10:10:06.767751 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-fpxws_722b7a46-52a1-49a7-a053-803334dc5964/console/0.log" Oct 06 10:10:06 crc kubenswrapper[4824]: I1006 10:10:06.768303 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-fpxws" Oct 06 10:10:06 crc kubenswrapper[4824]: I1006 10:10:06.905467 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/722b7a46-52a1-49a7-a053-803334dc5964-trusted-ca-bundle\") pod \"722b7a46-52a1-49a7-a053-803334dc5964\" (UID: \"722b7a46-52a1-49a7-a053-803334dc5964\") " Oct 06 10:10:06 crc kubenswrapper[4824]: I1006 10:10:06.905822 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/722b7a46-52a1-49a7-a053-803334dc5964-console-oauth-config\") pod \"722b7a46-52a1-49a7-a053-803334dc5964\" (UID: \"722b7a46-52a1-49a7-a053-803334dc5964\") " Oct 06 10:10:06 crc kubenswrapper[4824]: I1006 10:10:06.905902 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/722b7a46-52a1-49a7-a053-803334dc5964-oauth-serving-cert\") pod \"722b7a46-52a1-49a7-a053-803334dc5964\" (UID: \"722b7a46-52a1-49a7-a053-803334dc5964\") " Oct 06 10:10:06 crc kubenswrapper[4824]: I1006 10:10:06.906007 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/722b7a46-52a1-49a7-a053-803334dc5964-service-ca\") pod \"722b7a46-52a1-49a7-a053-803334dc5964\" (UID: \"722b7a46-52a1-49a7-a053-803334dc5964\") " Oct 06 10:10:06 crc kubenswrapper[4824]: I1006 10:10:06.906043 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/722b7a46-52a1-49a7-a053-803334dc5964-console-config\") pod \"722b7a46-52a1-49a7-a053-803334dc5964\" (UID: \"722b7a46-52a1-49a7-a053-803334dc5964\") " Oct 06 10:10:06 crc kubenswrapper[4824]: I1006 10:10:06.906110 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/722b7a46-52a1-49a7-a053-803334dc5964-console-serving-cert\") pod \"722b7a46-52a1-49a7-a053-803334dc5964\" (UID: \"722b7a46-52a1-49a7-a053-803334dc5964\") " Oct 06 10:10:06 crc kubenswrapper[4824]: I1006 10:10:06.906150 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nqdpp\" (UniqueName: \"kubernetes.io/projected/722b7a46-52a1-49a7-a053-803334dc5964-kube-api-access-nqdpp\") pod \"722b7a46-52a1-49a7-a053-803334dc5964\" (UID: \"722b7a46-52a1-49a7-a053-803334dc5964\") " Oct 06 10:10:06 crc kubenswrapper[4824]: I1006 10:10:06.907007 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/722b7a46-52a1-49a7-a053-803334dc5964-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "722b7a46-52a1-49a7-a053-803334dc5964" (UID: "722b7a46-52a1-49a7-a053-803334dc5964"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:10:06 crc kubenswrapper[4824]: I1006 10:10:06.908328 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/722b7a46-52a1-49a7-a053-803334dc5964-service-ca" (OuterVolumeSpecName: "service-ca") pod "722b7a46-52a1-49a7-a053-803334dc5964" (UID: "722b7a46-52a1-49a7-a053-803334dc5964"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:10:06 crc kubenswrapper[4824]: I1006 10:10:06.908501 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/722b7a46-52a1-49a7-a053-803334dc5964-console-config" (OuterVolumeSpecName: "console-config") pod "722b7a46-52a1-49a7-a053-803334dc5964" (UID: "722b7a46-52a1-49a7-a053-803334dc5964"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:10:06 crc kubenswrapper[4824]: I1006 10:10:06.908874 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/722b7a46-52a1-49a7-a053-803334dc5964-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "722b7a46-52a1-49a7-a053-803334dc5964" (UID: "722b7a46-52a1-49a7-a053-803334dc5964"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:10:06 crc kubenswrapper[4824]: I1006 10:10:06.915637 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/722b7a46-52a1-49a7-a053-803334dc5964-kube-api-access-nqdpp" (OuterVolumeSpecName: "kube-api-access-nqdpp") pod "722b7a46-52a1-49a7-a053-803334dc5964" (UID: "722b7a46-52a1-49a7-a053-803334dc5964"). InnerVolumeSpecName "kube-api-access-nqdpp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:10:06 crc kubenswrapper[4824]: I1006 10:10:06.917637 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/722b7a46-52a1-49a7-a053-803334dc5964-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "722b7a46-52a1-49a7-a053-803334dc5964" (UID: "722b7a46-52a1-49a7-a053-803334dc5964"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:10:06 crc kubenswrapper[4824]: I1006 10:10:06.917822 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/722b7a46-52a1-49a7-a053-803334dc5964-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "722b7a46-52a1-49a7-a053-803334dc5964" (UID: "722b7a46-52a1-49a7-a053-803334dc5964"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:10:07 crc kubenswrapper[4824]: I1006 10:10:07.008635 4824 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/722b7a46-52a1-49a7-a053-803334dc5964-console-oauth-config\") on node \"crc\" DevicePath \"\"" Oct 06 10:10:07 crc kubenswrapper[4824]: I1006 10:10:07.008715 4824 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/722b7a46-52a1-49a7-a053-803334dc5964-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 06 10:10:07 crc kubenswrapper[4824]: I1006 10:10:07.008742 4824 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/722b7a46-52a1-49a7-a053-803334dc5964-service-ca\") on node \"crc\" DevicePath \"\"" Oct 06 10:10:07 crc kubenswrapper[4824]: I1006 10:10:07.008763 4824 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/722b7a46-52a1-49a7-a053-803334dc5964-console-config\") on node \"crc\" DevicePath \"\"" Oct 06 10:10:07 crc kubenswrapper[4824]: I1006 10:10:07.008793 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nqdpp\" (UniqueName: \"kubernetes.io/projected/722b7a46-52a1-49a7-a053-803334dc5964-kube-api-access-nqdpp\") on node \"crc\" DevicePath \"\"" Oct 06 10:10:07 crc kubenswrapper[4824]: I1006 10:10:07.008819 4824 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/722b7a46-52a1-49a7-a053-803334dc5964-console-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 06 10:10:07 crc kubenswrapper[4824]: I1006 10:10:07.008839 4824 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/722b7a46-52a1-49a7-a053-803334dc5964-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 10:10:07 crc kubenswrapper[4824]: I1006 10:10:07.376172 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-fpxws_722b7a46-52a1-49a7-a053-803334dc5964/console/0.log" Oct 06 10:10:07 crc kubenswrapper[4824]: I1006 10:10:07.376931 4824 generic.go:334] "Generic (PLEG): container finished" podID="722b7a46-52a1-49a7-a053-803334dc5964" containerID="5a683afc848a67c0e8bf3bf69fd9a0c4ede0a48fc5f7c8ef2264f4232013651b" exitCode=2 Oct 06 10:10:07 crc kubenswrapper[4824]: I1006 10:10:07.377031 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-fpxws" event={"ID":"722b7a46-52a1-49a7-a053-803334dc5964","Type":"ContainerDied","Data":"5a683afc848a67c0e8bf3bf69fd9a0c4ede0a48fc5f7c8ef2264f4232013651b"} Oct 06 10:10:07 crc kubenswrapper[4824]: I1006 10:10:07.377099 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-fpxws" event={"ID":"722b7a46-52a1-49a7-a053-803334dc5964","Type":"ContainerDied","Data":"e7f551e87c80776ae6e7ead338ebe026b4ce3179396bfa6a61c62c451c93592c"} Oct 06 10:10:07 crc kubenswrapper[4824]: I1006 10:10:07.377101 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-fpxws" Oct 06 10:10:07 crc kubenswrapper[4824]: I1006 10:10:07.377140 4824 scope.go:117] "RemoveContainer" containerID="5a683afc848a67c0e8bf3bf69fd9a0c4ede0a48fc5f7c8ef2264f4232013651b" Oct 06 10:10:07 crc kubenswrapper[4824]: I1006 10:10:07.405781 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-fpxws"] Oct 06 10:10:07 crc kubenswrapper[4824]: I1006 10:10:07.411260 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-fpxws"] Oct 06 10:10:07 crc kubenswrapper[4824]: I1006 10:10:07.417334 4824 scope.go:117] "RemoveContainer" containerID="5a683afc848a67c0e8bf3bf69fd9a0c4ede0a48fc5f7c8ef2264f4232013651b" Oct 06 10:10:07 crc kubenswrapper[4824]: E1006 10:10:07.418146 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5a683afc848a67c0e8bf3bf69fd9a0c4ede0a48fc5f7c8ef2264f4232013651b\": container with ID starting with 5a683afc848a67c0e8bf3bf69fd9a0c4ede0a48fc5f7c8ef2264f4232013651b not found: ID does not exist" containerID="5a683afc848a67c0e8bf3bf69fd9a0c4ede0a48fc5f7c8ef2264f4232013651b" Oct 06 10:10:07 crc kubenswrapper[4824]: I1006 10:10:07.418315 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5a683afc848a67c0e8bf3bf69fd9a0c4ede0a48fc5f7c8ef2264f4232013651b"} err="failed to get container status \"5a683afc848a67c0e8bf3bf69fd9a0c4ede0a48fc5f7c8ef2264f4232013651b\": rpc error: code = NotFound desc = could not find container \"5a683afc848a67c0e8bf3bf69fd9a0c4ede0a48fc5f7c8ef2264f4232013651b\": container with ID starting with 5a683afc848a67c0e8bf3bf69fd9a0c4ede0a48fc5f7c8ef2264f4232013651b not found: ID does not exist" Oct 06 10:10:09 crc kubenswrapper[4824]: I1006 10:10:09.288197 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="722b7a46-52a1-49a7-a053-803334dc5964" path="/var/lib/kubelet/pods/722b7a46-52a1-49a7-a053-803334dc5964/volumes" Oct 06 10:10:10 crc kubenswrapper[4824]: I1006 10:10:10.351829 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2frg7w"] Oct 06 10:10:10 crc kubenswrapper[4824]: E1006 10:10:10.352327 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="722b7a46-52a1-49a7-a053-803334dc5964" containerName="console" Oct 06 10:10:10 crc kubenswrapper[4824]: I1006 10:10:10.352345 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="722b7a46-52a1-49a7-a053-803334dc5964" containerName="console" Oct 06 10:10:10 crc kubenswrapper[4824]: I1006 10:10:10.352799 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="722b7a46-52a1-49a7-a053-803334dc5964" containerName="console" Oct 06 10:10:10 crc kubenswrapper[4824]: I1006 10:10:10.360923 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2frg7w" Oct 06 10:10:10 crc kubenswrapper[4824]: I1006 10:10:10.364396 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 06 10:10:10 crc kubenswrapper[4824]: I1006 10:10:10.380720 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2frg7w"] Oct 06 10:10:10 crc kubenswrapper[4824]: I1006 10:10:10.477259 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gnp97\" (UniqueName: \"kubernetes.io/projected/b186e32b-a095-4c20-816c-9e83ea7e261f-kube-api-access-gnp97\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2frg7w\" (UID: \"b186e32b-a095-4c20-816c-9e83ea7e261f\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2frg7w" Oct 06 10:10:10 crc kubenswrapper[4824]: I1006 10:10:10.477330 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b186e32b-a095-4c20-816c-9e83ea7e261f-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2frg7w\" (UID: \"b186e32b-a095-4c20-816c-9e83ea7e261f\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2frg7w" Oct 06 10:10:10 crc kubenswrapper[4824]: I1006 10:10:10.477688 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b186e32b-a095-4c20-816c-9e83ea7e261f-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2frg7w\" (UID: \"b186e32b-a095-4c20-816c-9e83ea7e261f\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2frg7w" Oct 06 10:10:10 crc kubenswrapper[4824]: I1006 10:10:10.579352 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b186e32b-a095-4c20-816c-9e83ea7e261f-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2frg7w\" (UID: \"b186e32b-a095-4c20-816c-9e83ea7e261f\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2frg7w" Oct 06 10:10:10 crc kubenswrapper[4824]: I1006 10:10:10.579511 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gnp97\" (UniqueName: \"kubernetes.io/projected/b186e32b-a095-4c20-816c-9e83ea7e261f-kube-api-access-gnp97\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2frg7w\" (UID: \"b186e32b-a095-4c20-816c-9e83ea7e261f\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2frg7w" Oct 06 10:10:10 crc kubenswrapper[4824]: I1006 10:10:10.579560 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b186e32b-a095-4c20-816c-9e83ea7e261f-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2frg7w\" (UID: \"b186e32b-a095-4c20-816c-9e83ea7e261f\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2frg7w" Oct 06 10:10:10 crc kubenswrapper[4824]: I1006 10:10:10.580416 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b186e32b-a095-4c20-816c-9e83ea7e261f-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2frg7w\" (UID: \"b186e32b-a095-4c20-816c-9e83ea7e261f\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2frg7w" Oct 06 10:10:10 crc kubenswrapper[4824]: I1006 10:10:10.580566 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b186e32b-a095-4c20-816c-9e83ea7e261f-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2frg7w\" (UID: \"b186e32b-a095-4c20-816c-9e83ea7e261f\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2frg7w" Oct 06 10:10:10 crc kubenswrapper[4824]: I1006 10:10:10.620231 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gnp97\" (UniqueName: \"kubernetes.io/projected/b186e32b-a095-4c20-816c-9e83ea7e261f-kube-api-access-gnp97\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2frg7w\" (UID: \"b186e32b-a095-4c20-816c-9e83ea7e261f\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2frg7w" Oct 06 10:10:10 crc kubenswrapper[4824]: I1006 10:10:10.695164 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2frg7w" Oct 06 10:10:11 crc kubenswrapper[4824]: I1006 10:10:11.207879 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2frg7w"] Oct 06 10:10:11 crc kubenswrapper[4824]: I1006 10:10:11.409770 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2frg7w" event={"ID":"b186e32b-a095-4c20-816c-9e83ea7e261f","Type":"ContainerStarted","Data":"dfeb24790e992c2dbb4c9b00c54980f1a35ac5fdfc2f1a08fe276bbca8d4ad13"} Oct 06 10:10:11 crc kubenswrapper[4824]: I1006 10:10:11.409837 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2frg7w" event={"ID":"b186e32b-a095-4c20-816c-9e83ea7e261f","Type":"ContainerStarted","Data":"efc68a578ae287483eca6c14280ab1183571702407207cd4776365589a238e6b"} Oct 06 10:10:11 crc kubenswrapper[4824]: E1006 10:10:11.589405 4824 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb186e32b_a095_4c20_816c_9e83ea7e261f.slice/crio-dfeb24790e992c2dbb4c9b00c54980f1a35ac5fdfc2f1a08fe276bbca8d4ad13.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb186e32b_a095_4c20_816c_9e83ea7e261f.slice/crio-conmon-dfeb24790e992c2dbb4c9b00c54980f1a35ac5fdfc2f1a08fe276bbca8d4ad13.scope\": RecentStats: unable to find data in memory cache]" Oct 06 10:10:12 crc kubenswrapper[4824]: I1006 10:10:12.421926 4824 generic.go:334] "Generic (PLEG): container finished" podID="b186e32b-a095-4c20-816c-9e83ea7e261f" containerID="dfeb24790e992c2dbb4c9b00c54980f1a35ac5fdfc2f1a08fe276bbca8d4ad13" exitCode=0 Oct 06 10:10:12 crc kubenswrapper[4824]: I1006 10:10:12.422098 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2frg7w" event={"ID":"b186e32b-a095-4c20-816c-9e83ea7e261f","Type":"ContainerDied","Data":"dfeb24790e992c2dbb4c9b00c54980f1a35ac5fdfc2f1a08fe276bbca8d4ad13"} Oct 06 10:10:14 crc kubenswrapper[4824]: I1006 10:10:14.444482 4824 generic.go:334] "Generic (PLEG): container finished" podID="b186e32b-a095-4c20-816c-9e83ea7e261f" containerID="22de8daaa34f5b3e5d7bd93d0f25d434d6f31d65598a2c946e72340d9841504e" exitCode=0 Oct 06 10:10:14 crc kubenswrapper[4824]: I1006 10:10:14.444587 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2frg7w" event={"ID":"b186e32b-a095-4c20-816c-9e83ea7e261f","Type":"ContainerDied","Data":"22de8daaa34f5b3e5d7bd93d0f25d434d6f31d65598a2c946e72340d9841504e"} Oct 06 10:10:15 crc kubenswrapper[4824]: I1006 10:10:15.456249 4824 generic.go:334] "Generic (PLEG): container finished" podID="b186e32b-a095-4c20-816c-9e83ea7e261f" containerID="09368bede76fd91c5d8c66f4c8fcebf2572df6e65fd3e565eb3dd2e97b6eb3e6" exitCode=0 Oct 06 10:10:15 crc kubenswrapper[4824]: I1006 10:10:15.456332 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2frg7w" event={"ID":"b186e32b-a095-4c20-816c-9e83ea7e261f","Type":"ContainerDied","Data":"09368bede76fd91c5d8c66f4c8fcebf2572df6e65fd3e565eb3dd2e97b6eb3e6"} Oct 06 10:10:16 crc kubenswrapper[4824]: I1006 10:10:16.794952 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2frg7w" Oct 06 10:10:16 crc kubenswrapper[4824]: I1006 10:10:16.891499 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b186e32b-a095-4c20-816c-9e83ea7e261f-bundle\") pod \"b186e32b-a095-4c20-816c-9e83ea7e261f\" (UID: \"b186e32b-a095-4c20-816c-9e83ea7e261f\") " Oct 06 10:10:16 crc kubenswrapper[4824]: I1006 10:10:16.891788 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gnp97\" (UniqueName: \"kubernetes.io/projected/b186e32b-a095-4c20-816c-9e83ea7e261f-kube-api-access-gnp97\") pod \"b186e32b-a095-4c20-816c-9e83ea7e261f\" (UID: \"b186e32b-a095-4c20-816c-9e83ea7e261f\") " Oct 06 10:10:16 crc kubenswrapper[4824]: I1006 10:10:16.891897 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b186e32b-a095-4c20-816c-9e83ea7e261f-util\") pod \"b186e32b-a095-4c20-816c-9e83ea7e261f\" (UID: \"b186e32b-a095-4c20-816c-9e83ea7e261f\") " Oct 06 10:10:16 crc kubenswrapper[4824]: I1006 10:10:16.892708 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b186e32b-a095-4c20-816c-9e83ea7e261f-bundle" (OuterVolumeSpecName: "bundle") pod "b186e32b-a095-4c20-816c-9e83ea7e261f" (UID: "b186e32b-a095-4c20-816c-9e83ea7e261f"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:10:16 crc kubenswrapper[4824]: I1006 10:10:16.901913 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b186e32b-a095-4c20-816c-9e83ea7e261f-kube-api-access-gnp97" (OuterVolumeSpecName: "kube-api-access-gnp97") pod "b186e32b-a095-4c20-816c-9e83ea7e261f" (UID: "b186e32b-a095-4c20-816c-9e83ea7e261f"). InnerVolumeSpecName "kube-api-access-gnp97". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:10:16 crc kubenswrapper[4824]: I1006 10:10:16.916397 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b186e32b-a095-4c20-816c-9e83ea7e261f-util" (OuterVolumeSpecName: "util") pod "b186e32b-a095-4c20-816c-9e83ea7e261f" (UID: "b186e32b-a095-4c20-816c-9e83ea7e261f"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:10:16 crc kubenswrapper[4824]: I1006 10:10:16.993286 4824 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b186e32b-a095-4c20-816c-9e83ea7e261f-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 10:10:16 crc kubenswrapper[4824]: I1006 10:10:16.993362 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gnp97\" (UniqueName: \"kubernetes.io/projected/b186e32b-a095-4c20-816c-9e83ea7e261f-kube-api-access-gnp97\") on node \"crc\" DevicePath \"\"" Oct 06 10:10:16 crc kubenswrapper[4824]: I1006 10:10:16.993394 4824 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b186e32b-a095-4c20-816c-9e83ea7e261f-util\") on node \"crc\" DevicePath \"\"" Oct 06 10:10:17 crc kubenswrapper[4824]: I1006 10:10:17.473085 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2frg7w" event={"ID":"b186e32b-a095-4c20-816c-9e83ea7e261f","Type":"ContainerDied","Data":"efc68a578ae287483eca6c14280ab1183571702407207cd4776365589a238e6b"} Oct 06 10:10:17 crc kubenswrapper[4824]: I1006 10:10:17.473145 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="efc68a578ae287483eca6c14280ab1183571702407207cd4776365589a238e6b" Oct 06 10:10:17 crc kubenswrapper[4824]: I1006 10:10:17.473196 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2frg7w" Oct 06 10:10:25 crc kubenswrapper[4824]: I1006 10:10:25.703826 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-57744c8fb4-85v7z"] Oct 06 10:10:25 crc kubenswrapper[4824]: E1006 10:10:25.704845 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b186e32b-a095-4c20-816c-9e83ea7e261f" containerName="extract" Oct 06 10:10:25 crc kubenswrapper[4824]: I1006 10:10:25.704865 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="b186e32b-a095-4c20-816c-9e83ea7e261f" containerName="extract" Oct 06 10:10:25 crc kubenswrapper[4824]: E1006 10:10:25.704891 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b186e32b-a095-4c20-816c-9e83ea7e261f" containerName="util" Oct 06 10:10:25 crc kubenswrapper[4824]: I1006 10:10:25.704899 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="b186e32b-a095-4c20-816c-9e83ea7e261f" containerName="util" Oct 06 10:10:25 crc kubenswrapper[4824]: E1006 10:10:25.704911 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b186e32b-a095-4c20-816c-9e83ea7e261f" containerName="pull" Oct 06 10:10:25 crc kubenswrapper[4824]: I1006 10:10:25.704918 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="b186e32b-a095-4c20-816c-9e83ea7e261f" containerName="pull" Oct 06 10:10:25 crc kubenswrapper[4824]: I1006 10:10:25.705090 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="b186e32b-a095-4c20-816c-9e83ea7e261f" containerName="extract" Oct 06 10:10:25 crc kubenswrapper[4824]: I1006 10:10:25.705542 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-57744c8fb4-85v7z" Oct 06 10:10:25 crc kubenswrapper[4824]: I1006 10:10:25.709925 4824 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Oct 06 10:10:25 crc kubenswrapper[4824]: I1006 10:10:25.710219 4824 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Oct 06 10:10:25 crc kubenswrapper[4824]: I1006 10:10:25.710388 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Oct 06 10:10:25 crc kubenswrapper[4824]: I1006 10:10:25.710556 4824 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-lct5t" Oct 06 10:10:25 crc kubenswrapper[4824]: I1006 10:10:25.710750 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Oct 06 10:10:25 crc kubenswrapper[4824]: I1006 10:10:25.721894 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-57744c8fb4-85v7z"] Oct 06 10:10:25 crc kubenswrapper[4824]: I1006 10:10:25.733530 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/d8226542-256f-4e73-9d78-0d8f261108d4-apiservice-cert\") pod \"metallb-operator-controller-manager-57744c8fb4-85v7z\" (UID: \"d8226542-256f-4e73-9d78-0d8f261108d4\") " pod="metallb-system/metallb-operator-controller-manager-57744c8fb4-85v7z" Oct 06 10:10:25 crc kubenswrapper[4824]: I1006 10:10:25.733605 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/d8226542-256f-4e73-9d78-0d8f261108d4-webhook-cert\") pod \"metallb-operator-controller-manager-57744c8fb4-85v7z\" (UID: \"d8226542-256f-4e73-9d78-0d8f261108d4\") " pod="metallb-system/metallb-operator-controller-manager-57744c8fb4-85v7z" Oct 06 10:10:25 crc kubenswrapper[4824]: I1006 10:10:25.733646 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ls5ww\" (UniqueName: \"kubernetes.io/projected/d8226542-256f-4e73-9d78-0d8f261108d4-kube-api-access-ls5ww\") pod \"metallb-operator-controller-manager-57744c8fb4-85v7z\" (UID: \"d8226542-256f-4e73-9d78-0d8f261108d4\") " pod="metallb-system/metallb-operator-controller-manager-57744c8fb4-85v7z" Oct 06 10:10:25 crc kubenswrapper[4824]: I1006 10:10:25.846126 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/d8226542-256f-4e73-9d78-0d8f261108d4-webhook-cert\") pod \"metallb-operator-controller-manager-57744c8fb4-85v7z\" (UID: \"d8226542-256f-4e73-9d78-0d8f261108d4\") " pod="metallb-system/metallb-operator-controller-manager-57744c8fb4-85v7z" Oct 06 10:10:25 crc kubenswrapper[4824]: I1006 10:10:25.846288 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ls5ww\" (UniqueName: \"kubernetes.io/projected/d8226542-256f-4e73-9d78-0d8f261108d4-kube-api-access-ls5ww\") pod \"metallb-operator-controller-manager-57744c8fb4-85v7z\" (UID: \"d8226542-256f-4e73-9d78-0d8f261108d4\") " pod="metallb-system/metallb-operator-controller-manager-57744c8fb4-85v7z" Oct 06 10:10:25 crc kubenswrapper[4824]: I1006 10:10:25.846391 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/d8226542-256f-4e73-9d78-0d8f261108d4-apiservice-cert\") pod \"metallb-operator-controller-manager-57744c8fb4-85v7z\" (UID: \"d8226542-256f-4e73-9d78-0d8f261108d4\") " pod="metallb-system/metallb-operator-controller-manager-57744c8fb4-85v7z" Oct 06 10:10:25 crc kubenswrapper[4824]: I1006 10:10:25.858839 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/d8226542-256f-4e73-9d78-0d8f261108d4-apiservice-cert\") pod \"metallb-operator-controller-manager-57744c8fb4-85v7z\" (UID: \"d8226542-256f-4e73-9d78-0d8f261108d4\") " pod="metallb-system/metallb-operator-controller-manager-57744c8fb4-85v7z" Oct 06 10:10:25 crc kubenswrapper[4824]: I1006 10:10:25.863033 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/d8226542-256f-4e73-9d78-0d8f261108d4-webhook-cert\") pod \"metallb-operator-controller-manager-57744c8fb4-85v7z\" (UID: \"d8226542-256f-4e73-9d78-0d8f261108d4\") " pod="metallb-system/metallb-operator-controller-manager-57744c8fb4-85v7z" Oct 06 10:10:25 crc kubenswrapper[4824]: I1006 10:10:25.870575 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ls5ww\" (UniqueName: \"kubernetes.io/projected/d8226542-256f-4e73-9d78-0d8f261108d4-kube-api-access-ls5ww\") pod \"metallb-operator-controller-manager-57744c8fb4-85v7z\" (UID: \"d8226542-256f-4e73-9d78-0d8f261108d4\") " pod="metallb-system/metallb-operator-controller-manager-57744c8fb4-85v7z" Oct 06 10:10:26 crc kubenswrapper[4824]: I1006 10:10:26.027285 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-57744c8fb4-85v7z" Oct 06 10:10:26 crc kubenswrapper[4824]: I1006 10:10:26.101417 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-56788496c6-xdgkc"] Oct 06 10:10:26 crc kubenswrapper[4824]: I1006 10:10:26.102165 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-56788496c6-xdgkc" Oct 06 10:10:26 crc kubenswrapper[4824]: I1006 10:10:26.108472 4824 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Oct 06 10:10:26 crc kubenswrapper[4824]: I1006 10:10:26.108518 4824 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-f5f4k" Oct 06 10:10:26 crc kubenswrapper[4824]: I1006 10:10:26.117240 4824 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Oct 06 10:10:26 crc kubenswrapper[4824]: I1006 10:10:26.142602 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-56788496c6-xdgkc"] Oct 06 10:10:26 crc kubenswrapper[4824]: I1006 10:10:26.152690 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/1d2065c1-e8a8-4dfd-bf6e-a8f701cbe060-webhook-cert\") pod \"metallb-operator-webhook-server-56788496c6-xdgkc\" (UID: \"1d2065c1-e8a8-4dfd-bf6e-a8f701cbe060\") " pod="metallb-system/metallb-operator-webhook-server-56788496c6-xdgkc" Oct 06 10:10:26 crc kubenswrapper[4824]: I1006 10:10:26.152741 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/1d2065c1-e8a8-4dfd-bf6e-a8f701cbe060-apiservice-cert\") pod \"metallb-operator-webhook-server-56788496c6-xdgkc\" (UID: \"1d2065c1-e8a8-4dfd-bf6e-a8f701cbe060\") " pod="metallb-system/metallb-operator-webhook-server-56788496c6-xdgkc" Oct 06 10:10:26 crc kubenswrapper[4824]: I1006 10:10:26.152777 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lzd9p\" (UniqueName: \"kubernetes.io/projected/1d2065c1-e8a8-4dfd-bf6e-a8f701cbe060-kube-api-access-lzd9p\") pod \"metallb-operator-webhook-server-56788496c6-xdgkc\" (UID: \"1d2065c1-e8a8-4dfd-bf6e-a8f701cbe060\") " pod="metallb-system/metallb-operator-webhook-server-56788496c6-xdgkc" Oct 06 10:10:26 crc kubenswrapper[4824]: I1006 10:10:26.253815 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lzd9p\" (UniqueName: \"kubernetes.io/projected/1d2065c1-e8a8-4dfd-bf6e-a8f701cbe060-kube-api-access-lzd9p\") pod \"metallb-operator-webhook-server-56788496c6-xdgkc\" (UID: \"1d2065c1-e8a8-4dfd-bf6e-a8f701cbe060\") " pod="metallb-system/metallb-operator-webhook-server-56788496c6-xdgkc" Oct 06 10:10:26 crc kubenswrapper[4824]: I1006 10:10:26.253915 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/1d2065c1-e8a8-4dfd-bf6e-a8f701cbe060-webhook-cert\") pod \"metallb-operator-webhook-server-56788496c6-xdgkc\" (UID: \"1d2065c1-e8a8-4dfd-bf6e-a8f701cbe060\") " pod="metallb-system/metallb-operator-webhook-server-56788496c6-xdgkc" Oct 06 10:10:26 crc kubenswrapper[4824]: I1006 10:10:26.253944 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/1d2065c1-e8a8-4dfd-bf6e-a8f701cbe060-apiservice-cert\") pod \"metallb-operator-webhook-server-56788496c6-xdgkc\" (UID: \"1d2065c1-e8a8-4dfd-bf6e-a8f701cbe060\") " pod="metallb-system/metallb-operator-webhook-server-56788496c6-xdgkc" Oct 06 10:10:26 crc kubenswrapper[4824]: I1006 10:10:26.258752 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/1d2065c1-e8a8-4dfd-bf6e-a8f701cbe060-webhook-cert\") pod \"metallb-operator-webhook-server-56788496c6-xdgkc\" (UID: \"1d2065c1-e8a8-4dfd-bf6e-a8f701cbe060\") " pod="metallb-system/metallb-operator-webhook-server-56788496c6-xdgkc" Oct 06 10:10:26 crc kubenswrapper[4824]: I1006 10:10:26.262112 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/1d2065c1-e8a8-4dfd-bf6e-a8f701cbe060-apiservice-cert\") pod \"metallb-operator-webhook-server-56788496c6-xdgkc\" (UID: \"1d2065c1-e8a8-4dfd-bf6e-a8f701cbe060\") " pod="metallb-system/metallb-operator-webhook-server-56788496c6-xdgkc" Oct 06 10:10:26 crc kubenswrapper[4824]: I1006 10:10:26.285008 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lzd9p\" (UniqueName: \"kubernetes.io/projected/1d2065c1-e8a8-4dfd-bf6e-a8f701cbe060-kube-api-access-lzd9p\") pod \"metallb-operator-webhook-server-56788496c6-xdgkc\" (UID: \"1d2065c1-e8a8-4dfd-bf6e-a8f701cbe060\") " pod="metallb-system/metallb-operator-webhook-server-56788496c6-xdgkc" Oct 06 10:10:26 crc kubenswrapper[4824]: I1006 10:10:26.427462 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-56788496c6-xdgkc" Oct 06 10:10:26 crc kubenswrapper[4824]: I1006 10:10:26.429144 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-57744c8fb4-85v7z"] Oct 06 10:10:26 crc kubenswrapper[4824]: I1006 10:10:26.549080 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-57744c8fb4-85v7z" event={"ID":"d8226542-256f-4e73-9d78-0d8f261108d4","Type":"ContainerStarted","Data":"9628c92f9efd870ffd8d474a221907e6f38bb2939a627828a66fcde32d9e9c63"} Oct 06 10:10:26 crc kubenswrapper[4824]: I1006 10:10:26.915247 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-56788496c6-xdgkc"] Oct 06 10:10:26 crc kubenswrapper[4824]: W1006 10:10:26.928567 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1d2065c1_e8a8_4dfd_bf6e_a8f701cbe060.slice/crio-28b7862689ba143b6979ab211cea214014ec3770f24ec41ec6d3b6f5a9063c2d WatchSource:0}: Error finding container 28b7862689ba143b6979ab211cea214014ec3770f24ec41ec6d3b6f5a9063c2d: Status 404 returned error can't find the container with id 28b7862689ba143b6979ab211cea214014ec3770f24ec41ec6d3b6f5a9063c2d Oct 06 10:10:27 crc kubenswrapper[4824]: I1006 10:10:27.555816 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-56788496c6-xdgkc" event={"ID":"1d2065c1-e8a8-4dfd-bf6e-a8f701cbe060","Type":"ContainerStarted","Data":"28b7862689ba143b6979ab211cea214014ec3770f24ec41ec6d3b6f5a9063c2d"} Oct 06 10:10:30 crc kubenswrapper[4824]: I1006 10:10:30.586028 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-57744c8fb4-85v7z" event={"ID":"d8226542-256f-4e73-9d78-0d8f261108d4","Type":"ContainerStarted","Data":"80a474871e6337432af19649f4c4d872068f0d1be97599f0f1dc380c5fbec6c3"} Oct 06 10:10:30 crc kubenswrapper[4824]: I1006 10:10:30.586456 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-57744c8fb4-85v7z" Oct 06 10:10:30 crc kubenswrapper[4824]: I1006 10:10:30.640242 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-57744c8fb4-85v7z" podStartSLOduration=2.504543552 podStartE2EDuration="5.64021976s" podCreationTimestamp="2025-10-06 10:10:25 +0000 UTC" firstStartedPulling="2025-10-06 10:10:26.455541749 +0000 UTC m=+735.819964610" lastFinishedPulling="2025-10-06 10:10:29.591217937 +0000 UTC m=+738.955640818" observedRunningTime="2025-10-06 10:10:30.638294602 +0000 UTC m=+740.002717453" watchObservedRunningTime="2025-10-06 10:10:30.64021976 +0000 UTC m=+740.004642621" Oct 06 10:10:32 crc kubenswrapper[4824]: I1006 10:10:32.601298 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-56788496c6-xdgkc" event={"ID":"1d2065c1-e8a8-4dfd-bf6e-a8f701cbe060","Type":"ContainerStarted","Data":"691e2663f3e56ad9c925dbd205db2a17afc7fa0968cd9348d6f3daa6b5462a0a"} Oct 06 10:10:32 crc kubenswrapper[4824]: I1006 10:10:32.602968 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-56788496c6-xdgkc" Oct 06 10:10:32 crc kubenswrapper[4824]: I1006 10:10:32.632336 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-56788496c6-xdgkc" podStartSLOduration=1.84549213 podStartE2EDuration="6.632315227s" podCreationTimestamp="2025-10-06 10:10:26 +0000 UTC" firstStartedPulling="2025-10-06 10:10:26.935882981 +0000 UTC m=+736.300305842" lastFinishedPulling="2025-10-06 10:10:31.722706078 +0000 UTC m=+741.087128939" observedRunningTime="2025-10-06 10:10:32.627397204 +0000 UTC m=+741.991820075" watchObservedRunningTime="2025-10-06 10:10:32.632315227 +0000 UTC m=+741.996738098" Oct 06 10:10:46 crc kubenswrapper[4824]: I1006 10:10:46.450404 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-56788496c6-xdgkc" Oct 06 10:10:50 crc kubenswrapper[4824]: I1006 10:10:50.532037 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-nqmr6"] Oct 06 10:10:50 crc kubenswrapper[4824]: I1006 10:10:50.532574 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-nqmr6" podUID="b6f7d3cb-418e-4f73-95a0-2477e37976c3" containerName="controller-manager" containerID="cri-o://610b272f4cb7b19cc0d283e6e5d3a4489f50cd7ffafc76e70072bfa5b0da0396" gracePeriod=30 Oct 06 10:10:50 crc kubenswrapper[4824]: I1006 10:10:50.566226 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-wk4bd"] Oct 06 10:10:50 crc kubenswrapper[4824]: I1006 10:10:50.566491 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wk4bd" podUID="dd50f0ac-ac97-413f-a350-7496681f7f0a" containerName="route-controller-manager" containerID="cri-o://37781d4ace294c944018c677becc05dc9d3fcc964a5419c612190d8c821c4cb7" gracePeriod=30 Oct 06 10:10:50 crc kubenswrapper[4824]: I1006 10:10:50.735302 4824 generic.go:334] "Generic (PLEG): container finished" podID="b6f7d3cb-418e-4f73-95a0-2477e37976c3" containerID="610b272f4cb7b19cc0d283e6e5d3a4489f50cd7ffafc76e70072bfa5b0da0396" exitCode=0 Oct 06 10:10:50 crc kubenswrapper[4824]: I1006 10:10:50.735384 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-nqmr6" event={"ID":"b6f7d3cb-418e-4f73-95a0-2477e37976c3","Type":"ContainerDied","Data":"610b272f4cb7b19cc0d283e6e5d3a4489f50cd7ffafc76e70072bfa5b0da0396"} Oct 06 10:10:50 crc kubenswrapper[4824]: I1006 10:10:50.739074 4824 generic.go:334] "Generic (PLEG): container finished" podID="dd50f0ac-ac97-413f-a350-7496681f7f0a" containerID="37781d4ace294c944018c677becc05dc9d3fcc964a5419c612190d8c821c4cb7" exitCode=0 Oct 06 10:10:50 crc kubenswrapper[4824]: I1006 10:10:50.739119 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wk4bd" event={"ID":"dd50f0ac-ac97-413f-a350-7496681f7f0a","Type":"ContainerDied","Data":"37781d4ace294c944018c677becc05dc9d3fcc964a5419c612190d8c821c4cb7"} Oct 06 10:10:50 crc kubenswrapper[4824]: I1006 10:10:50.967926 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-nqmr6" Oct 06 10:10:51 crc kubenswrapper[4824]: I1006 10:10:51.007312 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wk4bd" Oct 06 10:10:51 crc kubenswrapper[4824]: I1006 10:10:51.137596 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dd50f0ac-ac97-413f-a350-7496681f7f0a-config\") pod \"dd50f0ac-ac97-413f-a350-7496681f7f0a\" (UID: \"dd50f0ac-ac97-413f-a350-7496681f7f0a\") " Oct 06 10:10:51 crc kubenswrapper[4824]: I1006 10:10:51.137694 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2qj4r\" (UniqueName: \"kubernetes.io/projected/dd50f0ac-ac97-413f-a350-7496681f7f0a-kube-api-access-2qj4r\") pod \"dd50f0ac-ac97-413f-a350-7496681f7f0a\" (UID: \"dd50f0ac-ac97-413f-a350-7496681f7f0a\") " Oct 06 10:10:51 crc kubenswrapper[4824]: I1006 10:10:51.137744 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b6f7d3cb-418e-4f73-95a0-2477e37976c3-client-ca\") pod \"b6f7d3cb-418e-4f73-95a0-2477e37976c3\" (UID: \"b6f7d3cb-418e-4f73-95a0-2477e37976c3\") " Oct 06 10:10:51 crc kubenswrapper[4824]: I1006 10:10:51.137798 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b6f7d3cb-418e-4f73-95a0-2477e37976c3-config\") pod \"b6f7d3cb-418e-4f73-95a0-2477e37976c3\" (UID: \"b6f7d3cb-418e-4f73-95a0-2477e37976c3\") " Oct 06 10:10:51 crc kubenswrapper[4824]: I1006 10:10:51.137836 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b6f7d3cb-418e-4f73-95a0-2477e37976c3-serving-cert\") pod \"b6f7d3cb-418e-4f73-95a0-2477e37976c3\" (UID: \"b6f7d3cb-418e-4f73-95a0-2477e37976c3\") " Oct 06 10:10:51 crc kubenswrapper[4824]: I1006 10:10:51.137876 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b6f7d3cb-418e-4f73-95a0-2477e37976c3-proxy-ca-bundles\") pod \"b6f7d3cb-418e-4f73-95a0-2477e37976c3\" (UID: \"b6f7d3cb-418e-4f73-95a0-2477e37976c3\") " Oct 06 10:10:51 crc kubenswrapper[4824]: I1006 10:10:51.137988 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dd50f0ac-ac97-413f-a350-7496681f7f0a-serving-cert\") pod \"dd50f0ac-ac97-413f-a350-7496681f7f0a\" (UID: \"dd50f0ac-ac97-413f-a350-7496681f7f0a\") " Oct 06 10:10:51 crc kubenswrapper[4824]: I1006 10:10:51.138051 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2h2fc\" (UniqueName: \"kubernetes.io/projected/b6f7d3cb-418e-4f73-95a0-2477e37976c3-kube-api-access-2h2fc\") pod \"b6f7d3cb-418e-4f73-95a0-2477e37976c3\" (UID: \"b6f7d3cb-418e-4f73-95a0-2477e37976c3\") " Oct 06 10:10:51 crc kubenswrapper[4824]: I1006 10:10:51.138093 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/dd50f0ac-ac97-413f-a350-7496681f7f0a-client-ca\") pod \"dd50f0ac-ac97-413f-a350-7496681f7f0a\" (UID: \"dd50f0ac-ac97-413f-a350-7496681f7f0a\") " Oct 06 10:10:51 crc kubenswrapper[4824]: I1006 10:10:51.138590 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6f7d3cb-418e-4f73-95a0-2477e37976c3-client-ca" (OuterVolumeSpecName: "client-ca") pod "b6f7d3cb-418e-4f73-95a0-2477e37976c3" (UID: "b6f7d3cb-418e-4f73-95a0-2477e37976c3"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:10:51 crc kubenswrapper[4824]: I1006 10:10:51.138724 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dd50f0ac-ac97-413f-a350-7496681f7f0a-config" (OuterVolumeSpecName: "config") pod "dd50f0ac-ac97-413f-a350-7496681f7f0a" (UID: "dd50f0ac-ac97-413f-a350-7496681f7f0a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:10:51 crc kubenswrapper[4824]: I1006 10:10:51.138802 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6f7d3cb-418e-4f73-95a0-2477e37976c3-config" (OuterVolumeSpecName: "config") pod "b6f7d3cb-418e-4f73-95a0-2477e37976c3" (UID: "b6f7d3cb-418e-4f73-95a0-2477e37976c3"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:10:51 crc kubenswrapper[4824]: I1006 10:10:51.138803 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6f7d3cb-418e-4f73-95a0-2477e37976c3-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "b6f7d3cb-418e-4f73-95a0-2477e37976c3" (UID: "b6f7d3cb-418e-4f73-95a0-2477e37976c3"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:10:51 crc kubenswrapper[4824]: I1006 10:10:51.139055 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dd50f0ac-ac97-413f-a350-7496681f7f0a-client-ca" (OuterVolumeSpecName: "client-ca") pod "dd50f0ac-ac97-413f-a350-7496681f7f0a" (UID: "dd50f0ac-ac97-413f-a350-7496681f7f0a"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:10:51 crc kubenswrapper[4824]: I1006 10:10:51.145702 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6f7d3cb-418e-4f73-95a0-2477e37976c3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "b6f7d3cb-418e-4f73-95a0-2477e37976c3" (UID: "b6f7d3cb-418e-4f73-95a0-2477e37976c3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:10:51 crc kubenswrapper[4824]: I1006 10:10:51.145949 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dd50f0ac-ac97-413f-a350-7496681f7f0a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "dd50f0ac-ac97-413f-a350-7496681f7f0a" (UID: "dd50f0ac-ac97-413f-a350-7496681f7f0a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:10:51 crc kubenswrapper[4824]: I1006 10:10:51.146035 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6f7d3cb-418e-4f73-95a0-2477e37976c3-kube-api-access-2h2fc" (OuterVolumeSpecName: "kube-api-access-2h2fc") pod "b6f7d3cb-418e-4f73-95a0-2477e37976c3" (UID: "b6f7d3cb-418e-4f73-95a0-2477e37976c3"). InnerVolumeSpecName "kube-api-access-2h2fc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:10:51 crc kubenswrapper[4824]: I1006 10:10:51.146106 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dd50f0ac-ac97-413f-a350-7496681f7f0a-kube-api-access-2qj4r" (OuterVolumeSpecName: "kube-api-access-2qj4r") pod "dd50f0ac-ac97-413f-a350-7496681f7f0a" (UID: "dd50f0ac-ac97-413f-a350-7496681f7f0a"). InnerVolumeSpecName "kube-api-access-2qj4r". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:10:51 crc kubenswrapper[4824]: I1006 10:10:51.239541 4824 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b6f7d3cb-418e-4f73-95a0-2477e37976c3-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 06 10:10:51 crc kubenswrapper[4824]: I1006 10:10:51.239584 4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dd50f0ac-ac97-413f-a350-7496681f7f0a-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 06 10:10:51 crc kubenswrapper[4824]: I1006 10:10:51.239594 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2h2fc\" (UniqueName: \"kubernetes.io/projected/b6f7d3cb-418e-4f73-95a0-2477e37976c3-kube-api-access-2h2fc\") on node \"crc\" DevicePath \"\"" Oct 06 10:10:51 crc kubenswrapper[4824]: I1006 10:10:51.239607 4824 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/dd50f0ac-ac97-413f-a350-7496681f7f0a-client-ca\") on node \"crc\" DevicePath \"\"" Oct 06 10:10:51 crc kubenswrapper[4824]: I1006 10:10:51.239617 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dd50f0ac-ac97-413f-a350-7496681f7f0a-config\") on node \"crc\" DevicePath \"\"" Oct 06 10:10:51 crc kubenswrapper[4824]: I1006 10:10:51.239626 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2qj4r\" (UniqueName: \"kubernetes.io/projected/dd50f0ac-ac97-413f-a350-7496681f7f0a-kube-api-access-2qj4r\") on node \"crc\" DevicePath \"\"" Oct 06 10:10:51 crc kubenswrapper[4824]: I1006 10:10:51.239638 4824 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b6f7d3cb-418e-4f73-95a0-2477e37976c3-client-ca\") on node \"crc\" DevicePath \"\"" Oct 06 10:10:51 crc kubenswrapper[4824]: I1006 10:10:51.239647 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b6f7d3cb-418e-4f73-95a0-2477e37976c3-config\") on node \"crc\" DevicePath \"\"" Oct 06 10:10:51 crc kubenswrapper[4824]: I1006 10:10:51.239656 4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b6f7d3cb-418e-4f73-95a0-2477e37976c3-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 06 10:10:51 crc kubenswrapper[4824]: I1006 10:10:51.749184 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-nqmr6" Oct 06 10:10:51 crc kubenswrapper[4824]: I1006 10:10:51.749415 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-nqmr6" event={"ID":"b6f7d3cb-418e-4f73-95a0-2477e37976c3","Type":"ContainerDied","Data":"243447b3c3ad6710011f6daf375f053746d337edca1d1ec2b5c40e13fb029dc6"} Oct 06 10:10:51 crc kubenswrapper[4824]: I1006 10:10:51.749485 4824 scope.go:117] "RemoveContainer" containerID="610b272f4cb7b19cc0d283e6e5d3a4489f50cd7ffafc76e70072bfa5b0da0396" Oct 06 10:10:51 crc kubenswrapper[4824]: I1006 10:10:51.751518 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wk4bd" event={"ID":"dd50f0ac-ac97-413f-a350-7496681f7f0a","Type":"ContainerDied","Data":"caa6556ff759f287324f7bb3ad5df851767a64c5aebfb75b072ba6fdd55b4ef8"} Oct 06 10:10:51 crc kubenswrapper[4824]: I1006 10:10:51.751580 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wk4bd" Oct 06 10:10:51 crc kubenswrapper[4824]: I1006 10:10:51.779896 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-nqmr6"] Oct 06 10:10:51 crc kubenswrapper[4824]: I1006 10:10:51.793584 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-nqmr6"] Oct 06 10:10:51 crc kubenswrapper[4824]: I1006 10:10:51.802321 4824 scope.go:117] "RemoveContainer" containerID="37781d4ace294c944018c677becc05dc9d3fcc964a5419c612190d8c821c4cb7" Oct 06 10:10:51 crc kubenswrapper[4824]: I1006 10:10:51.806327 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-wk4bd"] Oct 06 10:10:51 crc kubenswrapper[4824]: I1006 10:10:51.811947 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-6b4bcb86d5-b57w5"] Oct 06 10:10:51 crc kubenswrapper[4824]: E1006 10:10:51.812651 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd50f0ac-ac97-413f-a350-7496681f7f0a" containerName="route-controller-manager" Oct 06 10:10:51 crc kubenswrapper[4824]: I1006 10:10:51.812686 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd50f0ac-ac97-413f-a350-7496681f7f0a" containerName="route-controller-manager" Oct 06 10:10:51 crc kubenswrapper[4824]: E1006 10:10:51.812731 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6f7d3cb-418e-4f73-95a0-2477e37976c3" containerName="controller-manager" Oct 06 10:10:51 crc kubenswrapper[4824]: I1006 10:10:51.812778 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6f7d3cb-418e-4f73-95a0-2477e37976c3" containerName="controller-manager" Oct 06 10:10:51 crc kubenswrapper[4824]: I1006 10:10:51.813042 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="dd50f0ac-ac97-413f-a350-7496681f7f0a" containerName="route-controller-manager" Oct 06 10:10:51 crc kubenswrapper[4824]: I1006 10:10:51.813072 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="b6f7d3cb-418e-4f73-95a0-2477e37976c3" containerName="controller-manager" Oct 06 10:10:51 crc kubenswrapper[4824]: I1006 10:10:51.813746 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6b4bcb86d5-b57w5" Oct 06 10:10:51 crc kubenswrapper[4824]: I1006 10:10:51.817022 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Oct 06 10:10:51 crc kubenswrapper[4824]: I1006 10:10:51.820319 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Oct 06 10:10:51 crc kubenswrapper[4824]: I1006 10:10:51.820631 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Oct 06 10:10:51 crc kubenswrapper[4824]: I1006 10:10:51.821020 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Oct 06 10:10:51 crc kubenswrapper[4824]: I1006 10:10:51.821528 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Oct 06 10:10:51 crc kubenswrapper[4824]: I1006 10:10:51.821656 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6d646b97ff-jhtrw"] Oct 06 10:10:51 crc kubenswrapper[4824]: I1006 10:10:51.822627 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6d646b97ff-jhtrw" Oct 06 10:10:51 crc kubenswrapper[4824]: I1006 10:10:51.822686 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Oct 06 10:10:51 crc kubenswrapper[4824]: I1006 10:10:51.825156 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Oct 06 10:10:51 crc kubenswrapper[4824]: I1006 10:10:51.825710 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Oct 06 10:10:51 crc kubenswrapper[4824]: I1006 10:10:51.826292 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Oct 06 10:10:51 crc kubenswrapper[4824]: I1006 10:10:51.826771 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Oct 06 10:10:51 crc kubenswrapper[4824]: I1006 10:10:51.826856 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Oct 06 10:10:51 crc kubenswrapper[4824]: I1006 10:10:51.828644 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Oct 06 10:10:51 crc kubenswrapper[4824]: I1006 10:10:51.828934 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Oct 06 10:10:51 crc kubenswrapper[4824]: I1006 10:10:51.838674 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-wk4bd"] Oct 06 10:10:51 crc kubenswrapper[4824]: I1006 10:10:51.838763 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6b4bcb86d5-b57w5"] Oct 06 10:10:51 crc kubenswrapper[4824]: I1006 10:10:51.846889 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6d646b97ff-jhtrw"] Oct 06 10:10:51 crc kubenswrapper[4824]: I1006 10:10:51.948801 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qvhtz\" (UniqueName: \"kubernetes.io/projected/ab3e2669-11e7-4a27-89c3-4983798a8997-kube-api-access-qvhtz\") pod \"route-controller-manager-6d646b97ff-jhtrw\" (UID: \"ab3e2669-11e7-4a27-89c3-4983798a8997\") " pod="openshift-route-controller-manager/route-controller-manager-6d646b97ff-jhtrw" Oct 06 10:10:51 crc kubenswrapper[4824]: I1006 10:10:51.948858 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/cc2c5d65-334e-4005-a7a9-1c1ab5f4ad02-client-ca\") pod \"controller-manager-6b4bcb86d5-b57w5\" (UID: \"cc2c5d65-334e-4005-a7a9-1c1ab5f4ad02\") " pod="openshift-controller-manager/controller-manager-6b4bcb86d5-b57w5" Oct 06 10:10:51 crc kubenswrapper[4824]: I1006 10:10:51.948888 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/cc2c5d65-334e-4005-a7a9-1c1ab5f4ad02-proxy-ca-bundles\") pod \"controller-manager-6b4bcb86d5-b57w5\" (UID: \"cc2c5d65-334e-4005-a7a9-1c1ab5f4ad02\") " pod="openshift-controller-manager/controller-manager-6b4bcb86d5-b57w5" Oct 06 10:10:51 crc kubenswrapper[4824]: I1006 10:10:51.949472 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ab3e2669-11e7-4a27-89c3-4983798a8997-serving-cert\") pod \"route-controller-manager-6d646b97ff-jhtrw\" (UID: \"ab3e2669-11e7-4a27-89c3-4983798a8997\") " pod="openshift-route-controller-manager/route-controller-manager-6d646b97ff-jhtrw" Oct 06 10:10:51 crc kubenswrapper[4824]: I1006 10:10:51.949508 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cc2c5d65-334e-4005-a7a9-1c1ab5f4ad02-serving-cert\") pod \"controller-manager-6b4bcb86d5-b57w5\" (UID: \"cc2c5d65-334e-4005-a7a9-1c1ab5f4ad02\") " pod="openshift-controller-manager/controller-manager-6b4bcb86d5-b57w5" Oct 06 10:10:51 crc kubenswrapper[4824]: I1006 10:10:51.949533 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-78xgd\" (UniqueName: \"kubernetes.io/projected/cc2c5d65-334e-4005-a7a9-1c1ab5f4ad02-kube-api-access-78xgd\") pod \"controller-manager-6b4bcb86d5-b57w5\" (UID: \"cc2c5d65-334e-4005-a7a9-1c1ab5f4ad02\") " pod="openshift-controller-manager/controller-manager-6b4bcb86d5-b57w5" Oct 06 10:10:51 crc kubenswrapper[4824]: I1006 10:10:51.949559 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ab3e2669-11e7-4a27-89c3-4983798a8997-client-ca\") pod \"route-controller-manager-6d646b97ff-jhtrw\" (UID: \"ab3e2669-11e7-4a27-89c3-4983798a8997\") " pod="openshift-route-controller-manager/route-controller-manager-6d646b97ff-jhtrw" Oct 06 10:10:51 crc kubenswrapper[4824]: I1006 10:10:51.949592 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cc2c5d65-334e-4005-a7a9-1c1ab5f4ad02-config\") pod \"controller-manager-6b4bcb86d5-b57w5\" (UID: \"cc2c5d65-334e-4005-a7a9-1c1ab5f4ad02\") " pod="openshift-controller-manager/controller-manager-6b4bcb86d5-b57w5" Oct 06 10:10:51 crc kubenswrapper[4824]: I1006 10:10:51.949616 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ab3e2669-11e7-4a27-89c3-4983798a8997-config\") pod \"route-controller-manager-6d646b97ff-jhtrw\" (UID: \"ab3e2669-11e7-4a27-89c3-4983798a8997\") " pod="openshift-route-controller-manager/route-controller-manager-6d646b97ff-jhtrw" Oct 06 10:10:52 crc kubenswrapper[4824]: I1006 10:10:52.051812 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qvhtz\" (UniqueName: \"kubernetes.io/projected/ab3e2669-11e7-4a27-89c3-4983798a8997-kube-api-access-qvhtz\") pod \"route-controller-manager-6d646b97ff-jhtrw\" (UID: \"ab3e2669-11e7-4a27-89c3-4983798a8997\") " pod="openshift-route-controller-manager/route-controller-manager-6d646b97ff-jhtrw" Oct 06 10:10:52 crc kubenswrapper[4824]: I1006 10:10:52.051891 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/cc2c5d65-334e-4005-a7a9-1c1ab5f4ad02-client-ca\") pod \"controller-manager-6b4bcb86d5-b57w5\" (UID: \"cc2c5d65-334e-4005-a7a9-1c1ab5f4ad02\") " pod="openshift-controller-manager/controller-manager-6b4bcb86d5-b57w5" Oct 06 10:10:52 crc kubenswrapper[4824]: I1006 10:10:52.051919 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/cc2c5d65-334e-4005-a7a9-1c1ab5f4ad02-proxy-ca-bundles\") pod \"controller-manager-6b4bcb86d5-b57w5\" (UID: \"cc2c5d65-334e-4005-a7a9-1c1ab5f4ad02\") " pod="openshift-controller-manager/controller-manager-6b4bcb86d5-b57w5" Oct 06 10:10:52 crc kubenswrapper[4824]: I1006 10:10:52.051948 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ab3e2669-11e7-4a27-89c3-4983798a8997-serving-cert\") pod \"route-controller-manager-6d646b97ff-jhtrw\" (UID: \"ab3e2669-11e7-4a27-89c3-4983798a8997\") " pod="openshift-route-controller-manager/route-controller-manager-6d646b97ff-jhtrw" Oct 06 10:10:52 crc kubenswrapper[4824]: I1006 10:10:52.051997 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cc2c5d65-334e-4005-a7a9-1c1ab5f4ad02-serving-cert\") pod \"controller-manager-6b4bcb86d5-b57w5\" (UID: \"cc2c5d65-334e-4005-a7a9-1c1ab5f4ad02\") " pod="openshift-controller-manager/controller-manager-6b4bcb86d5-b57w5" Oct 06 10:10:52 crc kubenswrapper[4824]: I1006 10:10:52.052021 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-78xgd\" (UniqueName: \"kubernetes.io/projected/cc2c5d65-334e-4005-a7a9-1c1ab5f4ad02-kube-api-access-78xgd\") pod \"controller-manager-6b4bcb86d5-b57w5\" (UID: \"cc2c5d65-334e-4005-a7a9-1c1ab5f4ad02\") " pod="openshift-controller-manager/controller-manager-6b4bcb86d5-b57w5" Oct 06 10:10:52 crc kubenswrapper[4824]: I1006 10:10:52.052048 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ab3e2669-11e7-4a27-89c3-4983798a8997-client-ca\") pod \"route-controller-manager-6d646b97ff-jhtrw\" (UID: \"ab3e2669-11e7-4a27-89c3-4983798a8997\") " pod="openshift-route-controller-manager/route-controller-manager-6d646b97ff-jhtrw" Oct 06 10:10:52 crc kubenswrapper[4824]: I1006 10:10:52.052078 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cc2c5d65-334e-4005-a7a9-1c1ab5f4ad02-config\") pod \"controller-manager-6b4bcb86d5-b57w5\" (UID: \"cc2c5d65-334e-4005-a7a9-1c1ab5f4ad02\") " pod="openshift-controller-manager/controller-manager-6b4bcb86d5-b57w5" Oct 06 10:10:52 crc kubenswrapper[4824]: I1006 10:10:52.052099 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ab3e2669-11e7-4a27-89c3-4983798a8997-config\") pod \"route-controller-manager-6d646b97ff-jhtrw\" (UID: \"ab3e2669-11e7-4a27-89c3-4983798a8997\") " pod="openshift-route-controller-manager/route-controller-manager-6d646b97ff-jhtrw" Oct 06 10:10:52 crc kubenswrapper[4824]: I1006 10:10:52.053376 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ab3e2669-11e7-4a27-89c3-4983798a8997-config\") pod \"route-controller-manager-6d646b97ff-jhtrw\" (UID: \"ab3e2669-11e7-4a27-89c3-4983798a8997\") " pod="openshift-route-controller-manager/route-controller-manager-6d646b97ff-jhtrw" Oct 06 10:10:52 crc kubenswrapper[4824]: I1006 10:10:52.053431 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/cc2c5d65-334e-4005-a7a9-1c1ab5f4ad02-client-ca\") pod \"controller-manager-6b4bcb86d5-b57w5\" (UID: \"cc2c5d65-334e-4005-a7a9-1c1ab5f4ad02\") " pod="openshift-controller-manager/controller-manager-6b4bcb86d5-b57w5" Oct 06 10:10:52 crc kubenswrapper[4824]: I1006 10:10:52.054593 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/cc2c5d65-334e-4005-a7a9-1c1ab5f4ad02-proxy-ca-bundles\") pod \"controller-manager-6b4bcb86d5-b57w5\" (UID: \"cc2c5d65-334e-4005-a7a9-1c1ab5f4ad02\") " pod="openshift-controller-manager/controller-manager-6b4bcb86d5-b57w5" Oct 06 10:10:52 crc kubenswrapper[4824]: I1006 10:10:52.056096 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ab3e2669-11e7-4a27-89c3-4983798a8997-client-ca\") pod \"route-controller-manager-6d646b97ff-jhtrw\" (UID: \"ab3e2669-11e7-4a27-89c3-4983798a8997\") " pod="openshift-route-controller-manager/route-controller-manager-6d646b97ff-jhtrw" Oct 06 10:10:52 crc kubenswrapper[4824]: I1006 10:10:52.056953 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cc2c5d65-334e-4005-a7a9-1c1ab5f4ad02-config\") pod \"controller-manager-6b4bcb86d5-b57w5\" (UID: \"cc2c5d65-334e-4005-a7a9-1c1ab5f4ad02\") " pod="openshift-controller-manager/controller-manager-6b4bcb86d5-b57w5" Oct 06 10:10:52 crc kubenswrapper[4824]: I1006 10:10:52.059173 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cc2c5d65-334e-4005-a7a9-1c1ab5f4ad02-serving-cert\") pod \"controller-manager-6b4bcb86d5-b57w5\" (UID: \"cc2c5d65-334e-4005-a7a9-1c1ab5f4ad02\") " pod="openshift-controller-manager/controller-manager-6b4bcb86d5-b57w5" Oct 06 10:10:52 crc kubenswrapper[4824]: I1006 10:10:52.062147 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ab3e2669-11e7-4a27-89c3-4983798a8997-serving-cert\") pod \"route-controller-manager-6d646b97ff-jhtrw\" (UID: \"ab3e2669-11e7-4a27-89c3-4983798a8997\") " pod="openshift-route-controller-manager/route-controller-manager-6d646b97ff-jhtrw" Oct 06 10:10:52 crc kubenswrapper[4824]: I1006 10:10:52.074450 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qvhtz\" (UniqueName: \"kubernetes.io/projected/ab3e2669-11e7-4a27-89c3-4983798a8997-kube-api-access-qvhtz\") pod \"route-controller-manager-6d646b97ff-jhtrw\" (UID: \"ab3e2669-11e7-4a27-89c3-4983798a8997\") " pod="openshift-route-controller-manager/route-controller-manager-6d646b97ff-jhtrw" Oct 06 10:10:52 crc kubenswrapper[4824]: I1006 10:10:52.074900 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-78xgd\" (UniqueName: \"kubernetes.io/projected/cc2c5d65-334e-4005-a7a9-1c1ab5f4ad02-kube-api-access-78xgd\") pod \"controller-manager-6b4bcb86d5-b57w5\" (UID: \"cc2c5d65-334e-4005-a7a9-1c1ab5f4ad02\") " pod="openshift-controller-manager/controller-manager-6b4bcb86d5-b57w5" Oct 06 10:10:52 crc kubenswrapper[4824]: I1006 10:10:52.158615 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6b4bcb86d5-b57w5" Oct 06 10:10:52 crc kubenswrapper[4824]: I1006 10:10:52.176617 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6d646b97ff-jhtrw" Oct 06 10:10:52 crc kubenswrapper[4824]: I1006 10:10:52.442291 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-6b4bcb86d5-b57w5"] Oct 06 10:10:52 crc kubenswrapper[4824]: I1006 10:10:52.457049 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6d646b97ff-jhtrw"] Oct 06 10:10:52 crc kubenswrapper[4824]: I1006 10:10:52.676218 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-6b4bcb86d5-b57w5"] Oct 06 10:10:52 crc kubenswrapper[4824]: I1006 10:10:52.758011 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6d646b97ff-jhtrw"] Oct 06 10:10:52 crc kubenswrapper[4824]: I1006 10:10:52.766213 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6b4bcb86d5-b57w5" event={"ID":"cc2c5d65-334e-4005-a7a9-1c1ab5f4ad02","Type":"ContainerStarted","Data":"87cacfd81f328474c378f835fc0f4905b5bc5f8378f4f54708b2e2936685959f"} Oct 06 10:10:53 crc kubenswrapper[4824]: I1006 10:10:53.281040 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6f7d3cb-418e-4f73-95a0-2477e37976c3" path="/var/lib/kubelet/pods/b6f7d3cb-418e-4f73-95a0-2477e37976c3/volumes" Oct 06 10:10:53 crc kubenswrapper[4824]: I1006 10:10:53.282181 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dd50f0ac-ac97-413f-a350-7496681f7f0a" path="/var/lib/kubelet/pods/dd50f0ac-ac97-413f-a350-7496681f7f0a/volumes" Oct 06 10:10:53 crc kubenswrapper[4824]: I1006 10:10:53.776251 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6d646b97ff-jhtrw" event={"ID":"ab3e2669-11e7-4a27-89c3-4983798a8997","Type":"ContainerStarted","Data":"54332fbe4132bd45c60bdd287c9dac73bec459659826bfce4374c2a8563fdbad"} Oct 06 10:10:53 crc kubenswrapper[4824]: I1006 10:10:53.776321 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6d646b97ff-jhtrw" event={"ID":"ab3e2669-11e7-4a27-89c3-4983798a8997","Type":"ContainerStarted","Data":"4a3ab4b8d27e3017726f973a53e1db68d96f4095808ff84feef93b47a3481184"} Oct 06 10:10:53 crc kubenswrapper[4824]: I1006 10:10:53.776349 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6d646b97ff-jhtrw" Oct 06 10:10:53 crc kubenswrapper[4824]: I1006 10:10:53.776379 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6d646b97ff-jhtrw" podUID="ab3e2669-11e7-4a27-89c3-4983798a8997" containerName="route-controller-manager" containerID="cri-o://54332fbe4132bd45c60bdd287c9dac73bec459659826bfce4374c2a8563fdbad" gracePeriod=30 Oct 06 10:10:53 crc kubenswrapper[4824]: I1006 10:10:53.781394 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6b4bcb86d5-b57w5" event={"ID":"cc2c5d65-334e-4005-a7a9-1c1ab5f4ad02","Type":"ContainerStarted","Data":"82988673c87dd2c3b35c39ed2fea775bd4c8d67900a4076eab43f66e314f3128"} Oct 06 10:10:53 crc kubenswrapper[4824]: I1006 10:10:53.781777 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-6b4bcb86d5-b57w5" Oct 06 10:10:53 crc kubenswrapper[4824]: I1006 10:10:53.781549 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-6b4bcb86d5-b57w5" podUID="cc2c5d65-334e-4005-a7a9-1c1ab5f4ad02" containerName="controller-manager" containerID="cri-o://82988673c87dd2c3b35c39ed2fea775bd4c8d67900a4076eab43f66e314f3128" gracePeriod=30 Oct 06 10:10:53 crc kubenswrapper[4824]: I1006 10:10:53.783642 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6d646b97ff-jhtrw" Oct 06 10:10:53 crc kubenswrapper[4824]: I1006 10:10:53.790218 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-6b4bcb86d5-b57w5" Oct 06 10:10:53 crc kubenswrapper[4824]: I1006 10:10:53.804779 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6d646b97ff-jhtrw" podStartSLOduration=3.804746788 podStartE2EDuration="3.804746788s" podCreationTimestamp="2025-10-06 10:10:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:10:53.803818175 +0000 UTC m=+763.168241046" watchObservedRunningTime="2025-10-06 10:10:53.804746788 +0000 UTC m=+763.169169649" Oct 06 10:10:53 crc kubenswrapper[4824]: I1006 10:10:53.825276 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-6b4bcb86d5-b57w5" podStartSLOduration=3.825252279 podStartE2EDuration="3.825252279s" podCreationTimestamp="2025-10-06 10:10:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:10:53.82165536 +0000 UTC m=+763.186078241" watchObservedRunningTime="2025-10-06 10:10:53.825252279 +0000 UTC m=+763.189675140" Oct 06 10:10:53 crc kubenswrapper[4824]: I1006 10:10:53.868643 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-57vkd"] Oct 06 10:10:53 crc kubenswrapper[4824]: I1006 10:10:53.870688 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-57vkd" Oct 06 10:10:53 crc kubenswrapper[4824]: I1006 10:10:53.878604 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-57vkd"] Oct 06 10:10:53 crc kubenswrapper[4824]: I1006 10:10:53.885685 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b4856c75-71c0-4f10-9bb6-b2b04ac241ff-catalog-content\") pod \"certified-operators-57vkd\" (UID: \"b4856c75-71c0-4f10-9bb6-b2b04ac241ff\") " pod="openshift-marketplace/certified-operators-57vkd" Oct 06 10:10:53 crc kubenswrapper[4824]: I1006 10:10:53.885748 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hl598\" (UniqueName: \"kubernetes.io/projected/b4856c75-71c0-4f10-9bb6-b2b04ac241ff-kube-api-access-hl598\") pod \"certified-operators-57vkd\" (UID: \"b4856c75-71c0-4f10-9bb6-b2b04ac241ff\") " pod="openshift-marketplace/certified-operators-57vkd" Oct 06 10:10:53 crc kubenswrapper[4824]: I1006 10:10:53.885785 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b4856c75-71c0-4f10-9bb6-b2b04ac241ff-utilities\") pod \"certified-operators-57vkd\" (UID: \"b4856c75-71c0-4f10-9bb6-b2b04ac241ff\") " pod="openshift-marketplace/certified-operators-57vkd" Oct 06 10:10:53 crc kubenswrapper[4824]: I1006 10:10:53.986918 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b4856c75-71c0-4f10-9bb6-b2b04ac241ff-catalog-content\") pod \"certified-operators-57vkd\" (UID: \"b4856c75-71c0-4f10-9bb6-b2b04ac241ff\") " pod="openshift-marketplace/certified-operators-57vkd" Oct 06 10:10:53 crc kubenswrapper[4824]: I1006 10:10:53.986999 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hl598\" (UniqueName: \"kubernetes.io/projected/b4856c75-71c0-4f10-9bb6-b2b04ac241ff-kube-api-access-hl598\") pod \"certified-operators-57vkd\" (UID: \"b4856c75-71c0-4f10-9bb6-b2b04ac241ff\") " pod="openshift-marketplace/certified-operators-57vkd" Oct 06 10:10:53 crc kubenswrapper[4824]: I1006 10:10:53.987031 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b4856c75-71c0-4f10-9bb6-b2b04ac241ff-utilities\") pod \"certified-operators-57vkd\" (UID: \"b4856c75-71c0-4f10-9bb6-b2b04ac241ff\") " pod="openshift-marketplace/certified-operators-57vkd" Oct 06 10:10:53 crc kubenswrapper[4824]: I1006 10:10:53.987603 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b4856c75-71c0-4f10-9bb6-b2b04ac241ff-catalog-content\") pod \"certified-operators-57vkd\" (UID: \"b4856c75-71c0-4f10-9bb6-b2b04ac241ff\") " pod="openshift-marketplace/certified-operators-57vkd" Oct 06 10:10:53 crc kubenswrapper[4824]: I1006 10:10:53.987643 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b4856c75-71c0-4f10-9bb6-b2b04ac241ff-utilities\") pod \"certified-operators-57vkd\" (UID: \"b4856c75-71c0-4f10-9bb6-b2b04ac241ff\") " pod="openshift-marketplace/certified-operators-57vkd" Oct 06 10:10:54 crc kubenswrapper[4824]: I1006 10:10:54.035518 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hl598\" (UniqueName: \"kubernetes.io/projected/b4856c75-71c0-4f10-9bb6-b2b04ac241ff-kube-api-access-hl598\") pod \"certified-operators-57vkd\" (UID: \"b4856c75-71c0-4f10-9bb6-b2b04ac241ff\") " pod="openshift-marketplace/certified-operators-57vkd" Oct 06 10:10:54 crc kubenswrapper[4824]: I1006 10:10:54.202442 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6d646b97ff-jhtrw" Oct 06 10:10:54 crc kubenswrapper[4824]: I1006 10:10:54.237272 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-57vkd" Oct 06 10:10:54 crc kubenswrapper[4824]: I1006 10:10:54.241933 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-67b4ccc97f-l4pzs"] Oct 06 10:10:54 crc kubenswrapper[4824]: E1006 10:10:54.242284 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab3e2669-11e7-4a27-89c3-4983798a8997" containerName="route-controller-manager" Oct 06 10:10:54 crc kubenswrapper[4824]: I1006 10:10:54.242314 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab3e2669-11e7-4a27-89c3-4983798a8997" containerName="route-controller-manager" Oct 06 10:10:54 crc kubenswrapper[4824]: I1006 10:10:54.242461 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="ab3e2669-11e7-4a27-89c3-4983798a8997" containerName="route-controller-manager" Oct 06 10:10:54 crc kubenswrapper[4824]: I1006 10:10:54.245325 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-67b4ccc97f-l4pzs" Oct 06 10:10:54 crc kubenswrapper[4824]: I1006 10:10:54.271368 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-67b4ccc97f-l4pzs"] Oct 06 10:10:54 crc kubenswrapper[4824]: I1006 10:10:54.297572 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ab3e2669-11e7-4a27-89c3-4983798a8997-serving-cert\") pod \"ab3e2669-11e7-4a27-89c3-4983798a8997\" (UID: \"ab3e2669-11e7-4a27-89c3-4983798a8997\") " Oct 06 10:10:54 crc kubenswrapper[4824]: I1006 10:10:54.298201 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ab3e2669-11e7-4a27-89c3-4983798a8997-config\") pod \"ab3e2669-11e7-4a27-89c3-4983798a8997\" (UID: \"ab3e2669-11e7-4a27-89c3-4983798a8997\") " Oct 06 10:10:54 crc kubenswrapper[4824]: I1006 10:10:54.298246 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qvhtz\" (UniqueName: \"kubernetes.io/projected/ab3e2669-11e7-4a27-89c3-4983798a8997-kube-api-access-qvhtz\") pod \"ab3e2669-11e7-4a27-89c3-4983798a8997\" (UID: \"ab3e2669-11e7-4a27-89c3-4983798a8997\") " Oct 06 10:10:54 crc kubenswrapper[4824]: I1006 10:10:54.298311 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ab3e2669-11e7-4a27-89c3-4983798a8997-client-ca\") pod \"ab3e2669-11e7-4a27-89c3-4983798a8997\" (UID: \"ab3e2669-11e7-4a27-89c3-4983798a8997\") " Oct 06 10:10:54 crc kubenswrapper[4824]: I1006 10:10:54.298818 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1743c9ab-4f5c-4de2-9fc4-3f93e1bd23b7-config\") pod \"route-controller-manager-67b4ccc97f-l4pzs\" (UID: \"1743c9ab-4f5c-4de2-9fc4-3f93e1bd23b7\") " pod="openshift-route-controller-manager/route-controller-manager-67b4ccc97f-l4pzs" Oct 06 10:10:54 crc kubenswrapper[4824]: I1006 10:10:54.298904 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1743c9ab-4f5c-4de2-9fc4-3f93e1bd23b7-client-ca\") pod \"route-controller-manager-67b4ccc97f-l4pzs\" (UID: \"1743c9ab-4f5c-4de2-9fc4-3f93e1bd23b7\") " pod="openshift-route-controller-manager/route-controller-manager-67b4ccc97f-l4pzs" Oct 06 10:10:54 crc kubenswrapper[4824]: I1006 10:10:54.298937 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6dl6v\" (UniqueName: \"kubernetes.io/projected/1743c9ab-4f5c-4de2-9fc4-3f93e1bd23b7-kube-api-access-6dl6v\") pod \"route-controller-manager-67b4ccc97f-l4pzs\" (UID: \"1743c9ab-4f5c-4de2-9fc4-3f93e1bd23b7\") " pod="openshift-route-controller-manager/route-controller-manager-67b4ccc97f-l4pzs" Oct 06 10:10:54 crc kubenswrapper[4824]: I1006 10:10:54.299003 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1743c9ab-4f5c-4de2-9fc4-3f93e1bd23b7-serving-cert\") pod \"route-controller-manager-67b4ccc97f-l4pzs\" (UID: \"1743c9ab-4f5c-4de2-9fc4-3f93e1bd23b7\") " pod="openshift-route-controller-manager/route-controller-manager-67b4ccc97f-l4pzs" Oct 06 10:10:54 crc kubenswrapper[4824]: I1006 10:10:54.301717 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ab3e2669-11e7-4a27-89c3-4983798a8997-client-ca" (OuterVolumeSpecName: "client-ca") pod "ab3e2669-11e7-4a27-89c3-4983798a8997" (UID: "ab3e2669-11e7-4a27-89c3-4983798a8997"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:10:54 crc kubenswrapper[4824]: I1006 10:10:54.307239 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ab3e2669-11e7-4a27-89c3-4983798a8997-config" (OuterVolumeSpecName: "config") pod "ab3e2669-11e7-4a27-89c3-4983798a8997" (UID: "ab3e2669-11e7-4a27-89c3-4983798a8997"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:10:54 crc kubenswrapper[4824]: I1006 10:10:54.319175 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ab3e2669-11e7-4a27-89c3-4983798a8997-kube-api-access-qvhtz" (OuterVolumeSpecName: "kube-api-access-qvhtz") pod "ab3e2669-11e7-4a27-89c3-4983798a8997" (UID: "ab3e2669-11e7-4a27-89c3-4983798a8997"). InnerVolumeSpecName "kube-api-access-qvhtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:10:54 crc kubenswrapper[4824]: I1006 10:10:54.326163 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6b4bcb86d5-b57w5" Oct 06 10:10:54 crc kubenswrapper[4824]: I1006 10:10:54.331487 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab3e2669-11e7-4a27-89c3-4983798a8997-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "ab3e2669-11e7-4a27-89c3-4983798a8997" (UID: "ab3e2669-11e7-4a27-89c3-4983798a8997"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:10:54 crc kubenswrapper[4824]: I1006 10:10:54.399735 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/cc2c5d65-334e-4005-a7a9-1c1ab5f4ad02-client-ca\") pod \"cc2c5d65-334e-4005-a7a9-1c1ab5f4ad02\" (UID: \"cc2c5d65-334e-4005-a7a9-1c1ab5f4ad02\") " Oct 06 10:10:54 crc kubenswrapper[4824]: I1006 10:10:54.399846 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-78xgd\" (UniqueName: \"kubernetes.io/projected/cc2c5d65-334e-4005-a7a9-1c1ab5f4ad02-kube-api-access-78xgd\") pod \"cc2c5d65-334e-4005-a7a9-1c1ab5f4ad02\" (UID: \"cc2c5d65-334e-4005-a7a9-1c1ab5f4ad02\") " Oct 06 10:10:54 crc kubenswrapper[4824]: I1006 10:10:54.399943 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cc2c5d65-334e-4005-a7a9-1c1ab5f4ad02-config\") pod \"cc2c5d65-334e-4005-a7a9-1c1ab5f4ad02\" (UID: \"cc2c5d65-334e-4005-a7a9-1c1ab5f4ad02\") " Oct 06 10:10:54 crc kubenswrapper[4824]: I1006 10:10:54.399965 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/cc2c5d65-334e-4005-a7a9-1c1ab5f4ad02-proxy-ca-bundles\") pod \"cc2c5d65-334e-4005-a7a9-1c1ab5f4ad02\" (UID: \"cc2c5d65-334e-4005-a7a9-1c1ab5f4ad02\") " Oct 06 10:10:54 crc kubenswrapper[4824]: I1006 10:10:54.400031 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cc2c5d65-334e-4005-a7a9-1c1ab5f4ad02-serving-cert\") pod \"cc2c5d65-334e-4005-a7a9-1c1ab5f4ad02\" (UID: \"cc2c5d65-334e-4005-a7a9-1c1ab5f4ad02\") " Oct 06 10:10:54 crc kubenswrapper[4824]: I1006 10:10:54.401347 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cc2c5d65-334e-4005-a7a9-1c1ab5f4ad02-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "cc2c5d65-334e-4005-a7a9-1c1ab5f4ad02" (UID: "cc2c5d65-334e-4005-a7a9-1c1ab5f4ad02"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:10:54 crc kubenswrapper[4824]: I1006 10:10:54.401389 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cc2c5d65-334e-4005-a7a9-1c1ab5f4ad02-config" (OuterVolumeSpecName: "config") pod "cc2c5d65-334e-4005-a7a9-1c1ab5f4ad02" (UID: "cc2c5d65-334e-4005-a7a9-1c1ab5f4ad02"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:10:54 crc kubenswrapper[4824]: I1006 10:10:54.401995 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cc2c5d65-334e-4005-a7a9-1c1ab5f4ad02-client-ca" (OuterVolumeSpecName: "client-ca") pod "cc2c5d65-334e-4005-a7a9-1c1ab5f4ad02" (UID: "cc2c5d65-334e-4005-a7a9-1c1ab5f4ad02"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:10:54 crc kubenswrapper[4824]: I1006 10:10:54.402344 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1743c9ab-4f5c-4de2-9fc4-3f93e1bd23b7-config\") pod \"route-controller-manager-67b4ccc97f-l4pzs\" (UID: \"1743c9ab-4f5c-4de2-9fc4-3f93e1bd23b7\") " pod="openshift-route-controller-manager/route-controller-manager-67b4ccc97f-l4pzs" Oct 06 10:10:54 crc kubenswrapper[4824]: I1006 10:10:54.403630 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cc2c5d65-334e-4005-a7a9-1c1ab5f4ad02-kube-api-access-78xgd" (OuterVolumeSpecName: "kube-api-access-78xgd") pod "cc2c5d65-334e-4005-a7a9-1c1ab5f4ad02" (UID: "cc2c5d65-334e-4005-a7a9-1c1ab5f4ad02"). InnerVolumeSpecName "kube-api-access-78xgd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:10:54 crc kubenswrapper[4824]: I1006 10:10:54.404092 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1743c9ab-4f5c-4de2-9fc4-3f93e1bd23b7-config\") pod \"route-controller-manager-67b4ccc97f-l4pzs\" (UID: \"1743c9ab-4f5c-4de2-9fc4-3f93e1bd23b7\") " pod="openshift-route-controller-manager/route-controller-manager-67b4ccc97f-l4pzs" Oct 06 10:10:54 crc kubenswrapper[4824]: I1006 10:10:54.404377 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1743c9ab-4f5c-4de2-9fc4-3f93e1bd23b7-client-ca\") pod \"route-controller-manager-67b4ccc97f-l4pzs\" (UID: \"1743c9ab-4f5c-4de2-9fc4-3f93e1bd23b7\") " pod="openshift-route-controller-manager/route-controller-manager-67b4ccc97f-l4pzs" Oct 06 10:10:54 crc kubenswrapper[4824]: I1006 10:10:54.404454 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6dl6v\" (UniqueName: \"kubernetes.io/projected/1743c9ab-4f5c-4de2-9fc4-3f93e1bd23b7-kube-api-access-6dl6v\") pod \"route-controller-manager-67b4ccc97f-l4pzs\" (UID: \"1743c9ab-4f5c-4de2-9fc4-3f93e1bd23b7\") " pod="openshift-route-controller-manager/route-controller-manager-67b4ccc97f-l4pzs" Oct 06 10:10:54 crc kubenswrapper[4824]: I1006 10:10:54.404563 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1743c9ab-4f5c-4de2-9fc4-3f93e1bd23b7-serving-cert\") pod \"route-controller-manager-67b4ccc97f-l4pzs\" (UID: \"1743c9ab-4f5c-4de2-9fc4-3f93e1bd23b7\") " pod="openshift-route-controller-manager/route-controller-manager-67b4ccc97f-l4pzs" Oct 06 10:10:54 crc kubenswrapper[4824]: I1006 10:10:54.404686 4824 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ab3e2669-11e7-4a27-89c3-4983798a8997-client-ca\") on node \"crc\" DevicePath \"\"" Oct 06 10:10:54 crc kubenswrapper[4824]: I1006 10:10:54.404716 4824 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/cc2c5d65-334e-4005-a7a9-1c1ab5f4ad02-client-ca\") on node \"crc\" DevicePath \"\"" Oct 06 10:10:54 crc kubenswrapper[4824]: I1006 10:10:54.404730 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-78xgd\" (UniqueName: \"kubernetes.io/projected/cc2c5d65-334e-4005-a7a9-1c1ab5f4ad02-kube-api-access-78xgd\") on node \"crc\" DevicePath \"\"" Oct 06 10:10:54 crc kubenswrapper[4824]: I1006 10:10:54.404743 4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ab3e2669-11e7-4a27-89c3-4983798a8997-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 06 10:10:54 crc kubenswrapper[4824]: I1006 10:10:54.404875 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cc2c5d65-334e-4005-a7a9-1c1ab5f4ad02-config\") on node \"crc\" DevicePath \"\"" Oct 06 10:10:54 crc kubenswrapper[4824]: I1006 10:10:54.404891 4824 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/cc2c5d65-334e-4005-a7a9-1c1ab5f4ad02-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 06 10:10:54 crc kubenswrapper[4824]: I1006 10:10:54.404904 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ab3e2669-11e7-4a27-89c3-4983798a8997-config\") on node \"crc\" DevicePath \"\"" Oct 06 10:10:54 crc kubenswrapper[4824]: I1006 10:10:54.404917 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qvhtz\" (UniqueName: \"kubernetes.io/projected/ab3e2669-11e7-4a27-89c3-4983798a8997-kube-api-access-qvhtz\") on node \"crc\" DevicePath \"\"" Oct 06 10:10:54 crc kubenswrapper[4824]: I1006 10:10:54.405923 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1743c9ab-4f5c-4de2-9fc4-3f93e1bd23b7-client-ca\") pod \"route-controller-manager-67b4ccc97f-l4pzs\" (UID: \"1743c9ab-4f5c-4de2-9fc4-3f93e1bd23b7\") " pod="openshift-route-controller-manager/route-controller-manager-67b4ccc97f-l4pzs" Oct 06 10:10:54 crc kubenswrapper[4824]: I1006 10:10:54.408405 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cc2c5d65-334e-4005-a7a9-1c1ab5f4ad02-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "cc2c5d65-334e-4005-a7a9-1c1ab5f4ad02" (UID: "cc2c5d65-334e-4005-a7a9-1c1ab5f4ad02"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:10:54 crc kubenswrapper[4824]: I1006 10:10:54.416974 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1743c9ab-4f5c-4de2-9fc4-3f93e1bd23b7-serving-cert\") pod \"route-controller-manager-67b4ccc97f-l4pzs\" (UID: \"1743c9ab-4f5c-4de2-9fc4-3f93e1bd23b7\") " pod="openshift-route-controller-manager/route-controller-manager-67b4ccc97f-l4pzs" Oct 06 10:10:54 crc kubenswrapper[4824]: I1006 10:10:54.429640 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6dl6v\" (UniqueName: \"kubernetes.io/projected/1743c9ab-4f5c-4de2-9fc4-3f93e1bd23b7-kube-api-access-6dl6v\") pod \"route-controller-manager-67b4ccc97f-l4pzs\" (UID: \"1743c9ab-4f5c-4de2-9fc4-3f93e1bd23b7\") " pod="openshift-route-controller-manager/route-controller-manager-67b4ccc97f-l4pzs" Oct 06 10:10:54 crc kubenswrapper[4824]: I1006 10:10:54.506299 4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cc2c5d65-334e-4005-a7a9-1c1ab5f4ad02-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 06 10:10:54 crc kubenswrapper[4824]: I1006 10:10:54.575712 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-57vkd"] Oct 06 10:10:54 crc kubenswrapper[4824]: W1006 10:10:54.584817 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb4856c75_71c0_4f10_9bb6_b2b04ac241ff.slice/crio-9839544bb03556ad10a90437a5af64c701fa954f1d9661ff579b5e2529314751 WatchSource:0}: Error finding container 9839544bb03556ad10a90437a5af64c701fa954f1d9661ff579b5e2529314751: Status 404 returned error can't find the container with id 9839544bb03556ad10a90437a5af64c701fa954f1d9661ff579b5e2529314751 Oct 06 10:10:54 crc kubenswrapper[4824]: I1006 10:10:54.613624 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-67b4ccc97f-l4pzs" Oct 06 10:10:54 crc kubenswrapper[4824]: I1006 10:10:54.792674 4824 generic.go:334] "Generic (PLEG): container finished" podID="cc2c5d65-334e-4005-a7a9-1c1ab5f4ad02" containerID="82988673c87dd2c3b35c39ed2fea775bd4c8d67900a4076eab43f66e314f3128" exitCode=0 Oct 06 10:10:54 crc kubenswrapper[4824]: I1006 10:10:54.792762 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6b4bcb86d5-b57w5" event={"ID":"cc2c5d65-334e-4005-a7a9-1c1ab5f4ad02","Type":"ContainerDied","Data":"82988673c87dd2c3b35c39ed2fea775bd4c8d67900a4076eab43f66e314f3128"} Oct 06 10:10:54 crc kubenswrapper[4824]: I1006 10:10:54.792805 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6b4bcb86d5-b57w5" event={"ID":"cc2c5d65-334e-4005-a7a9-1c1ab5f4ad02","Type":"ContainerDied","Data":"87cacfd81f328474c378f835fc0f4905b5bc5f8378f4f54708b2e2936685959f"} Oct 06 10:10:54 crc kubenswrapper[4824]: I1006 10:10:54.792829 4824 scope.go:117] "RemoveContainer" containerID="82988673c87dd2c3b35c39ed2fea775bd4c8d67900a4076eab43f66e314f3128" Oct 06 10:10:54 crc kubenswrapper[4824]: I1006 10:10:54.793182 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6b4bcb86d5-b57w5" Oct 06 10:10:54 crc kubenswrapper[4824]: I1006 10:10:54.804458 4824 generic.go:334] "Generic (PLEG): container finished" podID="ab3e2669-11e7-4a27-89c3-4983798a8997" containerID="54332fbe4132bd45c60bdd287c9dac73bec459659826bfce4374c2a8563fdbad" exitCode=0 Oct 06 10:10:54 crc kubenswrapper[4824]: I1006 10:10:54.805097 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6d646b97ff-jhtrw" event={"ID":"ab3e2669-11e7-4a27-89c3-4983798a8997","Type":"ContainerDied","Data":"54332fbe4132bd45c60bdd287c9dac73bec459659826bfce4374c2a8563fdbad"} Oct 06 10:10:54 crc kubenswrapper[4824]: I1006 10:10:54.805144 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6d646b97ff-jhtrw" event={"ID":"ab3e2669-11e7-4a27-89c3-4983798a8997","Type":"ContainerDied","Data":"4a3ab4b8d27e3017726f973a53e1db68d96f4095808ff84feef93b47a3481184"} Oct 06 10:10:54 crc kubenswrapper[4824]: I1006 10:10:54.805247 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6d646b97ff-jhtrw" Oct 06 10:10:54 crc kubenswrapper[4824]: I1006 10:10:54.808908 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-57vkd" event={"ID":"b4856c75-71c0-4f10-9bb6-b2b04ac241ff","Type":"ContainerStarted","Data":"62713eb13924cbf2fbd78523f9397c8ecc313bcb4cde994132ecf0086f7e05bb"} Oct 06 10:10:54 crc kubenswrapper[4824]: I1006 10:10:54.808967 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-57vkd" event={"ID":"b4856c75-71c0-4f10-9bb6-b2b04ac241ff","Type":"ContainerStarted","Data":"9839544bb03556ad10a90437a5af64c701fa954f1d9661ff579b5e2529314751"} Oct 06 10:10:54 crc kubenswrapper[4824]: I1006 10:10:54.845092 4824 scope.go:117] "RemoveContainer" containerID="82988673c87dd2c3b35c39ed2fea775bd4c8d67900a4076eab43f66e314f3128" Oct 06 10:10:54 crc kubenswrapper[4824]: E1006 10:10:54.847449 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"82988673c87dd2c3b35c39ed2fea775bd4c8d67900a4076eab43f66e314f3128\": container with ID starting with 82988673c87dd2c3b35c39ed2fea775bd4c8d67900a4076eab43f66e314f3128 not found: ID does not exist" containerID="82988673c87dd2c3b35c39ed2fea775bd4c8d67900a4076eab43f66e314f3128" Oct 06 10:10:54 crc kubenswrapper[4824]: I1006 10:10:54.847488 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"82988673c87dd2c3b35c39ed2fea775bd4c8d67900a4076eab43f66e314f3128"} err="failed to get container status \"82988673c87dd2c3b35c39ed2fea775bd4c8d67900a4076eab43f66e314f3128\": rpc error: code = NotFound desc = could not find container \"82988673c87dd2c3b35c39ed2fea775bd4c8d67900a4076eab43f66e314f3128\": container with ID starting with 82988673c87dd2c3b35c39ed2fea775bd4c8d67900a4076eab43f66e314f3128 not found: ID does not exist" Oct 06 10:10:54 crc kubenswrapper[4824]: I1006 10:10:54.847517 4824 scope.go:117] "RemoveContainer" containerID="54332fbe4132bd45c60bdd287c9dac73bec459659826bfce4374c2a8563fdbad" Oct 06 10:10:54 crc kubenswrapper[4824]: I1006 10:10:54.891192 4824 scope.go:117] "RemoveContainer" containerID="54332fbe4132bd45c60bdd287c9dac73bec459659826bfce4374c2a8563fdbad" Oct 06 10:10:54 crc kubenswrapper[4824]: E1006 10:10:54.892584 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"54332fbe4132bd45c60bdd287c9dac73bec459659826bfce4374c2a8563fdbad\": container with ID starting with 54332fbe4132bd45c60bdd287c9dac73bec459659826bfce4374c2a8563fdbad not found: ID does not exist" containerID="54332fbe4132bd45c60bdd287c9dac73bec459659826bfce4374c2a8563fdbad" Oct 06 10:10:54 crc kubenswrapper[4824]: I1006 10:10:54.892621 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"54332fbe4132bd45c60bdd287c9dac73bec459659826bfce4374c2a8563fdbad"} err="failed to get container status \"54332fbe4132bd45c60bdd287c9dac73bec459659826bfce4374c2a8563fdbad\": rpc error: code = NotFound desc = could not find container \"54332fbe4132bd45c60bdd287c9dac73bec459659826bfce4374c2a8563fdbad\": container with ID starting with 54332fbe4132bd45c60bdd287c9dac73bec459659826bfce4374c2a8563fdbad not found: ID does not exist" Oct 06 10:10:54 crc kubenswrapper[4824]: I1006 10:10:54.904049 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-6b4bcb86d5-b57w5"] Oct 06 10:10:54 crc kubenswrapper[4824]: I1006 10:10:54.911164 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-6b4bcb86d5-b57w5"] Oct 06 10:10:54 crc kubenswrapper[4824]: I1006 10:10:54.927986 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6d646b97ff-jhtrw"] Oct 06 10:10:54 crc kubenswrapper[4824]: I1006 10:10:54.928056 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6d646b97ff-jhtrw"] Oct 06 10:10:54 crc kubenswrapper[4824]: I1006 10:10:54.958156 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-67b4ccc97f-l4pzs"] Oct 06 10:10:55 crc kubenswrapper[4824]: I1006 10:10:55.287230 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ab3e2669-11e7-4a27-89c3-4983798a8997" path="/var/lib/kubelet/pods/ab3e2669-11e7-4a27-89c3-4983798a8997/volumes" Oct 06 10:10:55 crc kubenswrapper[4824]: I1006 10:10:55.290735 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cc2c5d65-334e-4005-a7a9-1c1ab5f4ad02" path="/var/lib/kubelet/pods/cc2c5d65-334e-4005-a7a9-1c1ab5f4ad02/volumes" Oct 06 10:10:55 crc kubenswrapper[4824]: I1006 10:10:55.817203 4824 generic.go:334] "Generic (PLEG): container finished" podID="b4856c75-71c0-4f10-9bb6-b2b04ac241ff" containerID="62713eb13924cbf2fbd78523f9397c8ecc313bcb4cde994132ecf0086f7e05bb" exitCode=0 Oct 06 10:10:55 crc kubenswrapper[4824]: I1006 10:10:55.817272 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-57vkd" event={"ID":"b4856c75-71c0-4f10-9bb6-b2b04ac241ff","Type":"ContainerDied","Data":"62713eb13924cbf2fbd78523f9397c8ecc313bcb4cde994132ecf0086f7e05bb"} Oct 06 10:10:55 crc kubenswrapper[4824]: I1006 10:10:55.819879 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-67b4ccc97f-l4pzs" event={"ID":"1743c9ab-4f5c-4de2-9fc4-3f93e1bd23b7","Type":"ContainerStarted","Data":"9ab5290409116b14747fd304bd51905e7b55d922fa97dd80772d8cfd128d167e"} Oct 06 10:10:55 crc kubenswrapper[4824]: I1006 10:10:55.819939 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-67b4ccc97f-l4pzs" event={"ID":"1743c9ab-4f5c-4de2-9fc4-3f93e1bd23b7","Type":"ContainerStarted","Data":"81dc8811c8a10eb14418ebc65653394161f3d44588f99377dc2c7729cd73b519"} Oct 06 10:10:55 crc kubenswrapper[4824]: I1006 10:10:55.820722 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-67b4ccc97f-l4pzs" Oct 06 10:10:55 crc kubenswrapper[4824]: I1006 10:10:55.831384 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-67b4ccc97f-l4pzs" Oct 06 10:10:55 crc kubenswrapper[4824]: I1006 10:10:55.853660 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-67b4ccc97f-l4pzs" podStartSLOduration=3.8536389399999997 podStartE2EDuration="3.85363894s" podCreationTimestamp="2025-10-06 10:10:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:10:55.852721757 +0000 UTC m=+765.217144618" watchObservedRunningTime="2025-10-06 10:10:55.85363894 +0000 UTC m=+765.218061801" Oct 06 10:10:56 crc kubenswrapper[4824]: I1006 10:10:56.795028 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-7666f86cb5-qqdq7"] Oct 06 10:10:56 crc kubenswrapper[4824]: E1006 10:10:56.796125 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc2c5d65-334e-4005-a7a9-1c1ab5f4ad02" containerName="controller-manager" Oct 06 10:10:56 crc kubenswrapper[4824]: I1006 10:10:56.796217 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc2c5d65-334e-4005-a7a9-1c1ab5f4ad02" containerName="controller-manager" Oct 06 10:10:56 crc kubenswrapper[4824]: I1006 10:10:56.796420 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc2c5d65-334e-4005-a7a9-1c1ab5f4ad02" containerName="controller-manager" Oct 06 10:10:56 crc kubenswrapper[4824]: I1006 10:10:56.796948 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7666f86cb5-qqdq7" Oct 06 10:10:56 crc kubenswrapper[4824]: I1006 10:10:56.800362 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Oct 06 10:10:56 crc kubenswrapper[4824]: I1006 10:10:56.800514 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Oct 06 10:10:56 crc kubenswrapper[4824]: I1006 10:10:56.800958 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Oct 06 10:10:56 crc kubenswrapper[4824]: I1006 10:10:56.801131 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Oct 06 10:10:56 crc kubenswrapper[4824]: I1006 10:10:56.801114 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Oct 06 10:10:56 crc kubenswrapper[4824]: I1006 10:10:56.802546 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Oct 06 10:10:56 crc kubenswrapper[4824]: I1006 10:10:56.813949 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Oct 06 10:10:56 crc kubenswrapper[4824]: I1006 10:10:56.825300 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7666f86cb5-qqdq7"] Oct 06 10:10:56 crc kubenswrapper[4824]: I1006 10:10:56.831519 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-57vkd" event={"ID":"b4856c75-71c0-4f10-9bb6-b2b04ac241ff","Type":"ContainerStarted","Data":"c47dbd5ffaa8525e40242beea1eba51d121b4c772ba1dfa755938c3d704100fe"} Oct 06 10:10:56 crc kubenswrapper[4824]: I1006 10:10:56.842129 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1211c343-85b5-4fb3-9203-60ff7e36adbe-serving-cert\") pod \"controller-manager-7666f86cb5-qqdq7\" (UID: \"1211c343-85b5-4fb3-9203-60ff7e36adbe\") " pod="openshift-controller-manager/controller-manager-7666f86cb5-qqdq7" Oct 06 10:10:56 crc kubenswrapper[4824]: I1006 10:10:56.842215 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1211c343-85b5-4fb3-9203-60ff7e36adbe-proxy-ca-bundles\") pod \"controller-manager-7666f86cb5-qqdq7\" (UID: \"1211c343-85b5-4fb3-9203-60ff7e36adbe\") " pod="openshift-controller-manager/controller-manager-7666f86cb5-qqdq7" Oct 06 10:10:56 crc kubenswrapper[4824]: I1006 10:10:56.842360 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fvfp7\" (UniqueName: \"kubernetes.io/projected/1211c343-85b5-4fb3-9203-60ff7e36adbe-kube-api-access-fvfp7\") pod \"controller-manager-7666f86cb5-qqdq7\" (UID: \"1211c343-85b5-4fb3-9203-60ff7e36adbe\") " pod="openshift-controller-manager/controller-manager-7666f86cb5-qqdq7" Oct 06 10:10:56 crc kubenswrapper[4824]: I1006 10:10:56.842433 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1211c343-85b5-4fb3-9203-60ff7e36adbe-config\") pod \"controller-manager-7666f86cb5-qqdq7\" (UID: \"1211c343-85b5-4fb3-9203-60ff7e36adbe\") " pod="openshift-controller-manager/controller-manager-7666f86cb5-qqdq7" Oct 06 10:10:56 crc kubenswrapper[4824]: I1006 10:10:56.842497 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1211c343-85b5-4fb3-9203-60ff7e36adbe-client-ca\") pod \"controller-manager-7666f86cb5-qqdq7\" (UID: \"1211c343-85b5-4fb3-9203-60ff7e36adbe\") " pod="openshift-controller-manager/controller-manager-7666f86cb5-qqdq7" Oct 06 10:10:56 crc kubenswrapper[4824]: I1006 10:10:56.944423 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fvfp7\" (UniqueName: \"kubernetes.io/projected/1211c343-85b5-4fb3-9203-60ff7e36adbe-kube-api-access-fvfp7\") pod \"controller-manager-7666f86cb5-qqdq7\" (UID: \"1211c343-85b5-4fb3-9203-60ff7e36adbe\") " pod="openshift-controller-manager/controller-manager-7666f86cb5-qqdq7" Oct 06 10:10:56 crc kubenswrapper[4824]: I1006 10:10:56.944929 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1211c343-85b5-4fb3-9203-60ff7e36adbe-config\") pod \"controller-manager-7666f86cb5-qqdq7\" (UID: \"1211c343-85b5-4fb3-9203-60ff7e36adbe\") " pod="openshift-controller-manager/controller-manager-7666f86cb5-qqdq7" Oct 06 10:10:56 crc kubenswrapper[4824]: I1006 10:10:56.945026 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1211c343-85b5-4fb3-9203-60ff7e36adbe-client-ca\") pod \"controller-manager-7666f86cb5-qqdq7\" (UID: \"1211c343-85b5-4fb3-9203-60ff7e36adbe\") " pod="openshift-controller-manager/controller-manager-7666f86cb5-qqdq7" Oct 06 10:10:56 crc kubenswrapper[4824]: I1006 10:10:56.946131 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1211c343-85b5-4fb3-9203-60ff7e36adbe-client-ca\") pod \"controller-manager-7666f86cb5-qqdq7\" (UID: \"1211c343-85b5-4fb3-9203-60ff7e36adbe\") " pod="openshift-controller-manager/controller-manager-7666f86cb5-qqdq7" Oct 06 10:10:56 crc kubenswrapper[4824]: I1006 10:10:56.945198 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1211c343-85b5-4fb3-9203-60ff7e36adbe-serving-cert\") pod \"controller-manager-7666f86cb5-qqdq7\" (UID: \"1211c343-85b5-4fb3-9203-60ff7e36adbe\") " pod="openshift-controller-manager/controller-manager-7666f86cb5-qqdq7" Oct 06 10:10:56 crc kubenswrapper[4824]: I1006 10:10:56.946365 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1211c343-85b5-4fb3-9203-60ff7e36adbe-config\") pod \"controller-manager-7666f86cb5-qqdq7\" (UID: \"1211c343-85b5-4fb3-9203-60ff7e36adbe\") " pod="openshift-controller-manager/controller-manager-7666f86cb5-qqdq7" Oct 06 10:10:56 crc kubenswrapper[4824]: I1006 10:10:56.946639 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1211c343-85b5-4fb3-9203-60ff7e36adbe-proxy-ca-bundles\") pod \"controller-manager-7666f86cb5-qqdq7\" (UID: \"1211c343-85b5-4fb3-9203-60ff7e36adbe\") " pod="openshift-controller-manager/controller-manager-7666f86cb5-qqdq7" Oct 06 10:10:56 crc kubenswrapper[4824]: I1006 10:10:56.947424 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1211c343-85b5-4fb3-9203-60ff7e36adbe-proxy-ca-bundles\") pod \"controller-manager-7666f86cb5-qqdq7\" (UID: \"1211c343-85b5-4fb3-9203-60ff7e36adbe\") " pod="openshift-controller-manager/controller-manager-7666f86cb5-qqdq7" Oct 06 10:10:56 crc kubenswrapper[4824]: I1006 10:10:56.953644 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1211c343-85b5-4fb3-9203-60ff7e36adbe-serving-cert\") pod \"controller-manager-7666f86cb5-qqdq7\" (UID: \"1211c343-85b5-4fb3-9203-60ff7e36adbe\") " pod="openshift-controller-manager/controller-manager-7666f86cb5-qqdq7" Oct 06 10:10:56 crc kubenswrapper[4824]: I1006 10:10:56.965408 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fvfp7\" (UniqueName: \"kubernetes.io/projected/1211c343-85b5-4fb3-9203-60ff7e36adbe-kube-api-access-fvfp7\") pod \"controller-manager-7666f86cb5-qqdq7\" (UID: \"1211c343-85b5-4fb3-9203-60ff7e36adbe\") " pod="openshift-controller-manager/controller-manager-7666f86cb5-qqdq7" Oct 06 10:10:57 crc kubenswrapper[4824]: I1006 10:10:57.115691 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7666f86cb5-qqdq7" Oct 06 10:10:57 crc kubenswrapper[4824]: I1006 10:10:57.665711 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7666f86cb5-qqdq7"] Oct 06 10:10:57 crc kubenswrapper[4824]: I1006 10:10:57.846577 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7666f86cb5-qqdq7" event={"ID":"1211c343-85b5-4fb3-9203-60ff7e36adbe","Type":"ContainerStarted","Data":"cee416408dbf953e69fdbd88001428a681a847369218db8b49410f3da29123d8"} Oct 06 10:10:57 crc kubenswrapper[4824]: I1006 10:10:57.847289 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7666f86cb5-qqdq7" event={"ID":"1211c343-85b5-4fb3-9203-60ff7e36adbe","Type":"ContainerStarted","Data":"1b5065083325ae30e4b2248f28839636909b7bf25150c63a93055beeba5ada74"} Oct 06 10:10:57 crc kubenswrapper[4824]: I1006 10:10:57.848281 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-7666f86cb5-qqdq7" Oct 06 10:10:57 crc kubenswrapper[4824]: I1006 10:10:57.850472 4824 patch_prober.go:28] interesting pod/controller-manager-7666f86cb5-qqdq7 container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.53:8443/healthz\": dial tcp 10.217.0.53:8443: connect: connection refused" start-of-body= Oct 06 10:10:57 crc kubenswrapper[4824]: I1006 10:10:57.850550 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-7666f86cb5-qqdq7" podUID="1211c343-85b5-4fb3-9203-60ff7e36adbe" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.53:8443/healthz\": dial tcp 10.217.0.53:8443: connect: connection refused" Oct 06 10:10:57 crc kubenswrapper[4824]: I1006 10:10:57.852605 4824 generic.go:334] "Generic (PLEG): container finished" podID="b4856c75-71c0-4f10-9bb6-b2b04ac241ff" containerID="c47dbd5ffaa8525e40242beea1eba51d121b4c772ba1dfa755938c3d704100fe" exitCode=0 Oct 06 10:10:57 crc kubenswrapper[4824]: I1006 10:10:57.852687 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-57vkd" event={"ID":"b4856c75-71c0-4f10-9bb6-b2b04ac241ff","Type":"ContainerDied","Data":"c47dbd5ffaa8525e40242beea1eba51d121b4c772ba1dfa755938c3d704100fe"} Oct 06 10:10:57 crc kubenswrapper[4824]: I1006 10:10:57.885217 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-7666f86cb5-qqdq7" podStartSLOduration=5.885180321 podStartE2EDuration="5.885180321s" podCreationTimestamp="2025-10-06 10:10:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:10:57.880253197 +0000 UTC m=+767.244676098" watchObservedRunningTime="2025-10-06 10:10:57.885180321 +0000 UTC m=+767.249603212" Oct 06 10:10:58 crc kubenswrapper[4824]: I1006 10:10:58.086132 4824 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 06 10:10:58 crc kubenswrapper[4824]: I1006 10:10:58.863950 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-57vkd" event={"ID":"b4856c75-71c0-4f10-9bb6-b2b04ac241ff","Type":"ContainerStarted","Data":"01befbd5ab369bac7170ad64c0a007ef072acd2f32422e0f03501b40241de676"} Oct 06 10:10:58 crc kubenswrapper[4824]: I1006 10:10:58.870404 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-7666f86cb5-qqdq7" Oct 06 10:10:58 crc kubenswrapper[4824]: I1006 10:10:58.916349 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-57vkd" podStartSLOduration=3.439140161 podStartE2EDuration="5.916317128s" podCreationTimestamp="2025-10-06 10:10:53 +0000 UTC" firstStartedPulling="2025-10-06 10:10:55.819225782 +0000 UTC m=+765.183648643" lastFinishedPulling="2025-10-06 10:10:58.296402739 +0000 UTC m=+767.660825610" observedRunningTime="2025-10-06 10:10:58.893962871 +0000 UTC m=+768.258385742" watchObservedRunningTime="2025-10-06 10:10:58.916317128 +0000 UTC m=+768.280739989" Oct 06 10:11:04 crc kubenswrapper[4824]: I1006 10:11:04.238615 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-57vkd" Oct 06 10:11:04 crc kubenswrapper[4824]: I1006 10:11:04.239107 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-57vkd" Oct 06 10:11:04 crc kubenswrapper[4824]: I1006 10:11:04.312792 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-57vkd" Oct 06 10:11:04 crc kubenswrapper[4824]: I1006 10:11:04.970895 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-57vkd" Oct 06 10:11:05 crc kubenswrapper[4824]: I1006 10:11:05.046111 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-57vkd"] Oct 06 10:11:06 crc kubenswrapper[4824]: I1006 10:11:06.032409 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-57744c8fb4-85v7z" Oct 06 10:11:06 crc kubenswrapper[4824]: I1006 10:11:06.922776 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-57vkd" podUID="b4856c75-71c0-4f10-9bb6-b2b04ac241ff" containerName="registry-server" containerID="cri-o://01befbd5ab369bac7170ad64c0a007ef072acd2f32422e0f03501b40241de676" gracePeriod=2 Oct 06 10:11:06 crc kubenswrapper[4824]: I1006 10:11:06.927777 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-x6lsw"] Oct 06 10:11:06 crc kubenswrapper[4824]: I1006 10:11:06.935820 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-x6lsw" Oct 06 10:11:06 crc kubenswrapper[4824]: I1006 10:11:06.940710 4824 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-fqjd4" Oct 06 10:11:06 crc kubenswrapper[4824]: I1006 10:11:06.942956 4824 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Oct 06 10:11:06 crc kubenswrapper[4824]: I1006 10:11:06.944864 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Oct 06 10:11:06 crc kubenswrapper[4824]: I1006 10:11:06.947429 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-d9kq5"] Oct 06 10:11:06 crc kubenswrapper[4824]: I1006 10:11:06.948235 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-d9kq5" Oct 06 10:11:06 crc kubenswrapper[4824]: I1006 10:11:06.951064 4824 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Oct 06 10:11:06 crc kubenswrapper[4824]: I1006 10:11:06.951498 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/2f2ec220-4ca6-4b43-9e89-fdae758ab063-metrics\") pod \"frr-k8s-x6lsw\" (UID: \"2f2ec220-4ca6-4b43-9e89-fdae758ab063\") " pod="metallb-system/frr-k8s-x6lsw" Oct 06 10:11:06 crc kubenswrapper[4824]: I1006 10:11:06.951558 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e0836691-2fe5-4e7c-a374-abe71aeb2278-cert\") pod \"frr-k8s-webhook-server-64bf5d555-d9kq5\" (UID: \"e0836691-2fe5-4e7c-a374-abe71aeb2278\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-d9kq5" Oct 06 10:11:06 crc kubenswrapper[4824]: I1006 10:11:06.951586 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7gc96\" (UniqueName: \"kubernetes.io/projected/e0836691-2fe5-4e7c-a374-abe71aeb2278-kube-api-access-7gc96\") pod \"frr-k8s-webhook-server-64bf5d555-d9kq5\" (UID: \"e0836691-2fe5-4e7c-a374-abe71aeb2278\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-d9kq5" Oct 06 10:11:06 crc kubenswrapper[4824]: I1006 10:11:06.951634 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/2f2ec220-4ca6-4b43-9e89-fdae758ab063-reloader\") pod \"frr-k8s-x6lsw\" (UID: \"2f2ec220-4ca6-4b43-9e89-fdae758ab063\") " pod="metallb-system/frr-k8s-x6lsw" Oct 06 10:11:06 crc kubenswrapper[4824]: I1006 10:11:06.951656 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/2f2ec220-4ca6-4b43-9e89-fdae758ab063-frr-conf\") pod \"frr-k8s-x6lsw\" (UID: \"2f2ec220-4ca6-4b43-9e89-fdae758ab063\") " pod="metallb-system/frr-k8s-x6lsw" Oct 06 10:11:06 crc kubenswrapper[4824]: I1006 10:11:06.951675 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2f2ec220-4ca6-4b43-9e89-fdae758ab063-metrics-certs\") pod \"frr-k8s-x6lsw\" (UID: \"2f2ec220-4ca6-4b43-9e89-fdae758ab063\") " pod="metallb-system/frr-k8s-x6lsw" Oct 06 10:11:06 crc kubenswrapper[4824]: I1006 10:11:06.951743 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/2f2ec220-4ca6-4b43-9e89-fdae758ab063-frr-startup\") pod \"frr-k8s-x6lsw\" (UID: \"2f2ec220-4ca6-4b43-9e89-fdae758ab063\") " pod="metallb-system/frr-k8s-x6lsw" Oct 06 10:11:06 crc kubenswrapper[4824]: I1006 10:11:06.951771 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5czw6\" (UniqueName: \"kubernetes.io/projected/2f2ec220-4ca6-4b43-9e89-fdae758ab063-kube-api-access-5czw6\") pod \"frr-k8s-x6lsw\" (UID: \"2f2ec220-4ca6-4b43-9e89-fdae758ab063\") " pod="metallb-system/frr-k8s-x6lsw" Oct 06 10:11:06 crc kubenswrapper[4824]: I1006 10:11:06.951796 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/2f2ec220-4ca6-4b43-9e89-fdae758ab063-frr-sockets\") pod \"frr-k8s-x6lsw\" (UID: \"2f2ec220-4ca6-4b43-9e89-fdae758ab063\") " pod="metallb-system/frr-k8s-x6lsw" Oct 06 10:11:06 crc kubenswrapper[4824]: I1006 10:11:06.974752 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-d9kq5"] Oct 06 10:11:07 crc kubenswrapper[4824]: I1006 10:11:07.053410 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/2f2ec220-4ca6-4b43-9e89-fdae758ab063-reloader\") pod \"frr-k8s-x6lsw\" (UID: \"2f2ec220-4ca6-4b43-9e89-fdae758ab063\") " pod="metallb-system/frr-k8s-x6lsw" Oct 06 10:11:07 crc kubenswrapper[4824]: I1006 10:11:07.053698 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/2f2ec220-4ca6-4b43-9e89-fdae758ab063-frr-conf\") pod \"frr-k8s-x6lsw\" (UID: \"2f2ec220-4ca6-4b43-9e89-fdae758ab063\") " pod="metallb-system/frr-k8s-x6lsw" Oct 06 10:11:07 crc kubenswrapper[4824]: I1006 10:11:07.053717 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2f2ec220-4ca6-4b43-9e89-fdae758ab063-metrics-certs\") pod \"frr-k8s-x6lsw\" (UID: \"2f2ec220-4ca6-4b43-9e89-fdae758ab063\") " pod="metallb-system/frr-k8s-x6lsw" Oct 06 10:11:07 crc kubenswrapper[4824]: I1006 10:11:07.053772 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/2f2ec220-4ca6-4b43-9e89-fdae758ab063-frr-startup\") pod \"frr-k8s-x6lsw\" (UID: \"2f2ec220-4ca6-4b43-9e89-fdae758ab063\") " pod="metallb-system/frr-k8s-x6lsw" Oct 06 10:11:07 crc kubenswrapper[4824]: I1006 10:11:07.053791 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5czw6\" (UniqueName: \"kubernetes.io/projected/2f2ec220-4ca6-4b43-9e89-fdae758ab063-kube-api-access-5czw6\") pod \"frr-k8s-x6lsw\" (UID: \"2f2ec220-4ca6-4b43-9e89-fdae758ab063\") " pod="metallb-system/frr-k8s-x6lsw" Oct 06 10:11:07 crc kubenswrapper[4824]: I1006 10:11:07.053808 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/2f2ec220-4ca6-4b43-9e89-fdae758ab063-frr-sockets\") pod \"frr-k8s-x6lsw\" (UID: \"2f2ec220-4ca6-4b43-9e89-fdae758ab063\") " pod="metallb-system/frr-k8s-x6lsw" Oct 06 10:11:07 crc kubenswrapper[4824]: I1006 10:11:07.053833 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/2f2ec220-4ca6-4b43-9e89-fdae758ab063-metrics\") pod \"frr-k8s-x6lsw\" (UID: \"2f2ec220-4ca6-4b43-9e89-fdae758ab063\") " pod="metallb-system/frr-k8s-x6lsw" Oct 06 10:11:07 crc kubenswrapper[4824]: I1006 10:11:07.053854 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e0836691-2fe5-4e7c-a374-abe71aeb2278-cert\") pod \"frr-k8s-webhook-server-64bf5d555-d9kq5\" (UID: \"e0836691-2fe5-4e7c-a374-abe71aeb2278\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-d9kq5" Oct 06 10:11:07 crc kubenswrapper[4824]: I1006 10:11:07.053872 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7gc96\" (UniqueName: \"kubernetes.io/projected/e0836691-2fe5-4e7c-a374-abe71aeb2278-kube-api-access-7gc96\") pod \"frr-k8s-webhook-server-64bf5d555-d9kq5\" (UID: \"e0836691-2fe5-4e7c-a374-abe71aeb2278\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-d9kq5" Oct 06 10:11:07 crc kubenswrapper[4824]: I1006 10:11:07.055020 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/2f2ec220-4ca6-4b43-9e89-fdae758ab063-reloader\") pod \"frr-k8s-x6lsw\" (UID: \"2f2ec220-4ca6-4b43-9e89-fdae758ab063\") " pod="metallb-system/frr-k8s-x6lsw" Oct 06 10:11:07 crc kubenswrapper[4824]: I1006 10:11:07.055250 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/2f2ec220-4ca6-4b43-9e89-fdae758ab063-frr-conf\") pod \"frr-k8s-x6lsw\" (UID: \"2f2ec220-4ca6-4b43-9e89-fdae758ab063\") " pod="metallb-system/frr-k8s-x6lsw" Oct 06 10:11:07 crc kubenswrapper[4824]: I1006 10:11:07.056413 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/2f2ec220-4ca6-4b43-9e89-fdae758ab063-metrics\") pod \"frr-k8s-x6lsw\" (UID: \"2f2ec220-4ca6-4b43-9e89-fdae758ab063\") " pod="metallb-system/frr-k8s-x6lsw" Oct 06 10:11:07 crc kubenswrapper[4824]: I1006 10:11:07.056679 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/2f2ec220-4ca6-4b43-9e89-fdae758ab063-frr-sockets\") pod \"frr-k8s-x6lsw\" (UID: \"2f2ec220-4ca6-4b43-9e89-fdae758ab063\") " pod="metallb-system/frr-k8s-x6lsw" Oct 06 10:11:07 crc kubenswrapper[4824]: E1006 10:11:07.056771 4824 secret.go:188] Couldn't get secret metallb-system/frr-k8s-webhook-server-cert: secret "frr-k8s-webhook-server-cert" not found Oct 06 10:11:07 crc kubenswrapper[4824]: E1006 10:11:07.056824 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e0836691-2fe5-4e7c-a374-abe71aeb2278-cert podName:e0836691-2fe5-4e7c-a374-abe71aeb2278 nodeName:}" failed. No retries permitted until 2025-10-06 10:11:07.556804606 +0000 UTC m=+776.921227467 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/e0836691-2fe5-4e7c-a374-abe71aeb2278-cert") pod "frr-k8s-webhook-server-64bf5d555-d9kq5" (UID: "e0836691-2fe5-4e7c-a374-abe71aeb2278") : secret "frr-k8s-webhook-server-cert" not found Oct 06 10:11:07 crc kubenswrapper[4824]: I1006 10:11:07.056972 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/2f2ec220-4ca6-4b43-9e89-fdae758ab063-frr-startup\") pod \"frr-k8s-x6lsw\" (UID: \"2f2ec220-4ca6-4b43-9e89-fdae758ab063\") " pod="metallb-system/frr-k8s-x6lsw" Oct 06 10:11:07 crc kubenswrapper[4824]: I1006 10:11:07.070137 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2f2ec220-4ca6-4b43-9e89-fdae758ab063-metrics-certs\") pod \"frr-k8s-x6lsw\" (UID: \"2f2ec220-4ca6-4b43-9e89-fdae758ab063\") " pod="metallb-system/frr-k8s-x6lsw" Oct 06 10:11:07 crc kubenswrapper[4824]: I1006 10:11:07.084518 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-tcgrw"] Oct 06 10:11:07 crc kubenswrapper[4824]: I1006 10:11:07.085593 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-tcgrw" Oct 06 10:11:07 crc kubenswrapper[4824]: I1006 10:11:07.089583 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5czw6\" (UniqueName: \"kubernetes.io/projected/2f2ec220-4ca6-4b43-9e89-fdae758ab063-kube-api-access-5czw6\") pod \"frr-k8s-x6lsw\" (UID: \"2f2ec220-4ca6-4b43-9e89-fdae758ab063\") " pod="metallb-system/frr-k8s-x6lsw" Oct 06 10:11:07 crc kubenswrapper[4824]: I1006 10:11:07.090095 4824 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-x4kh9" Oct 06 10:11:07 crc kubenswrapper[4824]: I1006 10:11:07.090285 4824 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Oct 06 10:11:07 crc kubenswrapper[4824]: I1006 10:11:07.090467 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Oct 06 10:11:07 crc kubenswrapper[4824]: I1006 10:11:07.090576 4824 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Oct 06 10:11:07 crc kubenswrapper[4824]: I1006 10:11:07.100713 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7gc96\" (UniqueName: \"kubernetes.io/projected/e0836691-2fe5-4e7c-a374-abe71aeb2278-kube-api-access-7gc96\") pod \"frr-k8s-webhook-server-64bf5d555-d9kq5\" (UID: \"e0836691-2fe5-4e7c-a374-abe71aeb2278\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-d9kq5" Oct 06 10:11:07 crc kubenswrapper[4824]: I1006 10:11:07.118991 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-68d546b9d8-sldcd"] Oct 06 10:11:07 crc kubenswrapper[4824]: I1006 10:11:07.119891 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-68d546b9d8-sldcd" Oct 06 10:11:07 crc kubenswrapper[4824]: I1006 10:11:07.125147 4824 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Oct 06 10:11:07 crc kubenswrapper[4824]: I1006 10:11:07.156540 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/129eb734-6e0d-4785-b25a-4c68af80d8c6-metrics-certs\") pod \"controller-68d546b9d8-sldcd\" (UID: \"129eb734-6e0d-4785-b25a-4c68af80d8c6\") " pod="metallb-system/controller-68d546b9d8-sldcd" Oct 06 10:11:07 crc kubenswrapper[4824]: I1006 10:11:07.156602 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/5c0a8aa4-2b23-4265-8b07-e075b355fc6c-memberlist\") pod \"speaker-tcgrw\" (UID: \"5c0a8aa4-2b23-4265-8b07-e075b355fc6c\") " pod="metallb-system/speaker-tcgrw" Oct 06 10:11:07 crc kubenswrapper[4824]: I1006 10:11:07.156636 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5c0a8aa4-2b23-4265-8b07-e075b355fc6c-metrics-certs\") pod \"speaker-tcgrw\" (UID: \"5c0a8aa4-2b23-4265-8b07-e075b355fc6c\") " pod="metallb-system/speaker-tcgrw" Oct 06 10:11:07 crc kubenswrapper[4824]: I1006 10:11:07.156655 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/129eb734-6e0d-4785-b25a-4c68af80d8c6-cert\") pod \"controller-68d546b9d8-sldcd\" (UID: \"129eb734-6e0d-4785-b25a-4c68af80d8c6\") " pod="metallb-system/controller-68d546b9d8-sldcd" Oct 06 10:11:07 crc kubenswrapper[4824]: I1006 10:11:07.156713 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-88x2n\" (UniqueName: \"kubernetes.io/projected/5c0a8aa4-2b23-4265-8b07-e075b355fc6c-kube-api-access-88x2n\") pod \"speaker-tcgrw\" (UID: \"5c0a8aa4-2b23-4265-8b07-e075b355fc6c\") " pod="metallb-system/speaker-tcgrw" Oct 06 10:11:07 crc kubenswrapper[4824]: I1006 10:11:07.156762 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r6shz\" (UniqueName: \"kubernetes.io/projected/129eb734-6e0d-4785-b25a-4c68af80d8c6-kube-api-access-r6shz\") pod \"controller-68d546b9d8-sldcd\" (UID: \"129eb734-6e0d-4785-b25a-4c68af80d8c6\") " pod="metallb-system/controller-68d546b9d8-sldcd" Oct 06 10:11:07 crc kubenswrapper[4824]: I1006 10:11:07.156778 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/5c0a8aa4-2b23-4265-8b07-e075b355fc6c-metallb-excludel2\") pod \"speaker-tcgrw\" (UID: \"5c0a8aa4-2b23-4265-8b07-e075b355fc6c\") " pod="metallb-system/speaker-tcgrw" Oct 06 10:11:07 crc kubenswrapper[4824]: I1006 10:11:07.200380 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-68d546b9d8-sldcd"] Oct 06 10:11:07 crc kubenswrapper[4824]: I1006 10:11:07.258544 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-88x2n\" (UniqueName: \"kubernetes.io/projected/5c0a8aa4-2b23-4265-8b07-e075b355fc6c-kube-api-access-88x2n\") pod \"speaker-tcgrw\" (UID: \"5c0a8aa4-2b23-4265-8b07-e075b355fc6c\") " pod="metallb-system/speaker-tcgrw" Oct 06 10:11:07 crc kubenswrapper[4824]: I1006 10:11:07.258613 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r6shz\" (UniqueName: \"kubernetes.io/projected/129eb734-6e0d-4785-b25a-4c68af80d8c6-kube-api-access-r6shz\") pod \"controller-68d546b9d8-sldcd\" (UID: \"129eb734-6e0d-4785-b25a-4c68af80d8c6\") " pod="metallb-system/controller-68d546b9d8-sldcd" Oct 06 10:11:07 crc kubenswrapper[4824]: I1006 10:11:07.258640 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/5c0a8aa4-2b23-4265-8b07-e075b355fc6c-metallb-excludel2\") pod \"speaker-tcgrw\" (UID: \"5c0a8aa4-2b23-4265-8b07-e075b355fc6c\") " pod="metallb-system/speaker-tcgrw" Oct 06 10:11:07 crc kubenswrapper[4824]: I1006 10:11:07.258689 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/129eb734-6e0d-4785-b25a-4c68af80d8c6-metrics-certs\") pod \"controller-68d546b9d8-sldcd\" (UID: \"129eb734-6e0d-4785-b25a-4c68af80d8c6\") " pod="metallb-system/controller-68d546b9d8-sldcd" Oct 06 10:11:07 crc kubenswrapper[4824]: I1006 10:11:07.258715 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/5c0a8aa4-2b23-4265-8b07-e075b355fc6c-memberlist\") pod \"speaker-tcgrw\" (UID: \"5c0a8aa4-2b23-4265-8b07-e075b355fc6c\") " pod="metallb-system/speaker-tcgrw" Oct 06 10:11:07 crc kubenswrapper[4824]: I1006 10:11:07.258757 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5c0a8aa4-2b23-4265-8b07-e075b355fc6c-metrics-certs\") pod \"speaker-tcgrw\" (UID: \"5c0a8aa4-2b23-4265-8b07-e075b355fc6c\") " pod="metallb-system/speaker-tcgrw" Oct 06 10:11:07 crc kubenswrapper[4824]: I1006 10:11:07.258792 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/129eb734-6e0d-4785-b25a-4c68af80d8c6-cert\") pod \"controller-68d546b9d8-sldcd\" (UID: \"129eb734-6e0d-4785-b25a-4c68af80d8c6\") " pod="metallb-system/controller-68d546b9d8-sldcd" Oct 06 10:11:07 crc kubenswrapper[4824]: E1006 10:11:07.259243 4824 secret.go:188] Couldn't get secret metallb-system/speaker-certs-secret: secret "speaker-certs-secret" not found Oct 06 10:11:07 crc kubenswrapper[4824]: E1006 10:11:07.259328 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5c0a8aa4-2b23-4265-8b07-e075b355fc6c-metrics-certs podName:5c0a8aa4-2b23-4265-8b07-e075b355fc6c nodeName:}" failed. No retries permitted until 2025-10-06 10:11:07.759307773 +0000 UTC m=+777.123730634 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/5c0a8aa4-2b23-4265-8b07-e075b355fc6c-metrics-certs") pod "speaker-tcgrw" (UID: "5c0a8aa4-2b23-4265-8b07-e075b355fc6c") : secret "speaker-certs-secret" not found Oct 06 10:11:07 crc kubenswrapper[4824]: E1006 10:11:07.259243 4824 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Oct 06 10:11:07 crc kubenswrapper[4824]: I1006 10:11:07.259418 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-x6lsw" Oct 06 10:11:07 crc kubenswrapper[4824]: E1006 10:11:07.259509 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5c0a8aa4-2b23-4265-8b07-e075b355fc6c-memberlist podName:5c0a8aa4-2b23-4265-8b07-e075b355fc6c nodeName:}" failed. No retries permitted until 2025-10-06 10:11:07.759477567 +0000 UTC m=+777.123900438 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/5c0a8aa4-2b23-4265-8b07-e075b355fc6c-memberlist") pod "speaker-tcgrw" (UID: "5c0a8aa4-2b23-4265-8b07-e075b355fc6c") : secret "metallb-memberlist" not found Oct 06 10:11:07 crc kubenswrapper[4824]: I1006 10:11:07.260584 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/5c0a8aa4-2b23-4265-8b07-e075b355fc6c-metallb-excludel2\") pod \"speaker-tcgrw\" (UID: \"5c0a8aa4-2b23-4265-8b07-e075b355fc6c\") " pod="metallb-system/speaker-tcgrw" Oct 06 10:11:07 crc kubenswrapper[4824]: I1006 10:11:07.264811 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/129eb734-6e0d-4785-b25a-4c68af80d8c6-metrics-certs\") pod \"controller-68d546b9d8-sldcd\" (UID: \"129eb734-6e0d-4785-b25a-4c68af80d8c6\") " pod="metallb-system/controller-68d546b9d8-sldcd" Oct 06 10:11:07 crc kubenswrapper[4824]: I1006 10:11:07.270389 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/129eb734-6e0d-4785-b25a-4c68af80d8c6-cert\") pod \"controller-68d546b9d8-sldcd\" (UID: \"129eb734-6e0d-4785-b25a-4c68af80d8c6\") " pod="metallb-system/controller-68d546b9d8-sldcd" Oct 06 10:11:07 crc kubenswrapper[4824]: I1006 10:11:07.303043 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r6shz\" (UniqueName: \"kubernetes.io/projected/129eb734-6e0d-4785-b25a-4c68af80d8c6-kube-api-access-r6shz\") pod \"controller-68d546b9d8-sldcd\" (UID: \"129eb734-6e0d-4785-b25a-4c68af80d8c6\") " pod="metallb-system/controller-68d546b9d8-sldcd" Oct 06 10:11:07 crc kubenswrapper[4824]: I1006 10:11:07.303586 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-88x2n\" (UniqueName: \"kubernetes.io/projected/5c0a8aa4-2b23-4265-8b07-e075b355fc6c-kube-api-access-88x2n\") pod \"speaker-tcgrw\" (UID: \"5c0a8aa4-2b23-4265-8b07-e075b355fc6c\") " pod="metallb-system/speaker-tcgrw" Oct 06 10:11:07 crc kubenswrapper[4824]: I1006 10:11:07.495293 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-68d546b9d8-sldcd" Oct 06 10:11:07 crc kubenswrapper[4824]: I1006 10:11:07.568137 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e0836691-2fe5-4e7c-a374-abe71aeb2278-cert\") pod \"frr-k8s-webhook-server-64bf5d555-d9kq5\" (UID: \"e0836691-2fe5-4e7c-a374-abe71aeb2278\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-d9kq5" Oct 06 10:11:07 crc kubenswrapper[4824]: I1006 10:11:07.574580 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e0836691-2fe5-4e7c-a374-abe71aeb2278-cert\") pod \"frr-k8s-webhook-server-64bf5d555-d9kq5\" (UID: \"e0836691-2fe5-4e7c-a374-abe71aeb2278\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-d9kq5" Oct 06 10:11:07 crc kubenswrapper[4824]: I1006 10:11:07.771531 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/5c0a8aa4-2b23-4265-8b07-e075b355fc6c-memberlist\") pod \"speaker-tcgrw\" (UID: \"5c0a8aa4-2b23-4265-8b07-e075b355fc6c\") " pod="metallb-system/speaker-tcgrw" Oct 06 10:11:07 crc kubenswrapper[4824]: I1006 10:11:07.771608 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5c0a8aa4-2b23-4265-8b07-e075b355fc6c-metrics-certs\") pod \"speaker-tcgrw\" (UID: \"5c0a8aa4-2b23-4265-8b07-e075b355fc6c\") " pod="metallb-system/speaker-tcgrw" Oct 06 10:11:07 crc kubenswrapper[4824]: E1006 10:11:07.771846 4824 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Oct 06 10:11:07 crc kubenswrapper[4824]: E1006 10:11:07.771999 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5c0a8aa4-2b23-4265-8b07-e075b355fc6c-memberlist podName:5c0a8aa4-2b23-4265-8b07-e075b355fc6c nodeName:}" failed. No retries permitted until 2025-10-06 10:11:08.771946439 +0000 UTC m=+778.136369480 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/5c0a8aa4-2b23-4265-8b07-e075b355fc6c-memberlist") pod "speaker-tcgrw" (UID: "5c0a8aa4-2b23-4265-8b07-e075b355fc6c") : secret "metallb-memberlist" not found Oct 06 10:11:07 crc kubenswrapper[4824]: I1006 10:11:07.776435 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5c0a8aa4-2b23-4265-8b07-e075b355fc6c-metrics-certs\") pod \"speaker-tcgrw\" (UID: \"5c0a8aa4-2b23-4265-8b07-e075b355fc6c\") " pod="metallb-system/speaker-tcgrw" Oct 06 10:11:07 crc kubenswrapper[4824]: I1006 10:11:07.865799 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-d9kq5" Oct 06 10:11:07 crc kubenswrapper[4824]: I1006 10:11:07.931041 4824 generic.go:334] "Generic (PLEG): container finished" podID="b4856c75-71c0-4f10-9bb6-b2b04ac241ff" containerID="01befbd5ab369bac7170ad64c0a007ef072acd2f32422e0f03501b40241de676" exitCode=0 Oct 06 10:11:07 crc kubenswrapper[4824]: I1006 10:11:07.931104 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-57vkd" event={"ID":"b4856c75-71c0-4f10-9bb6-b2b04ac241ff","Type":"ContainerDied","Data":"01befbd5ab369bac7170ad64c0a007ef072acd2f32422e0f03501b40241de676"} Oct 06 10:11:07 crc kubenswrapper[4824]: I1006 10:11:07.931890 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-x6lsw" event={"ID":"2f2ec220-4ca6-4b43-9e89-fdae758ab063","Type":"ContainerStarted","Data":"b535e82ec941be8285072852bb1a9626794da0cb1fe15ac2cc9ddbd648b51379"} Oct 06 10:11:08 crc kubenswrapper[4824]: I1006 10:11:08.032833 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-68d546b9d8-sldcd"] Oct 06 10:11:08 crc kubenswrapper[4824]: W1006 10:11:08.053918 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod129eb734_6e0d_4785_b25a_4c68af80d8c6.slice/crio-97007f32f6f2736b435e5eebefd13c4f040adf096f948ccefd75560753512218 WatchSource:0}: Error finding container 97007f32f6f2736b435e5eebefd13c4f040adf096f948ccefd75560753512218: Status 404 returned error can't find the container with id 97007f32f6f2736b435e5eebefd13c4f040adf096f948ccefd75560753512218 Oct 06 10:11:08 crc kubenswrapper[4824]: I1006 10:11:08.248554 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-57vkd" Oct 06 10:11:08 crc kubenswrapper[4824]: I1006 10:11:08.381741 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b4856c75-71c0-4f10-9bb6-b2b04ac241ff-utilities\") pod \"b4856c75-71c0-4f10-9bb6-b2b04ac241ff\" (UID: \"b4856c75-71c0-4f10-9bb6-b2b04ac241ff\") " Oct 06 10:11:08 crc kubenswrapper[4824]: I1006 10:11:08.382420 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b4856c75-71c0-4f10-9bb6-b2b04ac241ff-catalog-content\") pod \"b4856c75-71c0-4f10-9bb6-b2b04ac241ff\" (UID: \"b4856c75-71c0-4f10-9bb6-b2b04ac241ff\") " Oct 06 10:11:08 crc kubenswrapper[4824]: I1006 10:11:08.382494 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hl598\" (UniqueName: \"kubernetes.io/projected/b4856c75-71c0-4f10-9bb6-b2b04ac241ff-kube-api-access-hl598\") pod \"b4856c75-71c0-4f10-9bb6-b2b04ac241ff\" (UID: \"b4856c75-71c0-4f10-9bb6-b2b04ac241ff\") " Oct 06 10:11:08 crc kubenswrapper[4824]: I1006 10:11:08.383459 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b4856c75-71c0-4f10-9bb6-b2b04ac241ff-utilities" (OuterVolumeSpecName: "utilities") pod "b4856c75-71c0-4f10-9bb6-b2b04ac241ff" (UID: "b4856c75-71c0-4f10-9bb6-b2b04ac241ff"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:11:08 crc kubenswrapper[4824]: I1006 10:11:08.391121 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b4856c75-71c0-4f10-9bb6-b2b04ac241ff-kube-api-access-hl598" (OuterVolumeSpecName: "kube-api-access-hl598") pod "b4856c75-71c0-4f10-9bb6-b2b04ac241ff" (UID: "b4856c75-71c0-4f10-9bb6-b2b04ac241ff"). InnerVolumeSpecName "kube-api-access-hl598". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:11:08 crc kubenswrapper[4824]: I1006 10:11:08.409501 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-d9kq5"] Oct 06 10:11:08 crc kubenswrapper[4824]: I1006 10:11:08.445455 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b4856c75-71c0-4f10-9bb6-b2b04ac241ff-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b4856c75-71c0-4f10-9bb6-b2b04ac241ff" (UID: "b4856c75-71c0-4f10-9bb6-b2b04ac241ff"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:11:08 crc kubenswrapper[4824]: I1006 10:11:08.484510 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b4856c75-71c0-4f10-9bb6-b2b04ac241ff-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 06 10:11:08 crc kubenswrapper[4824]: I1006 10:11:08.484561 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hl598\" (UniqueName: \"kubernetes.io/projected/b4856c75-71c0-4f10-9bb6-b2b04ac241ff-kube-api-access-hl598\") on node \"crc\" DevicePath \"\"" Oct 06 10:11:08 crc kubenswrapper[4824]: I1006 10:11:08.484573 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b4856c75-71c0-4f10-9bb6-b2b04ac241ff-utilities\") on node \"crc\" DevicePath \"\"" Oct 06 10:11:08 crc kubenswrapper[4824]: I1006 10:11:08.791938 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/5c0a8aa4-2b23-4265-8b07-e075b355fc6c-memberlist\") pod \"speaker-tcgrw\" (UID: \"5c0a8aa4-2b23-4265-8b07-e075b355fc6c\") " pod="metallb-system/speaker-tcgrw" Oct 06 10:11:08 crc kubenswrapper[4824]: I1006 10:11:08.797065 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/5c0a8aa4-2b23-4265-8b07-e075b355fc6c-memberlist\") pod \"speaker-tcgrw\" (UID: \"5c0a8aa4-2b23-4265-8b07-e075b355fc6c\") " pod="metallb-system/speaker-tcgrw" Oct 06 10:11:08 crc kubenswrapper[4824]: I1006 10:11:08.938920 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-d9kq5" event={"ID":"e0836691-2fe5-4e7c-a374-abe71aeb2278","Type":"ContainerStarted","Data":"745d17a8b77e3e3a705476194a94b99758faf5c69595268f839c1708fb93b45b"} Oct 06 10:11:08 crc kubenswrapper[4824]: I1006 10:11:08.940730 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-sldcd" event={"ID":"129eb734-6e0d-4785-b25a-4c68af80d8c6","Type":"ContainerStarted","Data":"d7ba1c824c19b0dff995e077889bae1c4007bdbf1269d1642a999c4d105b7578"} Oct 06 10:11:08 crc kubenswrapper[4824]: I1006 10:11:08.940757 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-sldcd" event={"ID":"129eb734-6e0d-4785-b25a-4c68af80d8c6","Type":"ContainerStarted","Data":"461755ff371387770a0a1d483e3dd99075e14f1ba6421495246a6304ed28c799"} Oct 06 10:11:08 crc kubenswrapper[4824]: I1006 10:11:08.940767 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-sldcd" event={"ID":"129eb734-6e0d-4785-b25a-4c68af80d8c6","Type":"ContainerStarted","Data":"97007f32f6f2736b435e5eebefd13c4f040adf096f948ccefd75560753512218"} Oct 06 10:11:08 crc kubenswrapper[4824]: I1006 10:11:08.941658 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-68d546b9d8-sldcd" Oct 06 10:11:08 crc kubenswrapper[4824]: I1006 10:11:08.946233 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-57vkd" event={"ID":"b4856c75-71c0-4f10-9bb6-b2b04ac241ff","Type":"ContainerDied","Data":"9839544bb03556ad10a90437a5af64c701fa954f1d9661ff579b5e2529314751"} Oct 06 10:11:08 crc kubenswrapper[4824]: I1006 10:11:08.946267 4824 scope.go:117] "RemoveContainer" containerID="01befbd5ab369bac7170ad64c0a007ef072acd2f32422e0f03501b40241de676" Oct 06 10:11:08 crc kubenswrapper[4824]: I1006 10:11:08.946742 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-57vkd" Oct 06 10:11:08 crc kubenswrapper[4824]: I1006 10:11:08.960742 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-68d546b9d8-sldcd" podStartSLOduration=1.960721945 podStartE2EDuration="1.960721945s" podCreationTimestamp="2025-10-06 10:11:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:11:08.958922451 +0000 UTC m=+778.323345322" watchObservedRunningTime="2025-10-06 10:11:08.960721945 +0000 UTC m=+778.325144806" Oct 06 10:11:08 crc kubenswrapper[4824]: I1006 10:11:08.969204 4824 scope.go:117] "RemoveContainer" containerID="c47dbd5ffaa8525e40242beea1eba51d121b4c772ba1dfa755938c3d704100fe" Oct 06 10:11:08 crc kubenswrapper[4824]: I1006 10:11:08.992356 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-tcgrw" Oct 06 10:11:09 crc kubenswrapper[4824]: I1006 10:11:09.038273 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-57vkd"] Oct 06 10:11:09 crc kubenswrapper[4824]: I1006 10:11:09.048646 4824 scope.go:117] "RemoveContainer" containerID="62713eb13924cbf2fbd78523f9397c8ecc313bcb4cde994132ecf0086f7e05bb" Oct 06 10:11:09 crc kubenswrapper[4824]: I1006 10:11:09.056057 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-57vkd"] Oct 06 10:11:09 crc kubenswrapper[4824]: I1006 10:11:09.288101 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b4856c75-71c0-4f10-9bb6-b2b04ac241ff" path="/var/lib/kubelet/pods/b4856c75-71c0-4f10-9bb6-b2b04ac241ff/volumes" Oct 06 10:11:09 crc kubenswrapper[4824]: I1006 10:11:09.960022 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-tcgrw" event={"ID":"5c0a8aa4-2b23-4265-8b07-e075b355fc6c","Type":"ContainerStarted","Data":"6545a7f496d0661a23076a5b632caaed6acf8fef86ec37df9a9858b297ad0bfa"} Oct 06 10:11:09 crc kubenswrapper[4824]: I1006 10:11:09.960370 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-tcgrw" event={"ID":"5c0a8aa4-2b23-4265-8b07-e075b355fc6c","Type":"ContainerStarted","Data":"a69448f982c9bce9a166e13801806c753ca0c15e56986e947aedf225f718d52c"} Oct 06 10:11:09 crc kubenswrapper[4824]: I1006 10:11:09.960383 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-tcgrw" event={"ID":"5c0a8aa4-2b23-4265-8b07-e075b355fc6c","Type":"ContainerStarted","Data":"f730a070b8a829d5b16c1505a06b4f08d60e81aa44c3205c7ed5cff54aeddb2f"} Oct 06 10:11:09 crc kubenswrapper[4824]: I1006 10:11:09.960594 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-tcgrw" Oct 06 10:11:09 crc kubenswrapper[4824]: I1006 10:11:09.996220 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-tcgrw" podStartSLOduration=2.996200542 podStartE2EDuration="2.996200542s" podCreationTimestamp="2025-10-06 10:11:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:11:09.994855578 +0000 UTC m=+779.359278439" watchObservedRunningTime="2025-10-06 10:11:09.996200542 +0000 UTC m=+779.360623393" Oct 06 10:11:13 crc kubenswrapper[4824]: I1006 10:11:13.915679 4824 patch_prober.go:28] interesting pod/machine-config-daemon-khgzw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 10:11:13 crc kubenswrapper[4824]: I1006 10:11:13.916698 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 10:11:16 crc kubenswrapper[4824]: I1006 10:11:16.013759 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-d9kq5" event={"ID":"e0836691-2fe5-4e7c-a374-abe71aeb2278","Type":"ContainerStarted","Data":"4e2b2ed6540d074f4268909f0edbb1a77e2f11af5deaa14205105ee9169c6cb9"} Oct 06 10:11:16 crc kubenswrapper[4824]: I1006 10:11:16.017453 4824 generic.go:334] "Generic (PLEG): container finished" podID="2f2ec220-4ca6-4b43-9e89-fdae758ab063" containerID="dcee1e16339356e2a384333a8eea5cd310064a272eb7ba4cd3c48970d2d0e874" exitCode=0 Oct 06 10:11:16 crc kubenswrapper[4824]: I1006 10:11:16.017493 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-x6lsw" event={"ID":"2f2ec220-4ca6-4b43-9e89-fdae758ab063","Type":"ContainerDied","Data":"dcee1e16339356e2a384333a8eea5cd310064a272eb7ba4cd3c48970d2d0e874"} Oct 06 10:11:16 crc kubenswrapper[4824]: I1006 10:11:16.044150 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-d9kq5" podStartSLOduration=3.598420168 podStartE2EDuration="10.044119448s" podCreationTimestamp="2025-10-06 10:11:06 +0000 UTC" firstStartedPulling="2025-10-06 10:11:08.437080125 +0000 UTC m=+777.801502986" lastFinishedPulling="2025-10-06 10:11:14.882779405 +0000 UTC m=+784.247202266" observedRunningTime="2025-10-06 10:11:16.037804751 +0000 UTC m=+785.402227682" watchObservedRunningTime="2025-10-06 10:11:16.044119448 +0000 UTC m=+785.408542319" Oct 06 10:11:17 crc kubenswrapper[4824]: I1006 10:11:17.031451 4824 generic.go:334] "Generic (PLEG): container finished" podID="2f2ec220-4ca6-4b43-9e89-fdae758ab063" containerID="7dbf2e088a50735ced922e38a35735619d71c540d160c05b07aac19261aa4511" exitCode=0 Oct 06 10:11:17 crc kubenswrapper[4824]: I1006 10:11:17.034383 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-x6lsw" event={"ID":"2f2ec220-4ca6-4b43-9e89-fdae758ab063","Type":"ContainerDied","Data":"7dbf2e088a50735ced922e38a35735619d71c540d160c05b07aac19261aa4511"} Oct 06 10:11:17 crc kubenswrapper[4824]: I1006 10:11:17.034448 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-d9kq5" Oct 06 10:11:18 crc kubenswrapper[4824]: I1006 10:11:18.043184 4824 generic.go:334] "Generic (PLEG): container finished" podID="2f2ec220-4ca6-4b43-9e89-fdae758ab063" containerID="d28ae83519850b12a9253a68e7c1f24ed19508e895ba792c123d6a006d87353f" exitCode=0 Oct 06 10:11:18 crc kubenswrapper[4824]: I1006 10:11:18.043283 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-x6lsw" event={"ID":"2f2ec220-4ca6-4b43-9e89-fdae758ab063","Type":"ContainerDied","Data":"d28ae83519850b12a9253a68e7c1f24ed19508e895ba792c123d6a006d87353f"} Oct 06 10:11:19 crc kubenswrapper[4824]: I1006 10:11:19.063845 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-x6lsw" event={"ID":"2f2ec220-4ca6-4b43-9e89-fdae758ab063","Type":"ContainerStarted","Data":"603feb7d0f348aff85748c77595b10cdaafa65724312c04e4e48e1917c1ac8f3"} Oct 06 10:11:19 crc kubenswrapper[4824]: I1006 10:11:19.064529 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-x6lsw" event={"ID":"2f2ec220-4ca6-4b43-9e89-fdae758ab063","Type":"ContainerStarted","Data":"6de10bdcec9b49cd2da696f110b518396e5e1892fcf854e0ae1c4356eed239d5"} Oct 06 10:11:19 crc kubenswrapper[4824]: I1006 10:11:19.064554 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-x6lsw" event={"ID":"2f2ec220-4ca6-4b43-9e89-fdae758ab063","Type":"ContainerStarted","Data":"41fec70fef4dea38911896b75b976e85d4a27345c1c50cf93a0c4a05cfb3a893"} Oct 06 10:11:19 crc kubenswrapper[4824]: I1006 10:11:19.064578 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-x6lsw" event={"ID":"2f2ec220-4ca6-4b43-9e89-fdae758ab063","Type":"ContainerStarted","Data":"5c3238641d3cd5ba659a91e82c69116ed116b2fe9c75230d63b2548b1d4e69d3"} Oct 06 10:11:19 crc kubenswrapper[4824]: I1006 10:11:19.064597 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-x6lsw" event={"ID":"2f2ec220-4ca6-4b43-9e89-fdae758ab063","Type":"ContainerStarted","Data":"1e63eaecdb2f44964f92de4132ff400f3422d2d4611034c5a7d9a1d4935884e5"} Oct 06 10:11:19 crc kubenswrapper[4824]: I1006 10:11:19.917672 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-g7f78"] Oct 06 10:11:19 crc kubenswrapper[4824]: E1006 10:11:19.918024 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4856c75-71c0-4f10-9bb6-b2b04ac241ff" containerName="extract-content" Oct 06 10:11:19 crc kubenswrapper[4824]: I1006 10:11:19.918045 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4856c75-71c0-4f10-9bb6-b2b04ac241ff" containerName="extract-content" Oct 06 10:11:19 crc kubenswrapper[4824]: E1006 10:11:19.918076 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4856c75-71c0-4f10-9bb6-b2b04ac241ff" containerName="registry-server" Oct 06 10:11:19 crc kubenswrapper[4824]: I1006 10:11:19.918089 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4856c75-71c0-4f10-9bb6-b2b04ac241ff" containerName="registry-server" Oct 06 10:11:19 crc kubenswrapper[4824]: E1006 10:11:19.918114 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4856c75-71c0-4f10-9bb6-b2b04ac241ff" containerName="extract-utilities" Oct 06 10:11:19 crc kubenswrapper[4824]: I1006 10:11:19.918126 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4856c75-71c0-4f10-9bb6-b2b04ac241ff" containerName="extract-utilities" Oct 06 10:11:19 crc kubenswrapper[4824]: I1006 10:11:19.918340 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="b4856c75-71c0-4f10-9bb6-b2b04ac241ff" containerName="registry-server" Oct 06 10:11:19 crc kubenswrapper[4824]: I1006 10:11:19.919937 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-g7f78" Oct 06 10:11:19 crc kubenswrapper[4824]: I1006 10:11:19.945650 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-g7f78"] Oct 06 10:11:20 crc kubenswrapper[4824]: I1006 10:11:20.035647 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/edb02879-620e-473c-aa8a-1f8aa4fffd6d-utilities\") pod \"redhat-marketplace-g7f78\" (UID: \"edb02879-620e-473c-aa8a-1f8aa4fffd6d\") " pod="openshift-marketplace/redhat-marketplace-g7f78" Oct 06 10:11:20 crc kubenswrapper[4824]: I1006 10:11:20.035833 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m7vv6\" (UniqueName: \"kubernetes.io/projected/edb02879-620e-473c-aa8a-1f8aa4fffd6d-kube-api-access-m7vv6\") pod \"redhat-marketplace-g7f78\" (UID: \"edb02879-620e-473c-aa8a-1f8aa4fffd6d\") " pod="openshift-marketplace/redhat-marketplace-g7f78" Oct 06 10:11:20 crc kubenswrapper[4824]: I1006 10:11:20.035899 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/edb02879-620e-473c-aa8a-1f8aa4fffd6d-catalog-content\") pod \"redhat-marketplace-g7f78\" (UID: \"edb02879-620e-473c-aa8a-1f8aa4fffd6d\") " pod="openshift-marketplace/redhat-marketplace-g7f78" Oct 06 10:11:20 crc kubenswrapper[4824]: I1006 10:11:20.078434 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-x6lsw" event={"ID":"2f2ec220-4ca6-4b43-9e89-fdae758ab063","Type":"ContainerStarted","Data":"f12dc7a562500739006a565d95cd3031f4a7c258c87a9a9e3751aa070814ec91"} Oct 06 10:11:20 crc kubenswrapper[4824]: I1006 10:11:20.078629 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-x6lsw" Oct 06 10:11:20 crc kubenswrapper[4824]: I1006 10:11:20.110484 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-x6lsw" podStartSLOduration=6.63956254 podStartE2EDuration="14.1104605s" podCreationTimestamp="2025-10-06 10:11:06 +0000 UTC" firstStartedPulling="2025-10-06 10:11:07.395252941 +0000 UTC m=+776.759675802" lastFinishedPulling="2025-10-06 10:11:14.866150901 +0000 UTC m=+784.230573762" observedRunningTime="2025-10-06 10:11:20.107578108 +0000 UTC m=+789.472001019" watchObservedRunningTime="2025-10-06 10:11:20.1104605 +0000 UTC m=+789.474883371" Oct 06 10:11:20 crc kubenswrapper[4824]: I1006 10:11:20.137742 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/edb02879-620e-473c-aa8a-1f8aa4fffd6d-utilities\") pod \"redhat-marketplace-g7f78\" (UID: \"edb02879-620e-473c-aa8a-1f8aa4fffd6d\") " pod="openshift-marketplace/redhat-marketplace-g7f78" Oct 06 10:11:20 crc kubenswrapper[4824]: I1006 10:11:20.137887 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m7vv6\" (UniqueName: \"kubernetes.io/projected/edb02879-620e-473c-aa8a-1f8aa4fffd6d-kube-api-access-m7vv6\") pod \"redhat-marketplace-g7f78\" (UID: \"edb02879-620e-473c-aa8a-1f8aa4fffd6d\") " pod="openshift-marketplace/redhat-marketplace-g7f78" Oct 06 10:11:20 crc kubenswrapper[4824]: I1006 10:11:20.137933 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/edb02879-620e-473c-aa8a-1f8aa4fffd6d-catalog-content\") pod \"redhat-marketplace-g7f78\" (UID: \"edb02879-620e-473c-aa8a-1f8aa4fffd6d\") " pod="openshift-marketplace/redhat-marketplace-g7f78" Oct 06 10:11:20 crc kubenswrapper[4824]: I1006 10:11:20.139147 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/edb02879-620e-473c-aa8a-1f8aa4fffd6d-utilities\") pod \"redhat-marketplace-g7f78\" (UID: \"edb02879-620e-473c-aa8a-1f8aa4fffd6d\") " pod="openshift-marketplace/redhat-marketplace-g7f78" Oct 06 10:11:20 crc kubenswrapper[4824]: I1006 10:11:20.139337 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/edb02879-620e-473c-aa8a-1f8aa4fffd6d-catalog-content\") pod \"redhat-marketplace-g7f78\" (UID: \"edb02879-620e-473c-aa8a-1f8aa4fffd6d\") " pod="openshift-marketplace/redhat-marketplace-g7f78" Oct 06 10:11:20 crc kubenswrapper[4824]: I1006 10:11:20.172301 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m7vv6\" (UniqueName: \"kubernetes.io/projected/edb02879-620e-473c-aa8a-1f8aa4fffd6d-kube-api-access-m7vv6\") pod \"redhat-marketplace-g7f78\" (UID: \"edb02879-620e-473c-aa8a-1f8aa4fffd6d\") " pod="openshift-marketplace/redhat-marketplace-g7f78" Oct 06 10:11:20 crc kubenswrapper[4824]: I1006 10:11:20.238190 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-g7f78" Oct 06 10:11:20 crc kubenswrapper[4824]: I1006 10:11:20.749726 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-g7f78"] Oct 06 10:11:21 crc kubenswrapper[4824]: I1006 10:11:21.103380 4824 generic.go:334] "Generic (PLEG): container finished" podID="edb02879-620e-473c-aa8a-1f8aa4fffd6d" containerID="31add07254df1b10e73843970fef95066c96749b88ef9b4a677de9caa306370c" exitCode=0 Oct 06 10:11:21 crc kubenswrapper[4824]: I1006 10:11:21.104046 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-g7f78" event={"ID":"edb02879-620e-473c-aa8a-1f8aa4fffd6d","Type":"ContainerDied","Data":"31add07254df1b10e73843970fef95066c96749b88ef9b4a677de9caa306370c"} Oct 06 10:11:21 crc kubenswrapper[4824]: I1006 10:11:21.105401 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-g7f78" event={"ID":"edb02879-620e-473c-aa8a-1f8aa4fffd6d","Type":"ContainerStarted","Data":"358346d19761c77c0fcac98cf778e534913836ce8e9b8e68c47533a3444f1d72"} Oct 06 10:11:22 crc kubenswrapper[4824]: I1006 10:11:22.260447 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-x6lsw" Oct 06 10:11:22 crc kubenswrapper[4824]: I1006 10:11:22.315458 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-x6lsw" Oct 06 10:11:23 crc kubenswrapper[4824]: I1006 10:11:23.125085 4824 generic.go:334] "Generic (PLEG): container finished" podID="edb02879-620e-473c-aa8a-1f8aa4fffd6d" containerID="700676481b53609568aaa87548dee9160418feb7a77525682783a6890ec68ea3" exitCode=0 Oct 06 10:11:23 crc kubenswrapper[4824]: I1006 10:11:23.127244 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-g7f78" event={"ID":"edb02879-620e-473c-aa8a-1f8aa4fffd6d","Type":"ContainerDied","Data":"700676481b53609568aaa87548dee9160418feb7a77525682783a6890ec68ea3"} Oct 06 10:11:24 crc kubenswrapper[4824]: I1006 10:11:24.148736 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-g7f78" event={"ID":"edb02879-620e-473c-aa8a-1f8aa4fffd6d","Type":"ContainerStarted","Data":"58c143e2f9da8ac94b66c2d4288484e15ffa3f4c8eb97e55ac139a05fe4e0ee3"} Oct 06 10:11:24 crc kubenswrapper[4824]: I1006 10:11:24.177926 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-g7f78" podStartSLOduration=2.7512053610000002 podStartE2EDuration="5.177899359s" podCreationTimestamp="2025-10-06 10:11:19 +0000 UTC" firstStartedPulling="2025-10-06 10:11:21.108834832 +0000 UTC m=+790.473257743" lastFinishedPulling="2025-10-06 10:11:23.53552884 +0000 UTC m=+792.899951741" observedRunningTime="2025-10-06 10:11:24.175425227 +0000 UTC m=+793.539848158" watchObservedRunningTime="2025-10-06 10:11:24.177899359 +0000 UTC m=+793.542322230" Oct 06 10:11:27 crc kubenswrapper[4824]: I1006 10:11:27.506262 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-68d546b9d8-sldcd" Oct 06 10:11:27 crc kubenswrapper[4824]: I1006 10:11:27.876125 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-d9kq5" Oct 06 10:11:29 crc kubenswrapper[4824]: I1006 10:11:29.002084 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-tcgrw" Oct 06 10:11:30 crc kubenswrapper[4824]: I1006 10:11:30.238563 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-g7f78" Oct 06 10:11:30 crc kubenswrapper[4824]: I1006 10:11:30.238642 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-g7f78" Oct 06 10:11:30 crc kubenswrapper[4824]: I1006 10:11:30.309967 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-g7f78" Oct 06 10:11:30 crc kubenswrapper[4824]: I1006 10:11:30.984819 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-vlxgd"] Oct 06 10:11:30 crc kubenswrapper[4824]: I1006 10:11:30.986694 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vlxgd" Oct 06 10:11:31 crc kubenswrapper[4824]: I1006 10:11:31.004402 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vlxgd"] Oct 06 10:11:31 crc kubenswrapper[4824]: I1006 10:11:31.144816 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b85edf34-0286-409d-907d-b97fceccdbae-catalog-content\") pod \"redhat-operators-vlxgd\" (UID: \"b85edf34-0286-409d-907d-b97fceccdbae\") " pod="openshift-marketplace/redhat-operators-vlxgd" Oct 06 10:11:31 crc kubenswrapper[4824]: I1006 10:11:31.144930 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b85edf34-0286-409d-907d-b97fceccdbae-utilities\") pod \"redhat-operators-vlxgd\" (UID: \"b85edf34-0286-409d-907d-b97fceccdbae\") " pod="openshift-marketplace/redhat-operators-vlxgd" Oct 06 10:11:31 crc kubenswrapper[4824]: I1006 10:11:31.144987 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vshhb\" (UniqueName: \"kubernetes.io/projected/b85edf34-0286-409d-907d-b97fceccdbae-kube-api-access-vshhb\") pod \"redhat-operators-vlxgd\" (UID: \"b85edf34-0286-409d-907d-b97fceccdbae\") " pod="openshift-marketplace/redhat-operators-vlxgd" Oct 06 10:11:31 crc kubenswrapper[4824]: I1006 10:11:31.245905 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b85edf34-0286-409d-907d-b97fceccdbae-utilities\") pod \"redhat-operators-vlxgd\" (UID: \"b85edf34-0286-409d-907d-b97fceccdbae\") " pod="openshift-marketplace/redhat-operators-vlxgd" Oct 06 10:11:31 crc kubenswrapper[4824]: I1006 10:11:31.246416 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vshhb\" (UniqueName: \"kubernetes.io/projected/b85edf34-0286-409d-907d-b97fceccdbae-kube-api-access-vshhb\") pod \"redhat-operators-vlxgd\" (UID: \"b85edf34-0286-409d-907d-b97fceccdbae\") " pod="openshift-marketplace/redhat-operators-vlxgd" Oct 06 10:11:31 crc kubenswrapper[4824]: I1006 10:11:31.246519 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b85edf34-0286-409d-907d-b97fceccdbae-catalog-content\") pod \"redhat-operators-vlxgd\" (UID: \"b85edf34-0286-409d-907d-b97fceccdbae\") " pod="openshift-marketplace/redhat-operators-vlxgd" Oct 06 10:11:31 crc kubenswrapper[4824]: I1006 10:11:31.246592 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b85edf34-0286-409d-907d-b97fceccdbae-utilities\") pod \"redhat-operators-vlxgd\" (UID: \"b85edf34-0286-409d-907d-b97fceccdbae\") " pod="openshift-marketplace/redhat-operators-vlxgd" Oct 06 10:11:31 crc kubenswrapper[4824]: I1006 10:11:31.246862 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b85edf34-0286-409d-907d-b97fceccdbae-catalog-content\") pod \"redhat-operators-vlxgd\" (UID: \"b85edf34-0286-409d-907d-b97fceccdbae\") " pod="openshift-marketplace/redhat-operators-vlxgd" Oct 06 10:11:31 crc kubenswrapper[4824]: I1006 10:11:31.258858 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-g7f78" Oct 06 10:11:31 crc kubenswrapper[4824]: I1006 10:11:31.274066 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vshhb\" (UniqueName: \"kubernetes.io/projected/b85edf34-0286-409d-907d-b97fceccdbae-kube-api-access-vshhb\") pod \"redhat-operators-vlxgd\" (UID: \"b85edf34-0286-409d-907d-b97fceccdbae\") " pod="openshift-marketplace/redhat-operators-vlxgd" Oct 06 10:11:31 crc kubenswrapper[4824]: I1006 10:11:31.312849 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vlxgd" Oct 06 10:11:31 crc kubenswrapper[4824]: I1006 10:11:31.773008 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vlxgd"] Oct 06 10:11:31 crc kubenswrapper[4824]: W1006 10:11:31.793511 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb85edf34_0286_409d_907d_b97fceccdbae.slice/crio-2ca56b22f5dccf759646c495f8aec128a4789415a0b657fe6ea5f1bfdf7223cf WatchSource:0}: Error finding container 2ca56b22f5dccf759646c495f8aec128a4789415a0b657fe6ea5f1bfdf7223cf: Status 404 returned error can't find the container with id 2ca56b22f5dccf759646c495f8aec128a4789415a0b657fe6ea5f1bfdf7223cf Oct 06 10:11:32 crc kubenswrapper[4824]: I1006 10:11:32.214911 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vlxgd" event={"ID":"b85edf34-0286-409d-907d-b97fceccdbae","Type":"ContainerStarted","Data":"2ca56b22f5dccf759646c495f8aec128a4789415a0b657fe6ea5f1bfdf7223cf"} Oct 06 10:11:32 crc kubenswrapper[4824]: E1006 10:11:32.969526 4824 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb85edf34_0286_409d_907d_b97fceccdbae.slice/crio-conmon-9194d3bd1c58e8292570da63ceff2b0fa373d545d01cf22100470624a8f7e963.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb85edf34_0286_409d_907d_b97fceccdbae.slice/crio-9194d3bd1c58e8292570da63ceff2b0fa373d545d01cf22100470624a8f7e963.scope\": RecentStats: unable to find data in memory cache]" Oct 06 10:11:33 crc kubenswrapper[4824]: I1006 10:11:33.229620 4824 generic.go:334] "Generic (PLEG): container finished" podID="b85edf34-0286-409d-907d-b97fceccdbae" containerID="9194d3bd1c58e8292570da63ceff2b0fa373d545d01cf22100470624a8f7e963" exitCode=0 Oct 06 10:11:33 crc kubenswrapper[4824]: I1006 10:11:33.229718 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vlxgd" event={"ID":"b85edf34-0286-409d-907d-b97fceccdbae","Type":"ContainerDied","Data":"9194d3bd1c58e8292570da63ceff2b0fa373d545d01cf22100470624a8f7e963"} Oct 06 10:11:34 crc kubenswrapper[4824]: I1006 10:11:34.764733 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-g7f78"] Oct 06 10:11:34 crc kubenswrapper[4824]: I1006 10:11:34.765587 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-g7f78" podUID="edb02879-620e-473c-aa8a-1f8aa4fffd6d" containerName="registry-server" containerID="cri-o://58c143e2f9da8ac94b66c2d4288484e15ffa3f4c8eb97e55ac139a05fe4e0ee3" gracePeriod=2 Oct 06 10:11:35 crc kubenswrapper[4824]: I1006 10:11:35.240712 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-g7f78" Oct 06 10:11:35 crc kubenswrapper[4824]: I1006 10:11:35.244759 4824 generic.go:334] "Generic (PLEG): container finished" podID="b85edf34-0286-409d-907d-b97fceccdbae" containerID="66835c780e6651eb2bd09e56ee7ea47a6beb2c63d37ae9847f162b5db98b253d" exitCode=0 Oct 06 10:11:35 crc kubenswrapper[4824]: I1006 10:11:35.244829 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vlxgd" event={"ID":"b85edf34-0286-409d-907d-b97fceccdbae","Type":"ContainerDied","Data":"66835c780e6651eb2bd09e56ee7ea47a6beb2c63d37ae9847f162b5db98b253d"} Oct 06 10:11:35 crc kubenswrapper[4824]: I1006 10:11:35.250707 4824 generic.go:334] "Generic (PLEG): container finished" podID="edb02879-620e-473c-aa8a-1f8aa4fffd6d" containerID="58c143e2f9da8ac94b66c2d4288484e15ffa3f4c8eb97e55ac139a05fe4e0ee3" exitCode=0 Oct 06 10:11:35 crc kubenswrapper[4824]: I1006 10:11:35.250767 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-g7f78" event={"ID":"edb02879-620e-473c-aa8a-1f8aa4fffd6d","Type":"ContainerDied","Data":"58c143e2f9da8ac94b66c2d4288484e15ffa3f4c8eb97e55ac139a05fe4e0ee3"} Oct 06 10:11:35 crc kubenswrapper[4824]: I1006 10:11:35.250803 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-g7f78" event={"ID":"edb02879-620e-473c-aa8a-1f8aa4fffd6d","Type":"ContainerDied","Data":"358346d19761c77c0fcac98cf778e534913836ce8e9b8e68c47533a3444f1d72"} Oct 06 10:11:35 crc kubenswrapper[4824]: I1006 10:11:35.250827 4824 scope.go:117] "RemoveContainer" containerID="58c143e2f9da8ac94b66c2d4288484e15ffa3f4c8eb97e55ac139a05fe4e0ee3" Oct 06 10:11:35 crc kubenswrapper[4824]: I1006 10:11:35.251002 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-g7f78" Oct 06 10:11:35 crc kubenswrapper[4824]: I1006 10:11:35.275823 4824 scope.go:117] "RemoveContainer" containerID="700676481b53609568aaa87548dee9160418feb7a77525682783a6890ec68ea3" Oct 06 10:11:35 crc kubenswrapper[4824]: I1006 10:11:35.308791 4824 scope.go:117] "RemoveContainer" containerID="31add07254df1b10e73843970fef95066c96749b88ef9b4a677de9caa306370c" Oct 06 10:11:35 crc kubenswrapper[4824]: I1006 10:11:35.343427 4824 scope.go:117] "RemoveContainer" containerID="58c143e2f9da8ac94b66c2d4288484e15ffa3f4c8eb97e55ac139a05fe4e0ee3" Oct 06 10:11:35 crc kubenswrapper[4824]: E1006 10:11:35.344164 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"58c143e2f9da8ac94b66c2d4288484e15ffa3f4c8eb97e55ac139a05fe4e0ee3\": container with ID starting with 58c143e2f9da8ac94b66c2d4288484e15ffa3f4c8eb97e55ac139a05fe4e0ee3 not found: ID does not exist" containerID="58c143e2f9da8ac94b66c2d4288484e15ffa3f4c8eb97e55ac139a05fe4e0ee3" Oct 06 10:11:35 crc kubenswrapper[4824]: I1006 10:11:35.344203 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"58c143e2f9da8ac94b66c2d4288484e15ffa3f4c8eb97e55ac139a05fe4e0ee3"} err="failed to get container status \"58c143e2f9da8ac94b66c2d4288484e15ffa3f4c8eb97e55ac139a05fe4e0ee3\": rpc error: code = NotFound desc = could not find container \"58c143e2f9da8ac94b66c2d4288484e15ffa3f4c8eb97e55ac139a05fe4e0ee3\": container with ID starting with 58c143e2f9da8ac94b66c2d4288484e15ffa3f4c8eb97e55ac139a05fe4e0ee3 not found: ID does not exist" Oct 06 10:11:35 crc kubenswrapper[4824]: I1006 10:11:35.344232 4824 scope.go:117] "RemoveContainer" containerID="700676481b53609568aaa87548dee9160418feb7a77525682783a6890ec68ea3" Oct 06 10:11:35 crc kubenswrapper[4824]: E1006 10:11:35.344671 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"700676481b53609568aaa87548dee9160418feb7a77525682783a6890ec68ea3\": container with ID starting with 700676481b53609568aaa87548dee9160418feb7a77525682783a6890ec68ea3 not found: ID does not exist" containerID="700676481b53609568aaa87548dee9160418feb7a77525682783a6890ec68ea3" Oct 06 10:11:35 crc kubenswrapper[4824]: I1006 10:11:35.344697 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"700676481b53609568aaa87548dee9160418feb7a77525682783a6890ec68ea3"} err="failed to get container status \"700676481b53609568aaa87548dee9160418feb7a77525682783a6890ec68ea3\": rpc error: code = NotFound desc = could not find container \"700676481b53609568aaa87548dee9160418feb7a77525682783a6890ec68ea3\": container with ID starting with 700676481b53609568aaa87548dee9160418feb7a77525682783a6890ec68ea3 not found: ID does not exist" Oct 06 10:11:35 crc kubenswrapper[4824]: I1006 10:11:35.344713 4824 scope.go:117] "RemoveContainer" containerID="31add07254df1b10e73843970fef95066c96749b88ef9b4a677de9caa306370c" Oct 06 10:11:35 crc kubenswrapper[4824]: E1006 10:11:35.345171 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"31add07254df1b10e73843970fef95066c96749b88ef9b4a677de9caa306370c\": container with ID starting with 31add07254df1b10e73843970fef95066c96749b88ef9b4a677de9caa306370c not found: ID does not exist" containerID="31add07254df1b10e73843970fef95066c96749b88ef9b4a677de9caa306370c" Oct 06 10:11:35 crc kubenswrapper[4824]: I1006 10:11:35.345264 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"31add07254df1b10e73843970fef95066c96749b88ef9b4a677de9caa306370c"} err="failed to get container status \"31add07254df1b10e73843970fef95066c96749b88ef9b4a677de9caa306370c\": rpc error: code = NotFound desc = could not find container \"31add07254df1b10e73843970fef95066c96749b88ef9b4a677de9caa306370c\": container with ID starting with 31add07254df1b10e73843970fef95066c96749b88ef9b4a677de9caa306370c not found: ID does not exist" Oct 06 10:11:35 crc kubenswrapper[4824]: I1006 10:11:35.427263 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/edb02879-620e-473c-aa8a-1f8aa4fffd6d-utilities\") pod \"edb02879-620e-473c-aa8a-1f8aa4fffd6d\" (UID: \"edb02879-620e-473c-aa8a-1f8aa4fffd6d\") " Oct 06 10:11:35 crc kubenswrapper[4824]: I1006 10:11:35.427345 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m7vv6\" (UniqueName: \"kubernetes.io/projected/edb02879-620e-473c-aa8a-1f8aa4fffd6d-kube-api-access-m7vv6\") pod \"edb02879-620e-473c-aa8a-1f8aa4fffd6d\" (UID: \"edb02879-620e-473c-aa8a-1f8aa4fffd6d\") " Oct 06 10:11:35 crc kubenswrapper[4824]: I1006 10:11:35.427421 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/edb02879-620e-473c-aa8a-1f8aa4fffd6d-catalog-content\") pod \"edb02879-620e-473c-aa8a-1f8aa4fffd6d\" (UID: \"edb02879-620e-473c-aa8a-1f8aa4fffd6d\") " Oct 06 10:11:35 crc kubenswrapper[4824]: I1006 10:11:35.428819 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/edb02879-620e-473c-aa8a-1f8aa4fffd6d-utilities" (OuterVolumeSpecName: "utilities") pod "edb02879-620e-473c-aa8a-1f8aa4fffd6d" (UID: "edb02879-620e-473c-aa8a-1f8aa4fffd6d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:11:35 crc kubenswrapper[4824]: I1006 10:11:35.436732 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/edb02879-620e-473c-aa8a-1f8aa4fffd6d-kube-api-access-m7vv6" (OuterVolumeSpecName: "kube-api-access-m7vv6") pod "edb02879-620e-473c-aa8a-1f8aa4fffd6d" (UID: "edb02879-620e-473c-aa8a-1f8aa4fffd6d"). InnerVolumeSpecName "kube-api-access-m7vv6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:11:35 crc kubenswrapper[4824]: I1006 10:11:35.457480 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/edb02879-620e-473c-aa8a-1f8aa4fffd6d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "edb02879-620e-473c-aa8a-1f8aa4fffd6d" (UID: "edb02879-620e-473c-aa8a-1f8aa4fffd6d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:11:35 crc kubenswrapper[4824]: I1006 10:11:35.529419 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/edb02879-620e-473c-aa8a-1f8aa4fffd6d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 06 10:11:35 crc kubenswrapper[4824]: I1006 10:11:35.529828 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/edb02879-620e-473c-aa8a-1f8aa4fffd6d-utilities\") on node \"crc\" DevicePath \"\"" Oct 06 10:11:35 crc kubenswrapper[4824]: I1006 10:11:35.529969 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m7vv6\" (UniqueName: \"kubernetes.io/projected/edb02879-620e-473c-aa8a-1f8aa4fffd6d-kube-api-access-m7vv6\") on node \"crc\" DevicePath \"\"" Oct 06 10:11:35 crc kubenswrapper[4824]: I1006 10:11:35.614624 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-g7f78"] Oct 06 10:11:35 crc kubenswrapper[4824]: I1006 10:11:35.622657 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-g7f78"] Oct 06 10:11:36 crc kubenswrapper[4824]: I1006 10:11:36.262881 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vlxgd" event={"ID":"b85edf34-0286-409d-907d-b97fceccdbae","Type":"ContainerStarted","Data":"c724e6aad9df321989a30b3ed8e95c48b9adb0a0225b65af3176a173459999a6"} Oct 06 10:11:36 crc kubenswrapper[4824]: I1006 10:11:36.295296 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-vlxgd" podStartSLOduration=3.83078072 podStartE2EDuration="6.29526807s" podCreationTimestamp="2025-10-06 10:11:30 +0000 UTC" firstStartedPulling="2025-10-06 10:11:33.233397071 +0000 UTC m=+802.597819962" lastFinishedPulling="2025-10-06 10:11:35.697884431 +0000 UTC m=+805.062307312" observedRunningTime="2025-10-06 10:11:36.292307586 +0000 UTC m=+805.656730477" watchObservedRunningTime="2025-10-06 10:11:36.29526807 +0000 UTC m=+805.659690941" Oct 06 10:11:37 crc kubenswrapper[4824]: I1006 10:11:37.262763 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-x6lsw" Oct 06 10:11:37 crc kubenswrapper[4824]: I1006 10:11:37.281464 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="edb02879-620e-473c-aa8a-1f8aa4fffd6d" path="/var/lib/kubelet/pods/edb02879-620e-473c-aa8a-1f8aa4fffd6d/volumes" Oct 06 10:11:38 crc kubenswrapper[4824]: I1006 10:11:38.772726 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-dlhm7"] Oct 06 10:11:38 crc kubenswrapper[4824]: E1006 10:11:38.773442 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="edb02879-620e-473c-aa8a-1f8aa4fffd6d" containerName="extract-content" Oct 06 10:11:38 crc kubenswrapper[4824]: I1006 10:11:38.773457 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="edb02879-620e-473c-aa8a-1f8aa4fffd6d" containerName="extract-content" Oct 06 10:11:38 crc kubenswrapper[4824]: E1006 10:11:38.773474 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="edb02879-620e-473c-aa8a-1f8aa4fffd6d" containerName="extract-utilities" Oct 06 10:11:38 crc kubenswrapper[4824]: I1006 10:11:38.773480 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="edb02879-620e-473c-aa8a-1f8aa4fffd6d" containerName="extract-utilities" Oct 06 10:11:38 crc kubenswrapper[4824]: E1006 10:11:38.773497 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="edb02879-620e-473c-aa8a-1f8aa4fffd6d" containerName="registry-server" Oct 06 10:11:38 crc kubenswrapper[4824]: I1006 10:11:38.773503 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="edb02879-620e-473c-aa8a-1f8aa4fffd6d" containerName="registry-server" Oct 06 10:11:38 crc kubenswrapper[4824]: I1006 10:11:38.773615 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="edb02879-620e-473c-aa8a-1f8aa4fffd6d" containerName="registry-server" Oct 06 10:11:38 crc kubenswrapper[4824]: I1006 10:11:38.774113 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-dlhm7" Oct 06 10:11:38 crc kubenswrapper[4824]: I1006 10:11:38.777222 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-vwrjs" Oct 06 10:11:38 crc kubenswrapper[4824]: I1006 10:11:38.778347 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Oct 06 10:11:38 crc kubenswrapper[4824]: I1006 10:11:38.780832 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Oct 06 10:11:38 crc kubenswrapper[4824]: I1006 10:11:38.780913 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h6q6h\" (UniqueName: \"kubernetes.io/projected/bd9ffa83-7e81-4322-9507-e92aeec0276d-kube-api-access-h6q6h\") pod \"openstack-operator-index-dlhm7\" (UID: \"bd9ffa83-7e81-4322-9507-e92aeec0276d\") " pod="openstack-operators/openstack-operator-index-dlhm7" Oct 06 10:11:38 crc kubenswrapper[4824]: I1006 10:11:38.791136 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-dlhm7"] Oct 06 10:11:38 crc kubenswrapper[4824]: I1006 10:11:38.882195 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h6q6h\" (UniqueName: \"kubernetes.io/projected/bd9ffa83-7e81-4322-9507-e92aeec0276d-kube-api-access-h6q6h\") pod \"openstack-operator-index-dlhm7\" (UID: \"bd9ffa83-7e81-4322-9507-e92aeec0276d\") " pod="openstack-operators/openstack-operator-index-dlhm7" Oct 06 10:11:38 crc kubenswrapper[4824]: I1006 10:11:38.907238 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h6q6h\" (UniqueName: \"kubernetes.io/projected/bd9ffa83-7e81-4322-9507-e92aeec0276d-kube-api-access-h6q6h\") pod \"openstack-operator-index-dlhm7\" (UID: \"bd9ffa83-7e81-4322-9507-e92aeec0276d\") " pod="openstack-operators/openstack-operator-index-dlhm7" Oct 06 10:11:39 crc kubenswrapper[4824]: I1006 10:11:39.116709 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-dlhm7" Oct 06 10:11:39 crc kubenswrapper[4824]: I1006 10:11:39.561401 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-dlhm7"] Oct 06 10:11:39 crc kubenswrapper[4824]: W1006 10:11:39.576744 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbd9ffa83_7e81_4322_9507_e92aeec0276d.slice/crio-0bdd17b544f8e656a908635400dc8790787449a74c9642f2847e440f56af0e20 WatchSource:0}: Error finding container 0bdd17b544f8e656a908635400dc8790787449a74c9642f2847e440f56af0e20: Status 404 returned error can't find the container with id 0bdd17b544f8e656a908635400dc8790787449a74c9642f2847e440f56af0e20 Oct 06 10:11:40 crc kubenswrapper[4824]: I1006 10:11:40.314345 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-dlhm7" event={"ID":"bd9ffa83-7e81-4322-9507-e92aeec0276d","Type":"ContainerStarted","Data":"0bdd17b544f8e656a908635400dc8790787449a74c9642f2847e440f56af0e20"} Oct 06 10:11:41 crc kubenswrapper[4824]: I1006 10:11:41.313895 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-vlxgd" Oct 06 10:11:41 crc kubenswrapper[4824]: I1006 10:11:41.314366 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-vlxgd" Oct 06 10:11:41 crc kubenswrapper[4824]: I1006 10:11:41.370527 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-vlxgd" Oct 06 10:11:42 crc kubenswrapper[4824]: I1006 10:11:42.390630 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-vlxgd" Oct 06 10:11:43 crc kubenswrapper[4824]: I1006 10:11:43.342655 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-dlhm7" event={"ID":"bd9ffa83-7e81-4322-9507-e92aeec0276d","Type":"ContainerStarted","Data":"6ef5337a5ce637ae6ab93d0a5b142f56fbbe4d58ddeac646aff40c07dfd98540"} Oct 06 10:11:43 crc kubenswrapper[4824]: I1006 10:11:43.372452 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-dlhm7" podStartSLOduration=2.758560404 podStartE2EDuration="5.372429727s" podCreationTimestamp="2025-10-06 10:11:38 +0000 UTC" firstStartedPulling="2025-10-06 10:11:39.584077894 +0000 UTC m=+808.948500765" lastFinishedPulling="2025-10-06 10:11:42.197947187 +0000 UTC m=+811.562370088" observedRunningTime="2025-10-06 10:11:43.364441268 +0000 UTC m=+812.728864129" watchObservedRunningTime="2025-10-06 10:11:43.372429727 +0000 UTC m=+812.736852588" Oct 06 10:11:43 crc kubenswrapper[4824]: I1006 10:11:43.915887 4824 patch_prober.go:28] interesting pod/machine-config-daemon-khgzw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 10:11:43 crc kubenswrapper[4824]: I1006 10:11:43.916077 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 10:11:46 crc kubenswrapper[4824]: I1006 10:11:46.165756 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-vlxgd"] Oct 06 10:11:46 crc kubenswrapper[4824]: I1006 10:11:46.166289 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-vlxgd" podUID="b85edf34-0286-409d-907d-b97fceccdbae" containerName="registry-server" containerID="cri-o://c724e6aad9df321989a30b3ed8e95c48b9adb0a0225b65af3176a173459999a6" gracePeriod=2 Oct 06 10:11:48 crc kubenswrapper[4824]: I1006 10:11:48.385049 4824 generic.go:334] "Generic (PLEG): container finished" podID="b85edf34-0286-409d-907d-b97fceccdbae" containerID="c724e6aad9df321989a30b3ed8e95c48b9adb0a0225b65af3176a173459999a6" exitCode=0 Oct 06 10:11:48 crc kubenswrapper[4824]: I1006 10:11:48.385872 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vlxgd" event={"ID":"b85edf34-0286-409d-907d-b97fceccdbae","Type":"ContainerDied","Data":"c724e6aad9df321989a30b3ed8e95c48b9adb0a0225b65af3176a173459999a6"} Oct 06 10:11:48 crc kubenswrapper[4824]: I1006 10:11:48.525427 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vlxgd" Oct 06 10:11:48 crc kubenswrapper[4824]: I1006 10:11:48.641605 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vshhb\" (UniqueName: \"kubernetes.io/projected/b85edf34-0286-409d-907d-b97fceccdbae-kube-api-access-vshhb\") pod \"b85edf34-0286-409d-907d-b97fceccdbae\" (UID: \"b85edf34-0286-409d-907d-b97fceccdbae\") " Oct 06 10:11:48 crc kubenswrapper[4824]: I1006 10:11:48.641679 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b85edf34-0286-409d-907d-b97fceccdbae-catalog-content\") pod \"b85edf34-0286-409d-907d-b97fceccdbae\" (UID: \"b85edf34-0286-409d-907d-b97fceccdbae\") " Oct 06 10:11:48 crc kubenswrapper[4824]: I1006 10:11:48.641767 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b85edf34-0286-409d-907d-b97fceccdbae-utilities\") pod \"b85edf34-0286-409d-907d-b97fceccdbae\" (UID: \"b85edf34-0286-409d-907d-b97fceccdbae\") " Oct 06 10:11:48 crc kubenswrapper[4824]: I1006 10:11:48.643046 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b85edf34-0286-409d-907d-b97fceccdbae-utilities" (OuterVolumeSpecName: "utilities") pod "b85edf34-0286-409d-907d-b97fceccdbae" (UID: "b85edf34-0286-409d-907d-b97fceccdbae"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:11:48 crc kubenswrapper[4824]: I1006 10:11:48.652257 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b85edf34-0286-409d-907d-b97fceccdbae-kube-api-access-vshhb" (OuterVolumeSpecName: "kube-api-access-vshhb") pod "b85edf34-0286-409d-907d-b97fceccdbae" (UID: "b85edf34-0286-409d-907d-b97fceccdbae"). InnerVolumeSpecName "kube-api-access-vshhb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:11:48 crc kubenswrapper[4824]: I1006 10:11:48.736283 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b85edf34-0286-409d-907d-b97fceccdbae-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b85edf34-0286-409d-907d-b97fceccdbae" (UID: "b85edf34-0286-409d-907d-b97fceccdbae"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:11:48 crc kubenswrapper[4824]: I1006 10:11:48.744470 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vshhb\" (UniqueName: \"kubernetes.io/projected/b85edf34-0286-409d-907d-b97fceccdbae-kube-api-access-vshhb\") on node \"crc\" DevicePath \"\"" Oct 06 10:11:48 crc kubenswrapper[4824]: I1006 10:11:48.744523 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b85edf34-0286-409d-907d-b97fceccdbae-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 06 10:11:48 crc kubenswrapper[4824]: I1006 10:11:48.744544 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b85edf34-0286-409d-907d-b97fceccdbae-utilities\") on node \"crc\" DevicePath \"\"" Oct 06 10:11:49 crc kubenswrapper[4824]: I1006 10:11:49.117578 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-dlhm7" Oct 06 10:11:49 crc kubenswrapper[4824]: I1006 10:11:49.117630 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-dlhm7" Oct 06 10:11:49 crc kubenswrapper[4824]: I1006 10:11:49.151629 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-dlhm7" Oct 06 10:11:49 crc kubenswrapper[4824]: I1006 10:11:49.394525 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vlxgd" event={"ID":"b85edf34-0286-409d-907d-b97fceccdbae","Type":"ContainerDied","Data":"2ca56b22f5dccf759646c495f8aec128a4789415a0b657fe6ea5f1bfdf7223cf"} Oct 06 10:11:49 crc kubenswrapper[4824]: I1006 10:11:49.394560 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vlxgd" Oct 06 10:11:49 crc kubenswrapper[4824]: I1006 10:11:49.394578 4824 scope.go:117] "RemoveContainer" containerID="c724e6aad9df321989a30b3ed8e95c48b9adb0a0225b65af3176a173459999a6" Oct 06 10:11:49 crc kubenswrapper[4824]: I1006 10:11:49.428276 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-vlxgd"] Oct 06 10:11:49 crc kubenswrapper[4824]: I1006 10:11:49.449787 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-vlxgd"] Oct 06 10:11:49 crc kubenswrapper[4824]: I1006 10:11:49.452621 4824 scope.go:117] "RemoveContainer" containerID="66835c780e6651eb2bd09e56ee7ea47a6beb2c63d37ae9847f162b5db98b253d" Oct 06 10:11:49 crc kubenswrapper[4824]: I1006 10:11:49.452891 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-dlhm7" Oct 06 10:11:49 crc kubenswrapper[4824]: I1006 10:11:49.475348 4824 scope.go:117] "RemoveContainer" containerID="9194d3bd1c58e8292570da63ceff2b0fa373d545d01cf22100470624a8f7e963" Oct 06 10:11:51 crc kubenswrapper[4824]: I1006 10:11:51.170069 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-mvt5t"] Oct 06 10:11:51 crc kubenswrapper[4824]: E1006 10:11:51.170304 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b85edf34-0286-409d-907d-b97fceccdbae" containerName="extract-utilities" Oct 06 10:11:51 crc kubenswrapper[4824]: I1006 10:11:51.170318 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="b85edf34-0286-409d-907d-b97fceccdbae" containerName="extract-utilities" Oct 06 10:11:51 crc kubenswrapper[4824]: E1006 10:11:51.170335 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b85edf34-0286-409d-907d-b97fceccdbae" containerName="registry-server" Oct 06 10:11:51 crc kubenswrapper[4824]: I1006 10:11:51.170341 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="b85edf34-0286-409d-907d-b97fceccdbae" containerName="registry-server" Oct 06 10:11:51 crc kubenswrapper[4824]: E1006 10:11:51.170355 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b85edf34-0286-409d-907d-b97fceccdbae" containerName="extract-content" Oct 06 10:11:51 crc kubenswrapper[4824]: I1006 10:11:51.170362 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="b85edf34-0286-409d-907d-b97fceccdbae" containerName="extract-content" Oct 06 10:11:51 crc kubenswrapper[4824]: I1006 10:11:51.170468 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="b85edf34-0286-409d-907d-b97fceccdbae" containerName="registry-server" Oct 06 10:11:51 crc kubenswrapper[4824]: I1006 10:11:51.171278 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mvt5t" Oct 06 10:11:51 crc kubenswrapper[4824]: I1006 10:11:51.193156 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-mvt5t"] Oct 06 10:11:51 crc kubenswrapper[4824]: I1006 10:11:51.282864 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b85edf34-0286-409d-907d-b97fceccdbae" path="/var/lib/kubelet/pods/b85edf34-0286-409d-907d-b97fceccdbae/volumes" Oct 06 10:11:51 crc kubenswrapper[4824]: I1006 10:11:51.285524 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rw2jm\" (UniqueName: \"kubernetes.io/projected/fc277ec2-0d30-472a-af67-7019f23e7a3c-kube-api-access-rw2jm\") pod \"community-operators-mvt5t\" (UID: \"fc277ec2-0d30-472a-af67-7019f23e7a3c\") " pod="openshift-marketplace/community-operators-mvt5t" Oct 06 10:11:51 crc kubenswrapper[4824]: I1006 10:11:51.285657 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fc277ec2-0d30-472a-af67-7019f23e7a3c-utilities\") pod \"community-operators-mvt5t\" (UID: \"fc277ec2-0d30-472a-af67-7019f23e7a3c\") " pod="openshift-marketplace/community-operators-mvt5t" Oct 06 10:11:51 crc kubenswrapper[4824]: I1006 10:11:51.285742 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fc277ec2-0d30-472a-af67-7019f23e7a3c-catalog-content\") pod \"community-operators-mvt5t\" (UID: \"fc277ec2-0d30-472a-af67-7019f23e7a3c\") " pod="openshift-marketplace/community-operators-mvt5t" Oct 06 10:11:51 crc kubenswrapper[4824]: I1006 10:11:51.386916 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fc277ec2-0d30-472a-af67-7019f23e7a3c-utilities\") pod \"community-operators-mvt5t\" (UID: \"fc277ec2-0d30-472a-af67-7019f23e7a3c\") " pod="openshift-marketplace/community-operators-mvt5t" Oct 06 10:11:51 crc kubenswrapper[4824]: I1006 10:11:51.387013 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fc277ec2-0d30-472a-af67-7019f23e7a3c-catalog-content\") pod \"community-operators-mvt5t\" (UID: \"fc277ec2-0d30-472a-af67-7019f23e7a3c\") " pod="openshift-marketplace/community-operators-mvt5t" Oct 06 10:11:51 crc kubenswrapper[4824]: I1006 10:11:51.387107 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rw2jm\" (UniqueName: \"kubernetes.io/projected/fc277ec2-0d30-472a-af67-7019f23e7a3c-kube-api-access-rw2jm\") pod \"community-operators-mvt5t\" (UID: \"fc277ec2-0d30-472a-af67-7019f23e7a3c\") " pod="openshift-marketplace/community-operators-mvt5t" Oct 06 10:11:51 crc kubenswrapper[4824]: I1006 10:11:51.387647 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fc277ec2-0d30-472a-af67-7019f23e7a3c-utilities\") pod \"community-operators-mvt5t\" (UID: \"fc277ec2-0d30-472a-af67-7019f23e7a3c\") " pod="openshift-marketplace/community-operators-mvt5t" Oct 06 10:11:51 crc kubenswrapper[4824]: I1006 10:11:51.387856 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fc277ec2-0d30-472a-af67-7019f23e7a3c-catalog-content\") pod \"community-operators-mvt5t\" (UID: \"fc277ec2-0d30-472a-af67-7019f23e7a3c\") " pod="openshift-marketplace/community-operators-mvt5t" Oct 06 10:11:51 crc kubenswrapper[4824]: I1006 10:11:51.413678 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rw2jm\" (UniqueName: \"kubernetes.io/projected/fc277ec2-0d30-472a-af67-7019f23e7a3c-kube-api-access-rw2jm\") pod \"community-operators-mvt5t\" (UID: \"fc277ec2-0d30-472a-af67-7019f23e7a3c\") " pod="openshift-marketplace/community-operators-mvt5t" Oct 06 10:11:51 crc kubenswrapper[4824]: I1006 10:11:51.496492 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mvt5t" Oct 06 10:11:52 crc kubenswrapper[4824]: I1006 10:11:52.873946 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-mvt5t"] Oct 06 10:11:53 crc kubenswrapper[4824]: I1006 10:11:53.424381 4824 generic.go:334] "Generic (PLEG): container finished" podID="fc277ec2-0d30-472a-af67-7019f23e7a3c" containerID="be16a5dc7bdfcf42b942d0b49c6afdf8f6c3ba07550089ab6dca2bbe82253933" exitCode=0 Oct 06 10:11:53 crc kubenswrapper[4824]: I1006 10:11:53.424773 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mvt5t" event={"ID":"fc277ec2-0d30-472a-af67-7019f23e7a3c","Type":"ContainerDied","Data":"be16a5dc7bdfcf42b942d0b49c6afdf8f6c3ba07550089ab6dca2bbe82253933"} Oct 06 10:11:53 crc kubenswrapper[4824]: I1006 10:11:53.424805 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mvt5t" event={"ID":"fc277ec2-0d30-472a-af67-7019f23e7a3c","Type":"ContainerStarted","Data":"6233fec9f83ed72e1341fc9bc4bdd4a4c7ee44df3d3343785c5c393a59141b9d"} Oct 06 10:11:53 crc kubenswrapper[4824]: I1006 10:11:53.859295 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/a842b0520c38c46f4cb51faa08e79e3c8193222ff536aee10057c0eb8cqhcsj"] Oct 06 10:11:53 crc kubenswrapper[4824]: I1006 10:11:53.861375 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/a842b0520c38c46f4cb51faa08e79e3c8193222ff536aee10057c0eb8cqhcsj" Oct 06 10:11:53 crc kubenswrapper[4824]: I1006 10:11:53.867775 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-54cnn" Oct 06 10:11:53 crc kubenswrapper[4824]: I1006 10:11:53.873670 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/a842b0520c38c46f4cb51faa08e79e3c8193222ff536aee10057c0eb8cqhcsj"] Oct 06 10:11:54 crc kubenswrapper[4824]: I1006 10:11:54.037235 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-djsm9\" (UniqueName: \"kubernetes.io/projected/65505c4e-e82d-485f-abc6-33be1406fd9a-kube-api-access-djsm9\") pod \"a842b0520c38c46f4cb51faa08e79e3c8193222ff536aee10057c0eb8cqhcsj\" (UID: \"65505c4e-e82d-485f-abc6-33be1406fd9a\") " pod="openstack-operators/a842b0520c38c46f4cb51faa08e79e3c8193222ff536aee10057c0eb8cqhcsj" Oct 06 10:11:54 crc kubenswrapper[4824]: I1006 10:11:54.037383 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/65505c4e-e82d-485f-abc6-33be1406fd9a-util\") pod \"a842b0520c38c46f4cb51faa08e79e3c8193222ff536aee10057c0eb8cqhcsj\" (UID: \"65505c4e-e82d-485f-abc6-33be1406fd9a\") " pod="openstack-operators/a842b0520c38c46f4cb51faa08e79e3c8193222ff536aee10057c0eb8cqhcsj" Oct 06 10:11:54 crc kubenswrapper[4824]: I1006 10:11:54.037417 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/65505c4e-e82d-485f-abc6-33be1406fd9a-bundle\") pod \"a842b0520c38c46f4cb51faa08e79e3c8193222ff536aee10057c0eb8cqhcsj\" (UID: \"65505c4e-e82d-485f-abc6-33be1406fd9a\") " pod="openstack-operators/a842b0520c38c46f4cb51faa08e79e3c8193222ff536aee10057c0eb8cqhcsj" Oct 06 10:11:54 crc kubenswrapper[4824]: I1006 10:11:54.138208 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-djsm9\" (UniqueName: \"kubernetes.io/projected/65505c4e-e82d-485f-abc6-33be1406fd9a-kube-api-access-djsm9\") pod \"a842b0520c38c46f4cb51faa08e79e3c8193222ff536aee10057c0eb8cqhcsj\" (UID: \"65505c4e-e82d-485f-abc6-33be1406fd9a\") " pod="openstack-operators/a842b0520c38c46f4cb51faa08e79e3c8193222ff536aee10057c0eb8cqhcsj" Oct 06 10:11:54 crc kubenswrapper[4824]: I1006 10:11:54.138295 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/65505c4e-e82d-485f-abc6-33be1406fd9a-util\") pod \"a842b0520c38c46f4cb51faa08e79e3c8193222ff536aee10057c0eb8cqhcsj\" (UID: \"65505c4e-e82d-485f-abc6-33be1406fd9a\") " pod="openstack-operators/a842b0520c38c46f4cb51faa08e79e3c8193222ff536aee10057c0eb8cqhcsj" Oct 06 10:11:54 crc kubenswrapper[4824]: I1006 10:11:54.138317 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/65505c4e-e82d-485f-abc6-33be1406fd9a-bundle\") pod \"a842b0520c38c46f4cb51faa08e79e3c8193222ff536aee10057c0eb8cqhcsj\" (UID: \"65505c4e-e82d-485f-abc6-33be1406fd9a\") " pod="openstack-operators/a842b0520c38c46f4cb51faa08e79e3c8193222ff536aee10057c0eb8cqhcsj" Oct 06 10:11:54 crc kubenswrapper[4824]: I1006 10:11:54.138846 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/65505c4e-e82d-485f-abc6-33be1406fd9a-bundle\") pod \"a842b0520c38c46f4cb51faa08e79e3c8193222ff536aee10057c0eb8cqhcsj\" (UID: \"65505c4e-e82d-485f-abc6-33be1406fd9a\") " pod="openstack-operators/a842b0520c38c46f4cb51faa08e79e3c8193222ff536aee10057c0eb8cqhcsj" Oct 06 10:11:54 crc kubenswrapper[4824]: I1006 10:11:54.139033 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/65505c4e-e82d-485f-abc6-33be1406fd9a-util\") pod \"a842b0520c38c46f4cb51faa08e79e3c8193222ff536aee10057c0eb8cqhcsj\" (UID: \"65505c4e-e82d-485f-abc6-33be1406fd9a\") " pod="openstack-operators/a842b0520c38c46f4cb51faa08e79e3c8193222ff536aee10057c0eb8cqhcsj" Oct 06 10:11:54 crc kubenswrapper[4824]: I1006 10:11:54.162206 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-djsm9\" (UniqueName: \"kubernetes.io/projected/65505c4e-e82d-485f-abc6-33be1406fd9a-kube-api-access-djsm9\") pod \"a842b0520c38c46f4cb51faa08e79e3c8193222ff536aee10057c0eb8cqhcsj\" (UID: \"65505c4e-e82d-485f-abc6-33be1406fd9a\") " pod="openstack-operators/a842b0520c38c46f4cb51faa08e79e3c8193222ff536aee10057c0eb8cqhcsj" Oct 06 10:11:54 crc kubenswrapper[4824]: I1006 10:11:54.187376 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/a842b0520c38c46f4cb51faa08e79e3c8193222ff536aee10057c0eb8cqhcsj" Oct 06 10:11:54 crc kubenswrapper[4824]: I1006 10:11:54.688917 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/a842b0520c38c46f4cb51faa08e79e3c8193222ff536aee10057c0eb8cqhcsj"] Oct 06 10:11:55 crc kubenswrapper[4824]: I1006 10:11:55.442750 4824 generic.go:334] "Generic (PLEG): container finished" podID="65505c4e-e82d-485f-abc6-33be1406fd9a" containerID="2749e0efbb7cbae4281f5cd7ac7df82c20aa19e93b235ff29107cb053ea7e7d9" exitCode=0 Oct 06 10:11:55 crc kubenswrapper[4824]: I1006 10:11:55.443009 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/a842b0520c38c46f4cb51faa08e79e3c8193222ff536aee10057c0eb8cqhcsj" event={"ID":"65505c4e-e82d-485f-abc6-33be1406fd9a","Type":"ContainerDied","Data":"2749e0efbb7cbae4281f5cd7ac7df82c20aa19e93b235ff29107cb053ea7e7d9"} Oct 06 10:11:55 crc kubenswrapper[4824]: I1006 10:11:55.443149 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/a842b0520c38c46f4cb51faa08e79e3c8193222ff536aee10057c0eb8cqhcsj" event={"ID":"65505c4e-e82d-485f-abc6-33be1406fd9a","Type":"ContainerStarted","Data":"97ce47cbb7fca8fd8dde595674315a7f200ce0991f807a7a03ebb7435f00a4c0"} Oct 06 10:11:55 crc kubenswrapper[4824]: I1006 10:11:55.449115 4824 generic.go:334] "Generic (PLEG): container finished" podID="fc277ec2-0d30-472a-af67-7019f23e7a3c" containerID="471b26f61254f725b22745dbf5022817e73ebc1ba5b85f52326c6b7307ff81ea" exitCode=0 Oct 06 10:11:55 crc kubenswrapper[4824]: I1006 10:11:55.449142 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mvt5t" event={"ID":"fc277ec2-0d30-472a-af67-7019f23e7a3c","Type":"ContainerDied","Data":"471b26f61254f725b22745dbf5022817e73ebc1ba5b85f52326c6b7307ff81ea"} Oct 06 10:11:56 crc kubenswrapper[4824]: I1006 10:11:56.458166 4824 generic.go:334] "Generic (PLEG): container finished" podID="65505c4e-e82d-485f-abc6-33be1406fd9a" containerID="1c186af5782ed65ef1eb1df4b47613ba8fd941b55508c2aa11ae5d28b1b60361" exitCode=0 Oct 06 10:11:56 crc kubenswrapper[4824]: I1006 10:11:56.458271 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/a842b0520c38c46f4cb51faa08e79e3c8193222ff536aee10057c0eb8cqhcsj" event={"ID":"65505c4e-e82d-485f-abc6-33be1406fd9a","Type":"ContainerDied","Data":"1c186af5782ed65ef1eb1df4b47613ba8fd941b55508c2aa11ae5d28b1b60361"} Oct 06 10:11:56 crc kubenswrapper[4824]: I1006 10:11:56.462111 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mvt5t" event={"ID":"fc277ec2-0d30-472a-af67-7019f23e7a3c","Type":"ContainerStarted","Data":"5f26630a6457aac156ba508c4f5a2af5ab64aecca7d7b12b883e34162387e91d"} Oct 06 10:11:56 crc kubenswrapper[4824]: I1006 10:11:56.509204 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-mvt5t" podStartSLOduration=2.885019662 podStartE2EDuration="5.509178171s" podCreationTimestamp="2025-10-06 10:11:51 +0000 UTC" firstStartedPulling="2025-10-06 10:11:53.426256389 +0000 UTC m=+822.790679290" lastFinishedPulling="2025-10-06 10:11:56.050414938 +0000 UTC m=+825.414837799" observedRunningTime="2025-10-06 10:11:56.502455704 +0000 UTC m=+825.866878585" watchObservedRunningTime="2025-10-06 10:11:56.509178171 +0000 UTC m=+825.873601052" Oct 06 10:11:57 crc kubenswrapper[4824]: I1006 10:11:57.470920 4824 generic.go:334] "Generic (PLEG): container finished" podID="65505c4e-e82d-485f-abc6-33be1406fd9a" containerID="17b67da48ddcf40b8e39aca9b98fc5fbc2b12c234da4b88e668b91616ab621f6" exitCode=0 Oct 06 10:11:57 crc kubenswrapper[4824]: I1006 10:11:57.472318 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/a842b0520c38c46f4cb51faa08e79e3c8193222ff536aee10057c0eb8cqhcsj" event={"ID":"65505c4e-e82d-485f-abc6-33be1406fd9a","Type":"ContainerDied","Data":"17b67da48ddcf40b8e39aca9b98fc5fbc2b12c234da4b88e668b91616ab621f6"} Oct 06 10:11:58 crc kubenswrapper[4824]: I1006 10:11:58.755179 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/a842b0520c38c46f4cb51faa08e79e3c8193222ff536aee10057c0eb8cqhcsj" Oct 06 10:11:58 crc kubenswrapper[4824]: I1006 10:11:58.818185 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/65505c4e-e82d-485f-abc6-33be1406fd9a-util\") pod \"65505c4e-e82d-485f-abc6-33be1406fd9a\" (UID: \"65505c4e-e82d-485f-abc6-33be1406fd9a\") " Oct 06 10:11:58 crc kubenswrapper[4824]: I1006 10:11:58.818305 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/65505c4e-e82d-485f-abc6-33be1406fd9a-bundle\") pod \"65505c4e-e82d-485f-abc6-33be1406fd9a\" (UID: \"65505c4e-e82d-485f-abc6-33be1406fd9a\") " Oct 06 10:11:58 crc kubenswrapper[4824]: I1006 10:11:58.818339 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-djsm9\" (UniqueName: \"kubernetes.io/projected/65505c4e-e82d-485f-abc6-33be1406fd9a-kube-api-access-djsm9\") pod \"65505c4e-e82d-485f-abc6-33be1406fd9a\" (UID: \"65505c4e-e82d-485f-abc6-33be1406fd9a\") " Oct 06 10:11:58 crc kubenswrapper[4824]: I1006 10:11:58.819190 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/65505c4e-e82d-485f-abc6-33be1406fd9a-bundle" (OuterVolumeSpecName: "bundle") pod "65505c4e-e82d-485f-abc6-33be1406fd9a" (UID: "65505c4e-e82d-485f-abc6-33be1406fd9a"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:11:58 crc kubenswrapper[4824]: I1006 10:11:58.826690 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/65505c4e-e82d-485f-abc6-33be1406fd9a-kube-api-access-djsm9" (OuterVolumeSpecName: "kube-api-access-djsm9") pod "65505c4e-e82d-485f-abc6-33be1406fd9a" (UID: "65505c4e-e82d-485f-abc6-33be1406fd9a"). InnerVolumeSpecName "kube-api-access-djsm9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:11:58 crc kubenswrapper[4824]: I1006 10:11:58.831567 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/65505c4e-e82d-485f-abc6-33be1406fd9a-util" (OuterVolumeSpecName: "util") pod "65505c4e-e82d-485f-abc6-33be1406fd9a" (UID: "65505c4e-e82d-485f-abc6-33be1406fd9a"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:11:58 crc kubenswrapper[4824]: I1006 10:11:58.919797 4824 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/65505c4e-e82d-485f-abc6-33be1406fd9a-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 10:11:58 crc kubenswrapper[4824]: I1006 10:11:58.919833 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-djsm9\" (UniqueName: \"kubernetes.io/projected/65505c4e-e82d-485f-abc6-33be1406fd9a-kube-api-access-djsm9\") on node \"crc\" DevicePath \"\"" Oct 06 10:11:58 crc kubenswrapper[4824]: I1006 10:11:58.919848 4824 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/65505c4e-e82d-485f-abc6-33be1406fd9a-util\") on node \"crc\" DevicePath \"\"" Oct 06 10:11:59 crc kubenswrapper[4824]: I1006 10:11:59.488653 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/a842b0520c38c46f4cb51faa08e79e3c8193222ff536aee10057c0eb8cqhcsj" event={"ID":"65505c4e-e82d-485f-abc6-33be1406fd9a","Type":"ContainerDied","Data":"97ce47cbb7fca8fd8dde595674315a7f200ce0991f807a7a03ebb7435f00a4c0"} Oct 06 10:11:59 crc kubenswrapper[4824]: I1006 10:11:59.488722 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="97ce47cbb7fca8fd8dde595674315a7f200ce0991f807a7a03ebb7435f00a4c0" Oct 06 10:11:59 crc kubenswrapper[4824]: I1006 10:11:59.488750 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/a842b0520c38c46f4cb51faa08e79e3c8193222ff536aee10057c0eb8cqhcsj" Oct 06 10:12:01 crc kubenswrapper[4824]: I1006 10:12:01.497630 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-mvt5t" Oct 06 10:12:01 crc kubenswrapper[4824]: I1006 10:12:01.498352 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-mvt5t" Oct 06 10:12:01 crc kubenswrapper[4824]: I1006 10:12:01.563794 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-mvt5t" Oct 06 10:12:02 crc kubenswrapper[4824]: I1006 10:12:02.589005 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-mvt5t" Oct 06 10:12:04 crc kubenswrapper[4824]: I1006 10:12:04.761866 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-mvt5t"] Oct 06 10:12:04 crc kubenswrapper[4824]: I1006 10:12:04.762785 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-mvt5t" podUID="fc277ec2-0d30-472a-af67-7019f23e7a3c" containerName="registry-server" containerID="cri-o://5f26630a6457aac156ba508c4f5a2af5ab64aecca7d7b12b883e34162387e91d" gracePeriod=2 Oct 06 10:12:05 crc kubenswrapper[4824]: I1006 10:12:05.232300 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-58f9547c95-dncvl"] Oct 06 10:12:05 crc kubenswrapper[4824]: E1006 10:12:05.232912 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65505c4e-e82d-485f-abc6-33be1406fd9a" containerName="pull" Oct 06 10:12:05 crc kubenswrapper[4824]: I1006 10:12:05.232929 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="65505c4e-e82d-485f-abc6-33be1406fd9a" containerName="pull" Oct 06 10:12:05 crc kubenswrapper[4824]: E1006 10:12:05.232941 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65505c4e-e82d-485f-abc6-33be1406fd9a" containerName="extract" Oct 06 10:12:05 crc kubenswrapper[4824]: I1006 10:12:05.232948 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="65505c4e-e82d-485f-abc6-33be1406fd9a" containerName="extract" Oct 06 10:12:05 crc kubenswrapper[4824]: E1006 10:12:05.232958 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65505c4e-e82d-485f-abc6-33be1406fd9a" containerName="util" Oct 06 10:12:05 crc kubenswrapper[4824]: I1006 10:12:05.232965 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="65505c4e-e82d-485f-abc6-33be1406fd9a" containerName="util" Oct 06 10:12:05 crc kubenswrapper[4824]: I1006 10:12:05.238249 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="65505c4e-e82d-485f-abc6-33be1406fd9a" containerName="extract" Oct 06 10:12:05 crc kubenswrapper[4824]: I1006 10:12:05.239009 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-58f9547c95-dncvl" Oct 06 10:12:05 crc kubenswrapper[4824]: I1006 10:12:05.244318 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-4d5wd" Oct 06 10:12:05 crc kubenswrapper[4824]: I1006 10:12:05.263637 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-58f9547c95-dncvl"] Oct 06 10:12:05 crc kubenswrapper[4824]: I1006 10:12:05.337762 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mvt5t" Oct 06 10:12:05 crc kubenswrapper[4824]: I1006 10:12:05.434005 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fc277ec2-0d30-472a-af67-7019f23e7a3c-utilities\") pod \"fc277ec2-0d30-472a-af67-7019f23e7a3c\" (UID: \"fc277ec2-0d30-472a-af67-7019f23e7a3c\") " Oct 06 10:12:05 crc kubenswrapper[4824]: I1006 10:12:05.434156 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rw2jm\" (UniqueName: \"kubernetes.io/projected/fc277ec2-0d30-472a-af67-7019f23e7a3c-kube-api-access-rw2jm\") pod \"fc277ec2-0d30-472a-af67-7019f23e7a3c\" (UID: \"fc277ec2-0d30-472a-af67-7019f23e7a3c\") " Oct 06 10:12:05 crc kubenswrapper[4824]: I1006 10:12:05.434217 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fc277ec2-0d30-472a-af67-7019f23e7a3c-catalog-content\") pod \"fc277ec2-0d30-472a-af67-7019f23e7a3c\" (UID: \"fc277ec2-0d30-472a-af67-7019f23e7a3c\") " Oct 06 10:12:05 crc kubenswrapper[4824]: I1006 10:12:05.434380 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xvzkl\" (UniqueName: \"kubernetes.io/projected/c1fed37a-a4ba-43eb-82fa-5b04f6a39ead-kube-api-access-xvzkl\") pod \"openstack-operator-controller-operator-58f9547c95-dncvl\" (UID: \"c1fed37a-a4ba-43eb-82fa-5b04f6a39ead\") " pod="openstack-operators/openstack-operator-controller-operator-58f9547c95-dncvl" Oct 06 10:12:05 crc kubenswrapper[4824]: I1006 10:12:05.434841 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fc277ec2-0d30-472a-af67-7019f23e7a3c-utilities" (OuterVolumeSpecName: "utilities") pod "fc277ec2-0d30-472a-af67-7019f23e7a3c" (UID: "fc277ec2-0d30-472a-af67-7019f23e7a3c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:12:05 crc kubenswrapper[4824]: I1006 10:12:05.442037 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fc277ec2-0d30-472a-af67-7019f23e7a3c-kube-api-access-rw2jm" (OuterVolumeSpecName: "kube-api-access-rw2jm") pod "fc277ec2-0d30-472a-af67-7019f23e7a3c" (UID: "fc277ec2-0d30-472a-af67-7019f23e7a3c"). InnerVolumeSpecName "kube-api-access-rw2jm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:12:05 crc kubenswrapper[4824]: I1006 10:12:05.476784 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fc277ec2-0d30-472a-af67-7019f23e7a3c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fc277ec2-0d30-472a-af67-7019f23e7a3c" (UID: "fc277ec2-0d30-472a-af67-7019f23e7a3c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:12:05 crc kubenswrapper[4824]: I1006 10:12:05.535383 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xvzkl\" (UniqueName: \"kubernetes.io/projected/c1fed37a-a4ba-43eb-82fa-5b04f6a39ead-kube-api-access-xvzkl\") pod \"openstack-operator-controller-operator-58f9547c95-dncvl\" (UID: \"c1fed37a-a4ba-43eb-82fa-5b04f6a39ead\") " pod="openstack-operators/openstack-operator-controller-operator-58f9547c95-dncvl" Oct 06 10:12:05 crc kubenswrapper[4824]: I1006 10:12:05.535488 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fc277ec2-0d30-472a-af67-7019f23e7a3c-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 06 10:12:05 crc kubenswrapper[4824]: I1006 10:12:05.535502 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fc277ec2-0d30-472a-af67-7019f23e7a3c-utilities\") on node \"crc\" DevicePath \"\"" Oct 06 10:12:05 crc kubenswrapper[4824]: I1006 10:12:05.535510 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rw2jm\" (UniqueName: \"kubernetes.io/projected/fc277ec2-0d30-472a-af67-7019f23e7a3c-kube-api-access-rw2jm\") on node \"crc\" DevicePath \"\"" Oct 06 10:12:05 crc kubenswrapper[4824]: I1006 10:12:05.543314 4824 generic.go:334] "Generic (PLEG): container finished" podID="fc277ec2-0d30-472a-af67-7019f23e7a3c" containerID="5f26630a6457aac156ba508c4f5a2af5ab64aecca7d7b12b883e34162387e91d" exitCode=0 Oct 06 10:12:05 crc kubenswrapper[4824]: I1006 10:12:05.543413 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mvt5t" event={"ID":"fc277ec2-0d30-472a-af67-7019f23e7a3c","Type":"ContainerDied","Data":"5f26630a6457aac156ba508c4f5a2af5ab64aecca7d7b12b883e34162387e91d"} Oct 06 10:12:05 crc kubenswrapper[4824]: I1006 10:12:05.543441 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mvt5t" Oct 06 10:12:05 crc kubenswrapper[4824]: I1006 10:12:05.543471 4824 scope.go:117] "RemoveContainer" containerID="5f26630a6457aac156ba508c4f5a2af5ab64aecca7d7b12b883e34162387e91d" Oct 06 10:12:05 crc kubenswrapper[4824]: I1006 10:12:05.543455 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mvt5t" event={"ID":"fc277ec2-0d30-472a-af67-7019f23e7a3c","Type":"ContainerDied","Data":"6233fec9f83ed72e1341fc9bc4bdd4a4c7ee44df3d3343785c5c393a59141b9d"} Oct 06 10:12:05 crc kubenswrapper[4824]: I1006 10:12:05.571362 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xvzkl\" (UniqueName: \"kubernetes.io/projected/c1fed37a-a4ba-43eb-82fa-5b04f6a39ead-kube-api-access-xvzkl\") pod \"openstack-operator-controller-operator-58f9547c95-dncvl\" (UID: \"c1fed37a-a4ba-43eb-82fa-5b04f6a39ead\") " pod="openstack-operators/openstack-operator-controller-operator-58f9547c95-dncvl" Oct 06 10:12:05 crc kubenswrapper[4824]: I1006 10:12:05.573317 4824 scope.go:117] "RemoveContainer" containerID="471b26f61254f725b22745dbf5022817e73ebc1ba5b85f52326c6b7307ff81ea" Oct 06 10:12:05 crc kubenswrapper[4824]: I1006 10:12:05.607151 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-mvt5t"] Oct 06 10:12:05 crc kubenswrapper[4824]: I1006 10:12:05.616639 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-mvt5t"] Oct 06 10:12:05 crc kubenswrapper[4824]: I1006 10:12:05.637131 4824 scope.go:117] "RemoveContainer" containerID="be16a5dc7bdfcf42b942d0b49c6afdf8f6c3ba07550089ab6dca2bbe82253933" Oct 06 10:12:05 crc kubenswrapper[4824]: I1006 10:12:05.664542 4824 scope.go:117] "RemoveContainer" containerID="5f26630a6457aac156ba508c4f5a2af5ab64aecca7d7b12b883e34162387e91d" Oct 06 10:12:05 crc kubenswrapper[4824]: E1006 10:12:05.670774 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5f26630a6457aac156ba508c4f5a2af5ab64aecca7d7b12b883e34162387e91d\": container with ID starting with 5f26630a6457aac156ba508c4f5a2af5ab64aecca7d7b12b883e34162387e91d not found: ID does not exist" containerID="5f26630a6457aac156ba508c4f5a2af5ab64aecca7d7b12b883e34162387e91d" Oct 06 10:12:05 crc kubenswrapper[4824]: I1006 10:12:05.670826 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5f26630a6457aac156ba508c4f5a2af5ab64aecca7d7b12b883e34162387e91d"} err="failed to get container status \"5f26630a6457aac156ba508c4f5a2af5ab64aecca7d7b12b883e34162387e91d\": rpc error: code = NotFound desc = could not find container \"5f26630a6457aac156ba508c4f5a2af5ab64aecca7d7b12b883e34162387e91d\": container with ID starting with 5f26630a6457aac156ba508c4f5a2af5ab64aecca7d7b12b883e34162387e91d not found: ID does not exist" Oct 06 10:12:05 crc kubenswrapper[4824]: I1006 10:12:05.670857 4824 scope.go:117] "RemoveContainer" containerID="471b26f61254f725b22745dbf5022817e73ebc1ba5b85f52326c6b7307ff81ea" Oct 06 10:12:05 crc kubenswrapper[4824]: E1006 10:12:05.671401 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"471b26f61254f725b22745dbf5022817e73ebc1ba5b85f52326c6b7307ff81ea\": container with ID starting with 471b26f61254f725b22745dbf5022817e73ebc1ba5b85f52326c6b7307ff81ea not found: ID does not exist" containerID="471b26f61254f725b22745dbf5022817e73ebc1ba5b85f52326c6b7307ff81ea" Oct 06 10:12:05 crc kubenswrapper[4824]: I1006 10:12:05.671460 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"471b26f61254f725b22745dbf5022817e73ebc1ba5b85f52326c6b7307ff81ea"} err="failed to get container status \"471b26f61254f725b22745dbf5022817e73ebc1ba5b85f52326c6b7307ff81ea\": rpc error: code = NotFound desc = could not find container \"471b26f61254f725b22745dbf5022817e73ebc1ba5b85f52326c6b7307ff81ea\": container with ID starting with 471b26f61254f725b22745dbf5022817e73ebc1ba5b85f52326c6b7307ff81ea not found: ID does not exist" Oct 06 10:12:05 crc kubenswrapper[4824]: I1006 10:12:05.671505 4824 scope.go:117] "RemoveContainer" containerID="be16a5dc7bdfcf42b942d0b49c6afdf8f6c3ba07550089ab6dca2bbe82253933" Oct 06 10:12:05 crc kubenswrapper[4824]: E1006 10:12:05.671821 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"be16a5dc7bdfcf42b942d0b49c6afdf8f6c3ba07550089ab6dca2bbe82253933\": container with ID starting with be16a5dc7bdfcf42b942d0b49c6afdf8f6c3ba07550089ab6dca2bbe82253933 not found: ID does not exist" containerID="be16a5dc7bdfcf42b942d0b49c6afdf8f6c3ba07550089ab6dca2bbe82253933" Oct 06 10:12:05 crc kubenswrapper[4824]: I1006 10:12:05.671849 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"be16a5dc7bdfcf42b942d0b49c6afdf8f6c3ba07550089ab6dca2bbe82253933"} err="failed to get container status \"be16a5dc7bdfcf42b942d0b49c6afdf8f6c3ba07550089ab6dca2bbe82253933\": rpc error: code = NotFound desc = could not find container \"be16a5dc7bdfcf42b942d0b49c6afdf8f6c3ba07550089ab6dca2bbe82253933\": container with ID starting with be16a5dc7bdfcf42b942d0b49c6afdf8f6c3ba07550089ab6dca2bbe82253933 not found: ID does not exist" Oct 06 10:12:05 crc kubenswrapper[4824]: I1006 10:12:05.861682 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-58f9547c95-dncvl" Oct 06 10:12:06 crc kubenswrapper[4824]: I1006 10:12:06.474615 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-58f9547c95-dncvl"] Oct 06 10:12:06 crc kubenswrapper[4824]: I1006 10:12:06.553935 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-58f9547c95-dncvl" event={"ID":"c1fed37a-a4ba-43eb-82fa-5b04f6a39ead","Type":"ContainerStarted","Data":"b34734af90d97558e5b44f03a46360040f7ee4e3ac3a485f08891a0710c99522"} Oct 06 10:12:07 crc kubenswrapper[4824]: I1006 10:12:07.286347 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fc277ec2-0d30-472a-af67-7019f23e7a3c" path="/var/lib/kubelet/pods/fc277ec2-0d30-472a-af67-7019f23e7a3c/volumes" Oct 06 10:12:11 crc kubenswrapper[4824]: I1006 10:12:11.593552 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-58f9547c95-dncvl" event={"ID":"c1fed37a-a4ba-43eb-82fa-5b04f6a39ead","Type":"ContainerStarted","Data":"e6050395d9a6946b8f69da2f8d68b8467815d3239890918abe8a85f7dd7aa105"} Oct 06 10:12:13 crc kubenswrapper[4824]: I1006 10:12:13.915719 4824 patch_prober.go:28] interesting pod/machine-config-daemon-khgzw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 10:12:13 crc kubenswrapper[4824]: I1006 10:12:13.916247 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 10:12:13 crc kubenswrapper[4824]: I1006 10:12:13.916318 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" Oct 06 10:12:13 crc kubenswrapper[4824]: I1006 10:12:13.917389 4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8fa1ede4cec498bdfe66760d5e74fd58534149105068e959659c8b228527cc04"} pod="openshift-machine-config-operator/machine-config-daemon-khgzw" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 06 10:12:13 crc kubenswrapper[4824]: I1006 10:12:13.917491 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" containerName="machine-config-daemon" containerID="cri-o://8fa1ede4cec498bdfe66760d5e74fd58534149105068e959659c8b228527cc04" gracePeriod=600 Oct 06 10:12:14 crc kubenswrapper[4824]: I1006 10:12:14.638884 4824 generic.go:334] "Generic (PLEG): container finished" podID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" containerID="8fa1ede4cec498bdfe66760d5e74fd58534149105068e959659c8b228527cc04" exitCode=0 Oct 06 10:12:14 crc kubenswrapper[4824]: I1006 10:12:14.638941 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" event={"ID":"e1140c8c-93f1-4fce-8c06-a69561ff8a6d","Type":"ContainerDied","Data":"8fa1ede4cec498bdfe66760d5e74fd58534149105068e959659c8b228527cc04"} Oct 06 10:12:14 crc kubenswrapper[4824]: I1006 10:12:14.639000 4824 scope.go:117] "RemoveContainer" containerID="d6acecda0f23f881c285040e9b7cdcc459c602934016ff58bff3f9c6edb5dbaf" Oct 06 10:12:16 crc kubenswrapper[4824]: I1006 10:12:16.662572 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-58f9547c95-dncvl" event={"ID":"c1fed37a-a4ba-43eb-82fa-5b04f6a39ead","Type":"ContainerStarted","Data":"cb7199950399848af4f58c665c895d549d0a1150f413556719b7fd1a2a87b5ab"} Oct 06 10:12:16 crc kubenswrapper[4824]: I1006 10:12:16.664196 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-58f9547c95-dncvl" Oct 06 10:12:16 crc kubenswrapper[4824]: I1006 10:12:16.666673 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" event={"ID":"e1140c8c-93f1-4fce-8c06-a69561ff8a6d","Type":"ContainerStarted","Data":"797b5c3c4be7e935c4c45dea62e8a371c48208db7d2ac607bd9b1a5fbc29630b"} Oct 06 10:12:16 crc kubenswrapper[4824]: I1006 10:12:16.667113 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-58f9547c95-dncvl" Oct 06 10:12:16 crc kubenswrapper[4824]: I1006 10:12:16.713826 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-58f9547c95-dncvl" podStartSLOduration=2.349048835 podStartE2EDuration="11.713805984s" podCreationTimestamp="2025-10-06 10:12:05 +0000 UTC" firstStartedPulling="2025-10-06 10:12:06.490365714 +0000 UTC m=+835.854788575" lastFinishedPulling="2025-10-06 10:12:15.855122853 +0000 UTC m=+845.219545724" observedRunningTime="2025-10-06 10:12:16.699905919 +0000 UTC m=+846.064328800" watchObservedRunningTime="2025-10-06 10:12:16.713805984 +0000 UTC m=+846.078228855" Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.155969 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-7d4d4f8d-4kn7z"] Oct 06 10:12:52 crc kubenswrapper[4824]: E1006 10:12:52.157179 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc277ec2-0d30-472a-af67-7019f23e7a3c" containerName="extract-content" Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.157202 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc277ec2-0d30-472a-af67-7019f23e7a3c" containerName="extract-content" Oct 06 10:12:52 crc kubenswrapper[4824]: E1006 10:12:52.157228 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc277ec2-0d30-472a-af67-7019f23e7a3c" containerName="registry-server" Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.157241 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc277ec2-0d30-472a-af67-7019f23e7a3c" containerName="registry-server" Oct 06 10:12:52 crc kubenswrapper[4824]: E1006 10:12:52.157267 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc277ec2-0d30-472a-af67-7019f23e7a3c" containerName="extract-utilities" Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.157281 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc277ec2-0d30-472a-af67-7019f23e7a3c" containerName="extract-utilities" Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.157529 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="fc277ec2-0d30-472a-af67-7019f23e7a3c" containerName="registry-server" Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.158752 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-7d4d4f8d-4kn7z" Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.160559 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-5f7c849b98-fn8nc"] Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.164161 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-5f7c849b98-fn8nc" Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.179771 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-2fzsv" Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.179795 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-qt2gc" Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.181462 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-5f7c849b98-fn8nc"] Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.191602 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-7d4d4f8d-4kn7z"] Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.201076 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-75dfd9b554-hktsj"] Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.202092 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-75dfd9b554-hktsj" Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.211453 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-698f7" Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.227559 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-5568b5d68-r996x"] Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.236303 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-75dfd9b554-hktsj"] Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.236437 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-5568b5d68-r996x" Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.241124 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-msxm5" Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.253797 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-5568b5d68-r996x"] Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.263852 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-st4dd\" (UniqueName: \"kubernetes.io/projected/e60e7a0d-ccc6-4b1a-b645-a4802b21d48e-kube-api-access-st4dd\") pod \"barbican-operator-controller-manager-5f7c849b98-fn8nc\" (UID: \"e60e7a0d-ccc6-4b1a-b645-a4802b21d48e\") " pod="openstack-operators/barbican-operator-controller-manager-5f7c849b98-fn8nc" Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.263953 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bqnsx\" (UniqueName: \"kubernetes.io/projected/b3c0daed-de3b-415e-9166-756cb0f5cab7-kube-api-access-bqnsx\") pod \"cinder-operator-controller-manager-7d4d4f8d-4kn7z\" (UID: \"b3c0daed-de3b-415e-9166-756cb0f5cab7\") " pod="openstack-operators/cinder-operator-controller-manager-7d4d4f8d-4kn7z" Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.280504 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-8f58bc9db-tp4b2"] Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.281677 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-8f58bc9db-tp4b2" Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.285335 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-ld7hx" Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.294938 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-8f58bc9db-tp4b2"] Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.306842 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-658588b8c9-6884h"] Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.308258 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-658588b8c9-6884h" Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.311417 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-wm6rq" Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.311681 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.320730 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-54876c876f-swsgf"] Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.323001 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-54876c876f-swsgf" Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.341238 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-pzw6t" Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.368798 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ddrtj\" (UniqueName: \"kubernetes.io/projected/4ca61d7c-741e-4028-8a75-d26a7b4cbd3a-kube-api-access-ddrtj\") pod \"designate-operator-controller-manager-75dfd9b554-hktsj\" (UID: \"4ca61d7c-741e-4028-8a75-d26a7b4cbd3a\") " pod="openstack-operators/designate-operator-controller-manager-75dfd9b554-hktsj" Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.369281 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-st4dd\" (UniqueName: \"kubernetes.io/projected/e60e7a0d-ccc6-4b1a-b645-a4802b21d48e-kube-api-access-st4dd\") pod \"barbican-operator-controller-manager-5f7c849b98-fn8nc\" (UID: \"e60e7a0d-ccc6-4b1a-b645-a4802b21d48e\") " pod="openstack-operators/barbican-operator-controller-manager-5f7c849b98-fn8nc" Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.369560 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bqnsx\" (UniqueName: \"kubernetes.io/projected/b3c0daed-de3b-415e-9166-756cb0f5cab7-kube-api-access-bqnsx\") pod \"cinder-operator-controller-manager-7d4d4f8d-4kn7z\" (UID: \"b3c0daed-de3b-415e-9166-756cb0f5cab7\") " pod="openstack-operators/cinder-operator-controller-manager-7d4d4f8d-4kn7z" Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.369627 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d7j2q\" (UniqueName: \"kubernetes.io/projected/e8ad16ea-027c-4ca7-accd-928ba6faf830-kube-api-access-d7j2q\") pod \"glance-operator-controller-manager-5568b5d68-r996x\" (UID: \"e8ad16ea-027c-4ca7-accd-928ba6faf830\") " pod="openstack-operators/glance-operator-controller-manager-5568b5d68-r996x" Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.387501 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-658588b8c9-6884h"] Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.419792 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-699b87f775-8mvml"] Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.421490 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-699b87f775-8mvml" Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.423478 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-ldjnw" Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.424795 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bqnsx\" (UniqueName: \"kubernetes.io/projected/b3c0daed-de3b-415e-9166-756cb0f5cab7-kube-api-access-bqnsx\") pod \"cinder-operator-controller-manager-7d4d4f8d-4kn7z\" (UID: \"b3c0daed-de3b-415e-9166-756cb0f5cab7\") " pod="openstack-operators/cinder-operator-controller-manager-7d4d4f8d-4kn7z" Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.429298 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-st4dd\" (UniqueName: \"kubernetes.io/projected/e60e7a0d-ccc6-4b1a-b645-a4802b21d48e-kube-api-access-st4dd\") pod \"barbican-operator-controller-manager-5f7c849b98-fn8nc\" (UID: \"e60e7a0d-ccc6-4b1a-b645-a4802b21d48e\") " pod="openstack-operators/barbican-operator-controller-manager-5f7c849b98-fn8nc" Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.438344 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-54876c876f-swsgf"] Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.454034 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-655d88ccb9-t8sjr"] Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.455317 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-655d88ccb9-t8sjr" Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.458831 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-hdxcw" Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.464611 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-655d88ccb9-t8sjr"] Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.477920 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d7j2q\" (UniqueName: \"kubernetes.io/projected/e8ad16ea-027c-4ca7-accd-928ba6faf830-kube-api-access-d7j2q\") pod \"glance-operator-controller-manager-5568b5d68-r996x\" (UID: \"e8ad16ea-027c-4ca7-accd-928ba6faf830\") " pod="openstack-operators/glance-operator-controller-manager-5568b5d68-r996x" Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.477999 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2l2sn\" (UniqueName: \"kubernetes.io/projected/09c09b76-0deb-44f1-bc93-b4479b7baa36-kube-api-access-2l2sn\") pod \"ironic-operator-controller-manager-699b87f775-8mvml\" (UID: \"09c09b76-0deb-44f1-bc93-b4479b7baa36\") " pod="openstack-operators/ironic-operator-controller-manager-699b87f775-8mvml" Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.478029 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/98cddaff-3395-473c-afd4-02c849951d51-cert\") pod \"infra-operator-controller-manager-658588b8c9-6884h\" (UID: \"98cddaff-3395-473c-afd4-02c849951d51\") " pod="openstack-operators/infra-operator-controller-manager-658588b8c9-6884h" Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.478075 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z77dc\" (UniqueName: \"kubernetes.io/projected/dd382562-c508-4b17-ae00-5db8abe20596-kube-api-access-z77dc\") pod \"keystone-operator-controller-manager-655d88ccb9-t8sjr\" (UID: \"dd382562-c508-4b17-ae00-5db8abe20596\") " pod="openstack-operators/keystone-operator-controller-manager-655d88ccb9-t8sjr" Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.478149 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vfl45\" (UniqueName: \"kubernetes.io/projected/76419802-f7a7-4270-9460-3cc1a9f7f667-kube-api-access-vfl45\") pod \"heat-operator-controller-manager-8f58bc9db-tp4b2\" (UID: \"76419802-f7a7-4270-9460-3cc1a9f7f667\") " pod="openstack-operators/heat-operator-controller-manager-8f58bc9db-tp4b2" Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.478192 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ddrtj\" (UniqueName: \"kubernetes.io/projected/4ca61d7c-741e-4028-8a75-d26a7b4cbd3a-kube-api-access-ddrtj\") pod \"designate-operator-controller-manager-75dfd9b554-hktsj\" (UID: \"4ca61d7c-741e-4028-8a75-d26a7b4cbd3a\") " pod="openstack-operators/designate-operator-controller-manager-75dfd9b554-hktsj" Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.478295 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gnx2m\" (UniqueName: \"kubernetes.io/projected/98cddaff-3395-473c-afd4-02c849951d51-kube-api-access-gnx2m\") pod \"infra-operator-controller-manager-658588b8c9-6884h\" (UID: \"98cddaff-3395-473c-afd4-02c849951d51\") " pod="openstack-operators/infra-operator-controller-manager-658588b8c9-6884h" Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.478362 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ldvzb\" (UniqueName: \"kubernetes.io/projected/67824860-53b9-4c9c-95a1-955d9139f6e8-kube-api-access-ldvzb\") pod \"horizon-operator-controller-manager-54876c876f-swsgf\" (UID: \"67824860-53b9-4c9c-95a1-955d9139f6e8\") " pod="openstack-operators/horizon-operator-controller-manager-54876c876f-swsgf" Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.478924 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-699b87f775-8mvml"] Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.496114 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-6cd6d7bdf5-4zc5b"] Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.497346 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-6cd6d7bdf5-4zc5b" Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.504206 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-v9z6t" Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.509420 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-65d89cfd9f-gts97"] Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.510381 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d7j2q\" (UniqueName: \"kubernetes.io/projected/e8ad16ea-027c-4ca7-accd-928ba6faf830-kube-api-access-d7j2q\") pod \"glance-operator-controller-manager-5568b5d68-r996x\" (UID: \"e8ad16ea-027c-4ca7-accd-928ba6faf830\") " pod="openstack-operators/glance-operator-controller-manager-5568b5d68-r996x" Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.510561 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-65d89cfd9f-gts97" Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.514501 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-7d4d4f8d-4kn7z" Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.515436 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-ndk4x" Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.519254 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ddrtj\" (UniqueName: \"kubernetes.io/projected/4ca61d7c-741e-4028-8a75-d26a7b4cbd3a-kube-api-access-ddrtj\") pod \"designate-operator-controller-manager-75dfd9b554-hktsj\" (UID: \"4ca61d7c-741e-4028-8a75-d26a7b4cbd3a\") " pod="openstack-operators/designate-operator-controller-manager-75dfd9b554-hktsj" Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.533073 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-8d984cc4d-82bf8"] Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.534445 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-8d984cc4d-82bf8" Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.535351 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-5f7c849b98-fn8nc" Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.536534 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-9szl9" Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.547136 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-6cd6d7bdf5-4zc5b"] Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.547502 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-75dfd9b554-hktsj" Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.563070 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-65d89cfd9f-gts97"] Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.578843 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2l2sn\" (UniqueName: \"kubernetes.io/projected/09c09b76-0deb-44f1-bc93-b4479b7baa36-kube-api-access-2l2sn\") pod \"ironic-operator-controller-manager-699b87f775-8mvml\" (UID: \"09c09b76-0deb-44f1-bc93-b4479b7baa36\") " pod="openstack-operators/ironic-operator-controller-manager-699b87f775-8mvml" Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.579475 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/98cddaff-3395-473c-afd4-02c849951d51-cert\") pod \"infra-operator-controller-manager-658588b8c9-6884h\" (UID: \"98cddaff-3395-473c-afd4-02c849951d51\") " pod="openstack-operators/infra-operator-controller-manager-658588b8c9-6884h" Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.579575 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z77dc\" (UniqueName: \"kubernetes.io/projected/dd382562-c508-4b17-ae00-5db8abe20596-kube-api-access-z77dc\") pod \"keystone-operator-controller-manager-655d88ccb9-t8sjr\" (UID: \"dd382562-c508-4b17-ae00-5db8abe20596\") " pod="openstack-operators/keystone-operator-controller-manager-655d88ccb9-t8sjr" Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.579661 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vfl45\" (UniqueName: \"kubernetes.io/projected/76419802-f7a7-4270-9460-3cc1a9f7f667-kube-api-access-vfl45\") pod \"heat-operator-controller-manager-8f58bc9db-tp4b2\" (UID: \"76419802-f7a7-4270-9460-3cc1a9f7f667\") " pod="openstack-operators/heat-operator-controller-manager-8f58bc9db-tp4b2" Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.579759 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gnx2m\" (UniqueName: \"kubernetes.io/projected/98cddaff-3395-473c-afd4-02c849951d51-kube-api-access-gnx2m\") pod \"infra-operator-controller-manager-658588b8c9-6884h\" (UID: \"98cddaff-3395-473c-afd4-02c849951d51\") " pod="openstack-operators/infra-operator-controller-manager-658588b8c9-6884h" Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.579841 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d2gvb\" (UniqueName: \"kubernetes.io/projected/84a43a6c-6456-47ed-ae93-e4476d3cc978-kube-api-access-d2gvb\") pod \"mariadb-operator-controller-manager-6cd6d7bdf5-4zc5b\" (UID: \"84a43a6c-6456-47ed-ae93-e4476d3cc978\") " pod="openstack-operators/mariadb-operator-controller-manager-6cd6d7bdf5-4zc5b" Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.579927 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ldvzb\" (UniqueName: \"kubernetes.io/projected/67824860-53b9-4c9c-95a1-955d9139f6e8-kube-api-access-ldvzb\") pod \"horizon-operator-controller-manager-54876c876f-swsgf\" (UID: \"67824860-53b9-4c9c-95a1-955d9139f6e8\") " pod="openstack-operators/horizon-operator-controller-manager-54876c876f-swsgf" Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.580027 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nwgzm\" (UniqueName: \"kubernetes.io/projected/36dbefcc-0dac-4435-aaac-5248cd2eb209-kube-api-access-nwgzm\") pod \"neutron-operator-controller-manager-8d984cc4d-82bf8\" (UID: \"36dbefcc-0dac-4435-aaac-5248cd2eb209\") " pod="openstack-operators/neutron-operator-controller-manager-8d984cc4d-82bf8" Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.580102 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qkwzn\" (UniqueName: \"kubernetes.io/projected/21123da0-cde6-45dd-9f50-f3e44a1f78c2-kube-api-access-qkwzn\") pod \"manila-operator-controller-manager-65d89cfd9f-gts97\" (UID: \"21123da0-cde6-45dd-9f50-f3e44a1f78c2\") " pod="openstack-operators/manila-operator-controller-manager-65d89cfd9f-gts97" Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.585061 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/98cddaff-3395-473c-afd4-02c849951d51-cert\") pod \"infra-operator-controller-manager-658588b8c9-6884h\" (UID: \"98cddaff-3395-473c-afd4-02c849951d51\") " pod="openstack-operators/infra-operator-controller-manager-658588b8c9-6884h" Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.586355 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-8d984cc4d-82bf8"] Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.594934 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-5568b5d68-r996x" Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.603670 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vfl45\" (UniqueName: \"kubernetes.io/projected/76419802-f7a7-4270-9460-3cc1a9f7f667-kube-api-access-vfl45\") pod \"heat-operator-controller-manager-8f58bc9db-tp4b2\" (UID: \"76419802-f7a7-4270-9460-3cc1a9f7f667\") " pod="openstack-operators/heat-operator-controller-manager-8f58bc9db-tp4b2" Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.607389 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ldvzb\" (UniqueName: \"kubernetes.io/projected/67824860-53b9-4c9c-95a1-955d9139f6e8-kube-api-access-ldvzb\") pod \"horizon-operator-controller-manager-54876c876f-swsgf\" (UID: \"67824860-53b9-4c9c-95a1-955d9139f6e8\") " pod="openstack-operators/horizon-operator-controller-manager-54876c876f-swsgf" Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.608196 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-8f58bc9db-tp4b2" Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.611016 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-7c7fc454ff-fjjsf"] Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.612548 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-7c7fc454ff-fjjsf" Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.616644 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-26kbw" Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.624777 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2l2sn\" (UniqueName: \"kubernetes.io/projected/09c09b76-0deb-44f1-bc93-b4479b7baa36-kube-api-access-2l2sn\") pod \"ironic-operator-controller-manager-699b87f775-8mvml\" (UID: \"09c09b76-0deb-44f1-bc93-b4479b7baa36\") " pod="openstack-operators/ironic-operator-controller-manager-699b87f775-8mvml" Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.627279 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z77dc\" (UniqueName: \"kubernetes.io/projected/dd382562-c508-4b17-ae00-5db8abe20596-kube-api-access-z77dc\") pod \"keystone-operator-controller-manager-655d88ccb9-t8sjr\" (UID: \"dd382562-c508-4b17-ae00-5db8abe20596\") " pod="openstack-operators/keystone-operator-controller-manager-655d88ccb9-t8sjr" Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.627347 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gnx2m\" (UniqueName: \"kubernetes.io/projected/98cddaff-3395-473c-afd4-02c849951d51-kube-api-access-gnx2m\") pod \"infra-operator-controller-manager-658588b8c9-6884h\" (UID: \"98cddaff-3395-473c-afd4-02c849951d51\") " pod="openstack-operators/infra-operator-controller-manager-658588b8c9-6884h" Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.630260 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-7c7fc454ff-fjjsf"] Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.651616 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-7468f855d8-kpbcv"] Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.652811 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-7468f855d8-kpbcv" Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.655692 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-76jbt" Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.662838 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-579449c7d5-w7k8s"] Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.664050 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-579449c7d5-w7k8s" Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.665856 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-g9zvg" Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.666121 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-658588b8c9-6884h" Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.667681 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665cxcwhp"] Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.668655 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665cxcwhp" Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.675049 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-v5dcs" Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.675343 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.682540 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-7468f855d8-kpbcv"] Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.683233 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nwgzm\" (UniqueName: \"kubernetes.io/projected/36dbefcc-0dac-4435-aaac-5248cd2eb209-kube-api-access-nwgzm\") pod \"neutron-operator-controller-manager-8d984cc4d-82bf8\" (UID: \"36dbefcc-0dac-4435-aaac-5248cd2eb209\") " pod="openstack-operators/neutron-operator-controller-manager-8d984cc4d-82bf8" Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.683261 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qkwzn\" (UniqueName: \"kubernetes.io/projected/21123da0-cde6-45dd-9f50-f3e44a1f78c2-kube-api-access-qkwzn\") pod \"manila-operator-controller-manager-65d89cfd9f-gts97\" (UID: \"21123da0-cde6-45dd-9f50-f3e44a1f78c2\") " pod="openstack-operators/manila-operator-controller-manager-65d89cfd9f-gts97" Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.683315 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/fa5da9bb-805d-420f-adc6-e948cf910b21-cert\") pod \"openstack-baremetal-operator-controller-manager-5dfbbd665cxcwhp\" (UID: \"fa5da9bb-805d-420f-adc6-e948cf910b21\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665cxcwhp" Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.683414 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d2gvb\" (UniqueName: \"kubernetes.io/projected/84a43a6c-6456-47ed-ae93-e4476d3cc978-kube-api-access-d2gvb\") pod \"mariadb-operator-controller-manager-6cd6d7bdf5-4zc5b\" (UID: \"84a43a6c-6456-47ed-ae93-e4476d3cc978\") " pod="openstack-operators/mariadb-operator-controller-manager-6cd6d7bdf5-4zc5b" Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.683435 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ptpkv\" (UniqueName: \"kubernetes.io/projected/12802c32-cb1d-452b-8240-9a5e1e40a163-kube-api-access-ptpkv\") pod \"ovn-operator-controller-manager-579449c7d5-w7k8s\" (UID: \"12802c32-cb1d-452b-8240-9a5e1e40a163\") " pod="openstack-operators/ovn-operator-controller-manager-579449c7d5-w7k8s" Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.683457 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dn8qm\" (UniqueName: \"kubernetes.io/projected/fa5da9bb-805d-420f-adc6-e948cf910b21-kube-api-access-dn8qm\") pod \"openstack-baremetal-operator-controller-manager-5dfbbd665cxcwhp\" (UID: \"fa5da9bb-805d-420f-adc6-e948cf910b21\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665cxcwhp" Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.683484 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lrwm6\" (UniqueName: \"kubernetes.io/projected/dff987b5-f7ce-4eb5-9778-3cb34663824a-kube-api-access-lrwm6\") pod \"nova-operator-controller-manager-7c7fc454ff-fjjsf\" (UID: \"dff987b5-f7ce-4eb5-9778-3cb34663824a\") " pod="openstack-operators/nova-operator-controller-manager-7c7fc454ff-fjjsf" Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.683501 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lgwt2\" (UniqueName: \"kubernetes.io/projected/c0da600c-cfd5-405a-82ca-ebbf3c474e44-kube-api-access-lgwt2\") pod \"octavia-operator-controller-manager-7468f855d8-kpbcv\" (UID: \"c0da600c-cfd5-405a-82ca-ebbf3c474e44\") " pod="openstack-operators/octavia-operator-controller-manager-7468f855d8-kpbcv" Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.687214 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-54876c876f-swsgf" Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.702848 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-579449c7d5-w7k8s"] Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.703368 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nwgzm\" (UniqueName: \"kubernetes.io/projected/36dbefcc-0dac-4435-aaac-5248cd2eb209-kube-api-access-nwgzm\") pod \"neutron-operator-controller-manager-8d984cc4d-82bf8\" (UID: \"36dbefcc-0dac-4435-aaac-5248cd2eb209\") " pod="openstack-operators/neutron-operator-controller-manager-8d984cc4d-82bf8" Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.706294 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qkwzn\" (UniqueName: \"kubernetes.io/projected/21123da0-cde6-45dd-9f50-f3e44a1f78c2-kube-api-access-qkwzn\") pod \"manila-operator-controller-manager-65d89cfd9f-gts97\" (UID: \"21123da0-cde6-45dd-9f50-f3e44a1f78c2\") " pod="openstack-operators/manila-operator-controller-manager-65d89cfd9f-gts97" Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.707064 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-54689d9f88-8qfvp"] Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.707501 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d2gvb\" (UniqueName: \"kubernetes.io/projected/84a43a6c-6456-47ed-ae93-e4476d3cc978-kube-api-access-d2gvb\") pod \"mariadb-operator-controller-manager-6cd6d7bdf5-4zc5b\" (UID: \"84a43a6c-6456-47ed-ae93-e4476d3cc978\") " pod="openstack-operators/mariadb-operator-controller-manager-6cd6d7bdf5-4zc5b" Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.708488 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-54689d9f88-8qfvp" Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.716219 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-grztj" Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.752836 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665cxcwhp"] Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.771190 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-54689d9f88-8qfvp"] Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.774274 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-699b87f775-8mvml" Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.778279 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-655d88ccb9-t8sjr" Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.785358 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-84c678c567-wzzrh"] Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.786489 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-84c678c567-wzzrh" Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.790800 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-9q8pw" Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.792141 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/fa5da9bb-805d-420f-adc6-e948cf910b21-cert\") pod \"openstack-baremetal-operator-controller-manager-5dfbbd665cxcwhp\" (UID: \"fa5da9bb-805d-420f-adc6-e948cf910b21\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665cxcwhp" Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.792217 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ptpkv\" (UniqueName: \"kubernetes.io/projected/12802c32-cb1d-452b-8240-9a5e1e40a163-kube-api-access-ptpkv\") pod \"ovn-operator-controller-manager-579449c7d5-w7k8s\" (UID: \"12802c32-cb1d-452b-8240-9a5e1e40a163\") " pod="openstack-operators/ovn-operator-controller-manager-579449c7d5-w7k8s" Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.792243 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dn8qm\" (UniqueName: \"kubernetes.io/projected/fa5da9bb-805d-420f-adc6-e948cf910b21-kube-api-access-dn8qm\") pod \"openstack-baremetal-operator-controller-manager-5dfbbd665cxcwhp\" (UID: \"fa5da9bb-805d-420f-adc6-e948cf910b21\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665cxcwhp" Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.792267 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lrwm6\" (UniqueName: \"kubernetes.io/projected/dff987b5-f7ce-4eb5-9778-3cb34663824a-kube-api-access-lrwm6\") pod \"nova-operator-controller-manager-7c7fc454ff-fjjsf\" (UID: \"dff987b5-f7ce-4eb5-9778-3cb34663824a\") " pod="openstack-operators/nova-operator-controller-manager-7c7fc454ff-fjjsf" Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.792285 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lgwt2\" (UniqueName: \"kubernetes.io/projected/c0da600c-cfd5-405a-82ca-ebbf3c474e44-kube-api-access-lgwt2\") pod \"octavia-operator-controller-manager-7468f855d8-kpbcv\" (UID: \"c0da600c-cfd5-405a-82ca-ebbf3c474e44\") " pod="openstack-operators/octavia-operator-controller-manager-7468f855d8-kpbcv" Oct 06 10:12:52 crc kubenswrapper[4824]: E1006 10:12:52.793172 4824 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 06 10:12:52 crc kubenswrapper[4824]: E1006 10:12:52.793493 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fa5da9bb-805d-420f-adc6-e948cf910b21-cert podName:fa5da9bb-805d-420f-adc6-e948cf910b21 nodeName:}" failed. No retries permitted until 2025-10-06 10:12:53.293474686 +0000 UTC m=+882.657897547 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/fa5da9bb-805d-420f-adc6-e948cf910b21-cert") pod "openstack-baremetal-operator-controller-manager-5dfbbd665cxcwhp" (UID: "fa5da9bb-805d-420f-adc6-e948cf910b21") : secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.809164 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-6859f9b676-n84pw"] Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.810331 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-6859f9b676-n84pw" Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.810632 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dn8qm\" (UniqueName: \"kubernetes.io/projected/fa5da9bb-805d-420f-adc6-e948cf910b21-kube-api-access-dn8qm\") pod \"openstack-baremetal-operator-controller-manager-5dfbbd665cxcwhp\" (UID: \"fa5da9bb-805d-420f-adc6-e948cf910b21\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665cxcwhp" Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.817561 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-frpj6" Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.825263 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lgwt2\" (UniqueName: \"kubernetes.io/projected/c0da600c-cfd5-405a-82ca-ebbf3c474e44-kube-api-access-lgwt2\") pod \"octavia-operator-controller-manager-7468f855d8-kpbcv\" (UID: \"c0da600c-cfd5-405a-82ca-ebbf3c474e44\") " pod="openstack-operators/octavia-operator-controller-manager-7468f855d8-kpbcv" Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.830065 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lrwm6\" (UniqueName: \"kubernetes.io/projected/dff987b5-f7ce-4eb5-9778-3cb34663824a-kube-api-access-lrwm6\") pod \"nova-operator-controller-manager-7c7fc454ff-fjjsf\" (UID: \"dff987b5-f7ce-4eb5-9778-3cb34663824a\") " pod="openstack-operators/nova-operator-controller-manager-7c7fc454ff-fjjsf" Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.830558 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ptpkv\" (UniqueName: \"kubernetes.io/projected/12802c32-cb1d-452b-8240-9a5e1e40a163-kube-api-access-ptpkv\") pod \"ovn-operator-controller-manager-579449c7d5-w7k8s\" (UID: \"12802c32-cb1d-452b-8240-9a5e1e40a163\") " pod="openstack-operators/ovn-operator-controller-manager-579449c7d5-w7k8s" Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.851165 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-84c678c567-wzzrh"] Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.886524 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-6859f9b676-n84pw"] Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.898016 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fn9xt\" (UniqueName: \"kubernetes.io/projected/7c4f21a3-07f8-46d8-bd7b-e2a24bf280d8-kube-api-access-fn9xt\") pod \"telemetry-operator-controller-manager-84c678c567-wzzrh\" (UID: \"7c4f21a3-07f8-46d8-bd7b-e2a24bf280d8\") " pod="openstack-operators/telemetry-operator-controller-manager-84c678c567-wzzrh" Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.898114 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ng82w\" (UniqueName: \"kubernetes.io/projected/1778e737-b572-417b-931d-0eec04a89ac9-kube-api-access-ng82w\") pod \"placement-operator-controller-manager-54689d9f88-8qfvp\" (UID: \"1778e737-b572-417b-931d-0eec04a89ac9\") " pod="openstack-operators/placement-operator-controller-manager-54689d9f88-8qfvp" Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.899310 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-6cd6d7bdf5-4zc5b" Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.912854 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-65d89cfd9f-gts97" Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.934294 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-8d984cc4d-82bf8" Oct 06 10:12:52 crc kubenswrapper[4824]: I1006 10:12:52.959515 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-7c7fc454ff-fjjsf" Oct 06 10:12:53 crc kubenswrapper[4824]: I1006 10:12:52.991363 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-7468f855d8-kpbcv" Oct 06 10:12:53 crc kubenswrapper[4824]: I1006 10:12:52.996539 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-5cd5cb47d7-6qgln"] Oct 06 10:12:53 crc kubenswrapper[4824]: I1006 10:12:53.001651 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ng82w\" (UniqueName: \"kubernetes.io/projected/1778e737-b572-417b-931d-0eec04a89ac9-kube-api-access-ng82w\") pod \"placement-operator-controller-manager-54689d9f88-8qfvp\" (UID: \"1778e737-b572-417b-931d-0eec04a89ac9\") " pod="openstack-operators/placement-operator-controller-manager-54689d9f88-8qfvp" Oct 06 10:12:53 crc kubenswrapper[4824]: I1006 10:12:53.001774 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fn9xt\" (UniqueName: \"kubernetes.io/projected/7c4f21a3-07f8-46d8-bd7b-e2a24bf280d8-kube-api-access-fn9xt\") pod \"telemetry-operator-controller-manager-84c678c567-wzzrh\" (UID: \"7c4f21a3-07f8-46d8-bd7b-e2a24bf280d8\") " pod="openstack-operators/telemetry-operator-controller-manager-84c678c567-wzzrh" Oct 06 10:12:53 crc kubenswrapper[4824]: I1006 10:12:53.001812 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kwk7h\" (UniqueName: \"kubernetes.io/projected/b32a09cc-901f-41f2-9912-628bbc33da7b-kube-api-access-kwk7h\") pod \"swift-operator-controller-manager-6859f9b676-n84pw\" (UID: \"b32a09cc-901f-41f2-9912-628bbc33da7b\") " pod="openstack-operators/swift-operator-controller-manager-6859f9b676-n84pw" Oct 06 10:12:53 crc kubenswrapper[4824]: I1006 10:12:53.003491 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-6qgln" Oct 06 10:12:53 crc kubenswrapper[4824]: I1006 10:12:53.007316 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-744c8" Oct 06 10:12:53 crc kubenswrapper[4824]: I1006 10:12:53.008771 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-579449c7d5-w7k8s" Oct 06 10:12:53 crc kubenswrapper[4824]: I1006 10:12:53.020746 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5cd5cb47d7-6qgln"] Oct 06 10:12:53 crc kubenswrapper[4824]: I1006 10:12:53.051417 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ng82w\" (UniqueName: \"kubernetes.io/projected/1778e737-b572-417b-931d-0eec04a89ac9-kube-api-access-ng82w\") pod \"placement-operator-controller-manager-54689d9f88-8qfvp\" (UID: \"1778e737-b572-417b-931d-0eec04a89ac9\") " pod="openstack-operators/placement-operator-controller-manager-54689d9f88-8qfvp" Oct 06 10:12:53 crc kubenswrapper[4824]: I1006 10:12:53.056065 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fn9xt\" (UniqueName: \"kubernetes.io/projected/7c4f21a3-07f8-46d8-bd7b-e2a24bf280d8-kube-api-access-fn9xt\") pod \"telemetry-operator-controller-manager-84c678c567-wzzrh\" (UID: \"7c4f21a3-07f8-46d8-bd7b-e2a24bf280d8\") " pod="openstack-operators/telemetry-operator-controller-manager-84c678c567-wzzrh" Oct 06 10:12:53 crc kubenswrapper[4824]: I1006 10:12:53.079013 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-6cbc6dd547-dglbk"] Oct 06 10:12:53 crc kubenswrapper[4824]: I1006 10:12:53.080263 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-6cbc6dd547-dglbk" Oct 06 10:12:53 crc kubenswrapper[4824]: I1006 10:12:53.082874 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-gv78m" Oct 06 10:12:53 crc kubenswrapper[4824]: I1006 10:12:53.095383 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-6cbc6dd547-dglbk"] Oct 06 10:12:53 crc kubenswrapper[4824]: I1006 10:12:53.102755 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kwk7h\" (UniqueName: \"kubernetes.io/projected/b32a09cc-901f-41f2-9912-628bbc33da7b-kube-api-access-kwk7h\") pod \"swift-operator-controller-manager-6859f9b676-n84pw\" (UID: \"b32a09cc-901f-41f2-9912-628bbc33da7b\") " pod="openstack-operators/swift-operator-controller-manager-6859f9b676-n84pw" Oct 06 10:12:53 crc kubenswrapper[4824]: I1006 10:12:53.125654 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-844bbc956-tjxqg"] Oct 06 10:12:53 crc kubenswrapper[4824]: I1006 10:12:53.126936 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-844bbc956-tjxqg" Oct 06 10:12:53 crc kubenswrapper[4824]: I1006 10:12:53.127908 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kwk7h\" (UniqueName: \"kubernetes.io/projected/b32a09cc-901f-41f2-9912-628bbc33da7b-kube-api-access-kwk7h\") pod \"swift-operator-controller-manager-6859f9b676-n84pw\" (UID: \"b32a09cc-901f-41f2-9912-628bbc33da7b\") " pod="openstack-operators/swift-operator-controller-manager-6859f9b676-n84pw" Oct 06 10:12:53 crc kubenswrapper[4824]: I1006 10:12:53.130035 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-844bbc956-tjxqg"] Oct 06 10:12:53 crc kubenswrapper[4824]: I1006 10:12:53.134135 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Oct 06 10:12:53 crc kubenswrapper[4824]: I1006 10:12:53.134441 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-mh7fg" Oct 06 10:12:53 crc kubenswrapper[4824]: I1006 10:12:53.135221 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-c8wx2"] Oct 06 10:12:53 crc kubenswrapper[4824]: I1006 10:12:53.136433 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-c8wx2" Oct 06 10:12:53 crc kubenswrapper[4824]: I1006 10:12:53.136903 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-84c678c567-wzzrh" Oct 06 10:12:53 crc kubenswrapper[4824]: I1006 10:12:53.138349 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-wsmk7" Oct 06 10:12:53 crc kubenswrapper[4824]: I1006 10:12:53.160332 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-c8wx2"] Oct 06 10:12:53 crc kubenswrapper[4824]: I1006 10:12:53.162692 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-6859f9b676-n84pw" Oct 06 10:12:53 crc kubenswrapper[4824]: I1006 10:12:53.183269 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-7d4d4f8d-4kn7z"] Oct 06 10:12:53 crc kubenswrapper[4824]: W1006 10:12:53.192263 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb3c0daed_de3b_415e_9166_756cb0f5cab7.slice/crio-feb64f7817b4504643a052d58d9eb87166bb737560e34315c92d1d8169aa4751 WatchSource:0}: Error finding container feb64f7817b4504643a052d58d9eb87166bb737560e34315c92d1d8169aa4751: Status 404 returned error can't find the container with id feb64f7817b4504643a052d58d9eb87166bb737560e34315c92d1d8169aa4751 Oct 06 10:12:53 crc kubenswrapper[4824]: I1006 10:12:53.204207 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kghzm\" (UniqueName: \"kubernetes.io/projected/755275e4-c198-49df-90b5-0688a4fb8228-kube-api-access-kghzm\") pod \"test-operator-controller-manager-5cd5cb47d7-6qgln\" (UID: \"755275e4-c198-49df-90b5-0688a4fb8228\") " pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-6qgln" Oct 06 10:12:53 crc kubenswrapper[4824]: I1006 10:12:53.204355 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6dmrw\" (UniqueName: \"kubernetes.io/projected/2a48640f-172c-4e24-8c75-5c36d26ae1aa-kube-api-access-6dmrw\") pod \"watcher-operator-controller-manager-6cbc6dd547-dglbk\" (UID: \"2a48640f-172c-4e24-8c75-5c36d26ae1aa\") " pod="openstack-operators/watcher-operator-controller-manager-6cbc6dd547-dglbk" Oct 06 10:12:53 crc kubenswrapper[4824]: I1006 10:12:53.323720 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/fa5da9bb-805d-420f-adc6-e948cf910b21-cert\") pod \"openstack-baremetal-operator-controller-manager-5dfbbd665cxcwhp\" (UID: \"fa5da9bb-805d-420f-adc6-e948cf910b21\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665cxcwhp" Oct 06 10:12:53 crc kubenswrapper[4824]: I1006 10:12:53.324098 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6dmrw\" (UniqueName: \"kubernetes.io/projected/2a48640f-172c-4e24-8c75-5c36d26ae1aa-kube-api-access-6dmrw\") pod \"watcher-operator-controller-manager-6cbc6dd547-dglbk\" (UID: \"2a48640f-172c-4e24-8c75-5c36d26ae1aa\") " pod="openstack-operators/watcher-operator-controller-manager-6cbc6dd547-dglbk" Oct 06 10:12:53 crc kubenswrapper[4824]: I1006 10:12:53.324164 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-68grf\" (UniqueName: \"kubernetes.io/projected/1f00bc57-93de-431a-adcf-d1e1221121a1-kube-api-access-68grf\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-c8wx2\" (UID: \"1f00bc57-93de-431a-adcf-d1e1221121a1\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-c8wx2" Oct 06 10:12:53 crc kubenswrapper[4824]: I1006 10:12:53.324195 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dl5r2\" (UniqueName: \"kubernetes.io/projected/993a8c7b-2aac-4de6-934d-b591cb6bfffe-kube-api-access-dl5r2\") pod \"openstack-operator-controller-manager-844bbc956-tjxqg\" (UID: \"993a8c7b-2aac-4de6-934d-b591cb6bfffe\") " pod="openstack-operators/openstack-operator-controller-manager-844bbc956-tjxqg" Oct 06 10:12:53 crc kubenswrapper[4824]: I1006 10:12:53.324222 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kghzm\" (UniqueName: \"kubernetes.io/projected/755275e4-c198-49df-90b5-0688a4fb8228-kube-api-access-kghzm\") pod \"test-operator-controller-manager-5cd5cb47d7-6qgln\" (UID: \"755275e4-c198-49df-90b5-0688a4fb8228\") " pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-6qgln" Oct 06 10:12:53 crc kubenswrapper[4824]: I1006 10:12:53.324241 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/993a8c7b-2aac-4de6-934d-b591cb6bfffe-cert\") pod \"openstack-operator-controller-manager-844bbc956-tjxqg\" (UID: \"993a8c7b-2aac-4de6-934d-b591cb6bfffe\") " pod="openstack-operators/openstack-operator-controller-manager-844bbc956-tjxqg" Oct 06 10:12:53 crc kubenswrapper[4824]: I1006 10:12:53.338201 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-5f7c849b98-fn8nc"] Oct 06 10:12:53 crc kubenswrapper[4824]: I1006 10:12:53.346885 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/fa5da9bb-805d-420f-adc6-e948cf910b21-cert\") pod \"openstack-baremetal-operator-controller-manager-5dfbbd665cxcwhp\" (UID: \"fa5da9bb-805d-420f-adc6-e948cf910b21\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665cxcwhp" Oct 06 10:12:53 crc kubenswrapper[4824]: I1006 10:12:53.352735 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-54689d9f88-8qfvp" Oct 06 10:12:53 crc kubenswrapper[4824]: I1006 10:12:53.355617 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6dmrw\" (UniqueName: \"kubernetes.io/projected/2a48640f-172c-4e24-8c75-5c36d26ae1aa-kube-api-access-6dmrw\") pod \"watcher-operator-controller-manager-6cbc6dd547-dglbk\" (UID: \"2a48640f-172c-4e24-8c75-5c36d26ae1aa\") " pod="openstack-operators/watcher-operator-controller-manager-6cbc6dd547-dglbk" Oct 06 10:12:53 crc kubenswrapper[4824]: I1006 10:12:53.370403 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kghzm\" (UniqueName: \"kubernetes.io/projected/755275e4-c198-49df-90b5-0688a4fb8228-kube-api-access-kghzm\") pod \"test-operator-controller-manager-5cd5cb47d7-6qgln\" (UID: \"755275e4-c198-49df-90b5-0688a4fb8228\") " pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-6qgln" Oct 06 10:12:53 crc kubenswrapper[4824]: I1006 10:12:53.423354 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-6cbc6dd547-dglbk" Oct 06 10:12:53 crc kubenswrapper[4824]: I1006 10:12:53.426289 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dl5r2\" (UniqueName: \"kubernetes.io/projected/993a8c7b-2aac-4de6-934d-b591cb6bfffe-kube-api-access-dl5r2\") pod \"openstack-operator-controller-manager-844bbc956-tjxqg\" (UID: \"993a8c7b-2aac-4de6-934d-b591cb6bfffe\") " pod="openstack-operators/openstack-operator-controller-manager-844bbc956-tjxqg" Oct 06 10:12:53 crc kubenswrapper[4824]: I1006 10:12:53.426342 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/993a8c7b-2aac-4de6-934d-b591cb6bfffe-cert\") pod \"openstack-operator-controller-manager-844bbc956-tjxqg\" (UID: \"993a8c7b-2aac-4de6-934d-b591cb6bfffe\") " pod="openstack-operators/openstack-operator-controller-manager-844bbc956-tjxqg" Oct 06 10:12:53 crc kubenswrapper[4824]: I1006 10:12:53.426413 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-68grf\" (UniqueName: \"kubernetes.io/projected/1f00bc57-93de-431a-adcf-d1e1221121a1-kube-api-access-68grf\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-c8wx2\" (UID: \"1f00bc57-93de-431a-adcf-d1e1221121a1\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-c8wx2" Oct 06 10:12:53 crc kubenswrapper[4824]: E1006 10:12:53.426837 4824 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Oct 06 10:12:53 crc kubenswrapper[4824]: E1006 10:12:53.426881 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/993a8c7b-2aac-4de6-934d-b591cb6bfffe-cert podName:993a8c7b-2aac-4de6-934d-b591cb6bfffe nodeName:}" failed. No retries permitted until 2025-10-06 10:12:53.92686633 +0000 UTC m=+883.291289191 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/993a8c7b-2aac-4de6-934d-b591cb6bfffe-cert") pod "openstack-operator-controller-manager-844bbc956-tjxqg" (UID: "993a8c7b-2aac-4de6-934d-b591cb6bfffe") : secret "webhook-server-cert" not found Oct 06 10:12:53 crc kubenswrapper[4824]: I1006 10:12:53.471278 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-68grf\" (UniqueName: \"kubernetes.io/projected/1f00bc57-93de-431a-adcf-d1e1221121a1-kube-api-access-68grf\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-c8wx2\" (UID: \"1f00bc57-93de-431a-adcf-d1e1221121a1\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-c8wx2" Oct 06 10:12:53 crc kubenswrapper[4824]: I1006 10:12:53.474942 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dl5r2\" (UniqueName: \"kubernetes.io/projected/993a8c7b-2aac-4de6-934d-b591cb6bfffe-kube-api-access-dl5r2\") pod \"openstack-operator-controller-manager-844bbc956-tjxqg\" (UID: \"993a8c7b-2aac-4de6-934d-b591cb6bfffe\") " pod="openstack-operators/openstack-operator-controller-manager-844bbc956-tjxqg" Oct 06 10:12:53 crc kubenswrapper[4824]: I1006 10:12:53.502151 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-c8wx2" Oct 06 10:12:53 crc kubenswrapper[4824]: I1006 10:12:53.511482 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-5568b5d68-r996x"] Oct 06 10:12:53 crc kubenswrapper[4824]: I1006 10:12:53.538018 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-8f58bc9db-tp4b2"] Oct 06 10:12:53 crc kubenswrapper[4824]: I1006 10:12:53.541914 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-75dfd9b554-hktsj"] Oct 06 10:12:53 crc kubenswrapper[4824]: I1006 10:12:53.621382 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665cxcwhp" Oct 06 10:12:53 crc kubenswrapper[4824]: I1006 10:12:53.653487 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-6qgln" Oct 06 10:12:53 crc kubenswrapper[4824]: I1006 10:12:53.933536 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/993a8c7b-2aac-4de6-934d-b591cb6bfffe-cert\") pod \"openstack-operator-controller-manager-844bbc956-tjxqg\" (UID: \"993a8c7b-2aac-4de6-934d-b591cb6bfffe\") " pod="openstack-operators/openstack-operator-controller-manager-844bbc956-tjxqg" Oct 06 10:12:53 crc kubenswrapper[4824]: I1006 10:12:53.939766 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/993a8c7b-2aac-4de6-934d-b591cb6bfffe-cert\") pod \"openstack-operator-controller-manager-844bbc956-tjxqg\" (UID: \"993a8c7b-2aac-4de6-934d-b591cb6bfffe\") " pod="openstack-operators/openstack-operator-controller-manager-844bbc956-tjxqg" Oct 06 10:12:53 crc kubenswrapper[4824]: I1006 10:12:53.972608 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-699b87f775-8mvml"] Oct 06 10:12:53 crc kubenswrapper[4824]: I1006 10:12:53.975121 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-7d4d4f8d-4kn7z" event={"ID":"b3c0daed-de3b-415e-9166-756cb0f5cab7","Type":"ContainerStarted","Data":"feb64f7817b4504643a052d58d9eb87166bb737560e34315c92d1d8169aa4751"} Oct 06 10:12:53 crc kubenswrapper[4824]: I1006 10:12:53.976592 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-75dfd9b554-hktsj" event={"ID":"4ca61d7c-741e-4028-8a75-d26a7b4cbd3a","Type":"ContainerStarted","Data":"a5ed73a6f9409aeedfa53d2e3420bd54ec4b316a08edb3f7ff7ba8f2d9f357ce"} Oct 06 10:12:53 crc kubenswrapper[4824]: I1006 10:12:53.977406 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-5f7c849b98-fn8nc" event={"ID":"e60e7a0d-ccc6-4b1a-b645-a4802b21d48e","Type":"ContainerStarted","Data":"206db553a24a5e1c51b7322630f4bf62cc0e843c3947a9143af23d77d4d706f3"} Oct 06 10:12:53 crc kubenswrapper[4824]: I1006 10:12:53.978160 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-8f58bc9db-tp4b2" event={"ID":"76419802-f7a7-4270-9460-3cc1a9f7f667","Type":"ContainerStarted","Data":"0c770244c551e78153fa900e07642902b9bad082891a37dd1b845e6f31cf9c94"} Oct 06 10:12:53 crc kubenswrapper[4824]: I1006 10:12:53.978836 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-5568b5d68-r996x" event={"ID":"e8ad16ea-027c-4ca7-accd-928ba6faf830","Type":"ContainerStarted","Data":"ddebd4ec1acb0bc4746e9c956d8bbee8e17585f67eb966baaecd6afe5d745c98"} Oct 06 10:12:53 crc kubenswrapper[4824]: I1006 10:12:53.993376 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-655d88ccb9-t8sjr"] Oct 06 10:12:54 crc kubenswrapper[4824]: W1006 10:12:54.010746 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod09c09b76_0deb_44f1_bc93_b4479b7baa36.slice/crio-100db90c33ac6555165b2681373418ad2acdd3c2816b5bab600e9589425eac3e WatchSource:0}: Error finding container 100db90c33ac6555165b2681373418ad2acdd3c2816b5bab600e9589425eac3e: Status 404 returned error can't find the container with id 100db90c33ac6555165b2681373418ad2acdd3c2816b5bab600e9589425eac3e Oct 06 10:12:54 crc kubenswrapper[4824]: W1006 10:12:54.013432 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddd382562_c508_4b17_ae00_5db8abe20596.slice/crio-10eb2b148f3165418e90a8485b8002dfe243fd8f8eb5fd4cb44a5cbf54ea2331 WatchSource:0}: Error finding container 10eb2b148f3165418e90a8485b8002dfe243fd8f8eb5fd4cb44a5cbf54ea2331: Status 404 returned error can't find the container with id 10eb2b148f3165418e90a8485b8002dfe243fd8f8eb5fd4cb44a5cbf54ea2331 Oct 06 10:12:54 crc kubenswrapper[4824]: I1006 10:12:54.019279 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-54876c876f-swsgf"] Oct 06 10:12:54 crc kubenswrapper[4824]: I1006 10:12:54.033766 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-658588b8c9-6884h"] Oct 06 10:12:54 crc kubenswrapper[4824]: W1006 10:12:54.033887 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod67824860_53b9_4c9c_95a1_955d9139f6e8.slice/crio-74f72ee4c4b5cf93271f117a499de8de6081426d690fc2a28d6e2ba5a77aa155 WatchSource:0}: Error finding container 74f72ee4c4b5cf93271f117a499de8de6081426d690fc2a28d6e2ba5a77aa155: Status 404 returned error can't find the container with id 74f72ee4c4b5cf93271f117a499de8de6081426d690fc2a28d6e2ba5a77aa155 Oct 06 10:12:54 crc kubenswrapper[4824]: I1006 10:12:54.052945 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-7c7fc454ff-fjjsf"] Oct 06 10:12:54 crc kubenswrapper[4824]: I1006 10:12:54.076609 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-844bbc956-tjxqg" Oct 06 10:12:54 crc kubenswrapper[4824]: I1006 10:12:54.101504 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-579449c7d5-w7k8s"] Oct 06 10:12:54 crc kubenswrapper[4824]: W1006 10:12:54.107329 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod84a43a6c_6456_47ed_ae93_e4476d3cc978.slice/crio-50d8b30830a0fbf685b243d3b0d57f4088de6f7be13aa7af3738469f45eabb72 WatchSource:0}: Error finding container 50d8b30830a0fbf685b243d3b0d57f4088de6f7be13aa7af3738469f45eabb72: Status 404 returned error can't find the container with id 50d8b30830a0fbf685b243d3b0d57f4088de6f7be13aa7af3738469f45eabb72 Oct 06 10:12:54 crc kubenswrapper[4824]: I1006 10:12:54.115079 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-8d984cc4d-82bf8"] Oct 06 10:12:54 crc kubenswrapper[4824]: I1006 10:12:54.120164 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-6cd6d7bdf5-4zc5b"] Oct 06 10:12:54 crc kubenswrapper[4824]: W1006 10:12:54.120598 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc0da600c_cfd5_405a_82ca_ebbf3c474e44.slice/crio-d6941dd2a4ccf9d1fbc6860ad65c612465afa15ad62dbb39e60ca4ce2d589d5d WatchSource:0}: Error finding container d6941dd2a4ccf9d1fbc6860ad65c612465afa15ad62dbb39e60ca4ce2d589d5d: Status 404 returned error can't find the container with id d6941dd2a4ccf9d1fbc6860ad65c612465afa15ad62dbb39e60ca4ce2d589d5d Oct 06 10:12:54 crc kubenswrapper[4824]: I1006 10:12:54.123724 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-7468f855d8-kpbcv"] Oct 06 10:12:54 crc kubenswrapper[4824]: I1006 10:12:54.484854 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-84c678c567-wzzrh"] Oct 06 10:12:54 crc kubenswrapper[4824]: I1006 10:12:54.506233 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-65d89cfd9f-gts97"] Oct 06 10:12:54 crc kubenswrapper[4824]: I1006 10:12:54.519012 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-6859f9b676-n84pw"] Oct 06 10:12:54 crc kubenswrapper[4824]: I1006 10:12:54.521680 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-6cbc6dd547-dglbk"] Oct 06 10:12:54 crc kubenswrapper[4824]: I1006 10:12:54.539441 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-c8wx2"] Oct 06 10:12:54 crc kubenswrapper[4824]: W1006 10:12:54.544290 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod21123da0_cde6_45dd_9f50_f3e44a1f78c2.slice/crio-8e94391ca0b11e34c8ac458fb62dabcd65d488e0c3fe3747de4215c66052a9bb WatchSource:0}: Error finding container 8e94391ca0b11e34c8ac458fb62dabcd65d488e0c3fe3747de4215c66052a9bb: Status 404 returned error can't find the container with id 8e94391ca0b11e34c8ac458fb62dabcd65d488e0c3fe3747de4215c66052a9bb Oct 06 10:12:54 crc kubenswrapper[4824]: W1006 10:12:54.547368 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb32a09cc_901f_41f2_9912_628bbc33da7b.slice/crio-fc7db450927a5b19beed78b1cd8190c6b78caf7e347f2fe74b10b12153376213 WatchSource:0}: Error finding container fc7db450927a5b19beed78b1cd8190c6b78caf7e347f2fe74b10b12153376213: Status 404 returned error can't find the container with id fc7db450927a5b19beed78b1cd8190c6b78caf7e347f2fe74b10b12153376213 Oct 06 10:12:54 crc kubenswrapper[4824]: I1006 10:12:54.549314 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-54689d9f88-8qfvp"] Oct 06 10:12:54 crc kubenswrapper[4824]: E1006 10:12:54.550778 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:637bb7b9ac308bc1e323391a3593b824f688090a856c83385814c17a571b1eed,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-kwk7h,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-6859f9b676-n84pw_openstack-operators(b32a09cc-901f-41f2-9912-628bbc33da7b): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 06 10:12:54 crc kubenswrapper[4824]: I1006 10:12:54.553848 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5cd5cb47d7-6qgln"] Oct 06 10:12:54 crc kubenswrapper[4824]: W1006 10:12:54.555484 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod755275e4_c198_49df_90b5_0688a4fb8228.slice/crio-c32600ba4c0557748cc1e564b3b5e13cbe3f3bf16cfa96e86bcb8c02ef4bef00 WatchSource:0}: Error finding container c32600ba4c0557748cc1e564b3b5e13cbe3f3bf16cfa96e86bcb8c02ef4bef00: Status 404 returned error can't find the container with id c32600ba4c0557748cc1e564b3b5e13cbe3f3bf16cfa96e86bcb8c02ef4bef00 Oct 06 10:12:54 crc kubenswrapper[4824]: I1006 10:12:54.558700 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665cxcwhp"] Oct 06 10:12:54 crc kubenswrapper[4824]: E1006 10:12:54.559585 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:0daf76cc40ab619ae266b11defcc1b65beb22d859369e7b1b04de9169089a4cb,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-kghzm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5cd5cb47d7-6qgln_openstack-operators(755275e4-c198-49df-90b5-0688a4fb8228): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 06 10:12:54 crc kubenswrapper[4824]: W1006 10:12:54.563442 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1778e737_b572_417b_931d_0eec04a89ac9.slice/crio-a813fdd438d104b359052352fe1b5147a1b266fc0d558e2119bce62c9154eb35 WatchSource:0}: Error finding container a813fdd438d104b359052352fe1b5147a1b266fc0d558e2119bce62c9154eb35: Status 404 returned error can't find the container with id a813fdd438d104b359052352fe1b5147a1b266fc0d558e2119bce62c9154eb35 Oct 06 10:12:54 crc kubenswrapper[4824]: E1006 10:12:54.572538 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:adc23c5fd1aece2b16dc8e22ceed628f9a719455e39d3f98c77544665c6749e1,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-ng82w,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-54689d9f88-8qfvp_openstack-operators(1778e737-b572-417b-931d-0eec04a89ac9): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 06 10:12:54 crc kubenswrapper[4824]: W1006 10:12:54.583289 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2a48640f_172c_4e24_8c75_5c36d26ae1aa.slice/crio-69282c12582597a84f2236fab5d20583d5970c3c3255391ca8faf7e1f8f35bce WatchSource:0}: Error finding container 69282c12582597a84f2236fab5d20583d5970c3c3255391ca8faf7e1f8f35bce: Status 404 returned error can't find the container with id 69282c12582597a84f2236fab5d20583d5970c3c3255391ca8faf7e1f8f35bce Oct 06 10:12:54 crc kubenswrapper[4824]: W1006 10:12:54.586102 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfa5da9bb_805d_420f_adc6_e948cf910b21.slice/crio-82008f8f3498c234263f989b854988f19cbf34e9aa6f675254fc8f30e44de7a6 WatchSource:0}: Error finding container 82008f8f3498c234263f989b854988f19cbf34e9aa6f675254fc8f30e44de7a6: Status 404 returned error can't find the container with id 82008f8f3498c234263f989b854988f19cbf34e9aa6f675254fc8f30e44de7a6 Oct 06 10:12:54 crc kubenswrapper[4824]: W1006 10:12:54.586576 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1f00bc57_93de_431a_adcf_d1e1221121a1.slice/crio-f357e2df9bef4dfae882833ded008026e97b2b42a74cf6fcdc9d5864ecb4b02c WatchSource:0}: Error finding container f357e2df9bef4dfae882833ded008026e97b2b42a74cf6fcdc9d5864ecb4b02c: Status 404 returned error can't find the container with id f357e2df9bef4dfae882833ded008026e97b2b42a74cf6fcdc9d5864ecb4b02c Oct 06 10:12:54 crc kubenswrapper[4824]: I1006 10:12:54.603855 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-844bbc956-tjxqg"] Oct 06 10:12:54 crc kubenswrapper[4824]: E1006 10:12:54.607938 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:bcd1acac74e68eea5a9c3b7ba1bcb29d3a5b43423fc23c19ad4715bdac41f799,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:true,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-baremetal-operator-agent:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_ANSIBLEEE_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-ansibleee-runner:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_EVALUATOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-evaluator:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_LISTENER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-listener:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_NOTIFIER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-notifier:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_APACHE_IMAGE_URL_DEFAULT,Value:registry.redhat.io/ubi9/httpd-24:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_KEYSTONE_LISTENER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-keystone-listener:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-worker:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_CENTRAL_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_COMPUTE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_IPMI_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-ipmi:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_MYSQLD_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/prometheus/mysqld-exporter:v0.15.1,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_NOTIFICATION_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-notification:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_SGCORE_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/sg-core:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_BACKUP_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-backup:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-scheduler:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_VOLUME_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-volume:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_BACKENDBIND9_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-backend-bind9:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_CENTRAL_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-central:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_MDNS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-mdns:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_PRODUCER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-producer:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_UNBOUND_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-unbound:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-worker:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_FRR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-frr:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_ISCSID_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-iscsid:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_KEPLER_IMAGE_URL_DEFAULT,Value:quay.io/sustainable_computing_io/kepler:release-0.7.12,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_LOGROTATE_CROND_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cron:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_MULTIPATHD_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-multipathd:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_DHCP_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_METADATA_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_OVN_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-ovn-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_SRIOV_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-sriov-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NODE_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/prometheus/node-exporter:v1.5.0,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_OVN_BGP_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-bgp-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_PODMAN_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/navidys/prometheus-podman-exporter:v1.10.1,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_GLANCE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-glance-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_CFNAPI_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-api-cfn:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_ENGINE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-engine:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HORIZON_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_INFRA_MEMCACHED_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-memcached:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_INFRA_REDIS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-redis:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_CONDUCTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-conductor:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_INSPECTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-inspector:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_NEUTRON_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-neutron-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_PXE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-pxe:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_PYTHON_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/ironic-python-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_KEYSTONE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-keystone:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_KSM_IMAGE_URL_DEFAULT,Value:registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.15.0,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-scheduler:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_SHARE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-share:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MARIADB_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-mariadb:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NET_UTILS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-netutils:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NEUTRON_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_COMPUTE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-compute:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_CONDUCTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-conductor:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_NOVNC_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-novncproxy:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-scheduler:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_HEALTHMANAGER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-health-manager:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_HOUSEKEEPING_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-housekeeping:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_RSYSLOG_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-rsyslog:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-worker:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_CLIENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-openstackclient:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_MUST_GATHER_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-must-gather:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_NETWORK_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-network-exporter:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OS_CONTAINER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/edpm-hardened-uefi:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_CONTROLLER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_CONTROLLER_OVS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-base:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_NB_DBCLUSTER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-nb-db-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_NORTHD_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-northd:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_SB_DBCLUSTER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-sb-db-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_PLACEMENT_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-placement-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_RABBITMQ_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_ACCOUNT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-account:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_CONTAINER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-container:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_OBJECT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-object:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_PROXY_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-proxy-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_TEST_TEMPEST_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_APPLIER_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-applier:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_DECISION_ENGINE_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-decision-engine:current-podified,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cert,ReadOnly:true,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-dn8qm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-baremetal-operator-controller-manager-5dfbbd665cxcwhp_openstack-operators(fa5da9bb-805d-420f-adc6-e948cf910b21): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 06 10:12:54 crc kubenswrapper[4824]: E1006 10:12:54.608158 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-68grf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-5f97d8c699-c8wx2_openstack-operators(1f00bc57-93de-431a-adcf-d1e1221121a1): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 06 10:12:54 crc kubenswrapper[4824]: E1006 10:12:54.608164 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:64f57b2b59dea2bd9fae91490c5bec2687131884a049e6579819d9f951b877c6,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-6dmrw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-6cbc6dd547-dglbk_openstack-operators(2a48640f-172c-4e24-8c75-5c36d26ae1aa): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 06 10:12:54 crc kubenswrapper[4824]: E1006 10:12:54.610001 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-c8wx2" podUID="1f00bc57-93de-431a-adcf-d1e1221121a1" Oct 06 10:12:54 crc kubenswrapper[4824]: W1006 10:12:54.654428 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod993a8c7b_2aac_4de6_934d_b591cb6bfffe.slice/crio-d07b6d1df01b72e15655bf13ff025c5c6fda1aa31afc80107813c1c54c20d5b7 WatchSource:0}: Error finding container d07b6d1df01b72e15655bf13ff025c5c6fda1aa31afc80107813c1c54c20d5b7: Status 404 returned error can't find the container with id d07b6d1df01b72e15655bf13ff025c5c6fda1aa31afc80107813c1c54c20d5b7 Oct 06 10:12:54 crc kubenswrapper[4824]: E1006 10:12:54.909713 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-6qgln" podUID="755275e4-c198-49df-90b5-0688a4fb8228" Oct 06 10:12:54 crc kubenswrapper[4824]: E1006 10:12:54.938025 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/placement-operator-controller-manager-54689d9f88-8qfvp" podUID="1778e737-b572-417b-931d-0eec04a89ac9" Oct 06 10:12:54 crc kubenswrapper[4824]: E1006 10:12:54.951139 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/swift-operator-controller-manager-6859f9b676-n84pw" podUID="b32a09cc-901f-41f2-9912-628bbc33da7b" Oct 06 10:12:54 crc kubenswrapper[4824]: E1006 10:12:54.954828 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/watcher-operator-controller-manager-6cbc6dd547-dglbk" podUID="2a48640f-172c-4e24-8c75-5c36d26ae1aa" Oct 06 10:12:54 crc kubenswrapper[4824]: E1006 10:12:54.972081 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665cxcwhp" podUID="fa5da9bb-805d-420f-adc6-e948cf910b21" Oct 06 10:12:54 crc kubenswrapper[4824]: I1006 10:12:54.993334 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-7468f855d8-kpbcv" event={"ID":"c0da600c-cfd5-405a-82ca-ebbf3c474e44","Type":"ContainerStarted","Data":"d6941dd2a4ccf9d1fbc6860ad65c612465afa15ad62dbb39e60ca4ce2d589d5d"} Oct 06 10:12:55 crc kubenswrapper[4824]: I1006 10:12:55.002555 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-6cd6d7bdf5-4zc5b" event={"ID":"84a43a6c-6456-47ed-ae93-e4476d3cc978","Type":"ContainerStarted","Data":"50d8b30830a0fbf685b243d3b0d57f4088de6f7be13aa7af3738469f45eabb72"} Oct 06 10:12:55 crc kubenswrapper[4824]: I1006 10:12:55.004467 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-579449c7d5-w7k8s" event={"ID":"12802c32-cb1d-452b-8240-9a5e1e40a163","Type":"ContainerStarted","Data":"ffb350bb1b6fd11865269b42e874f7ef53a098f47d5665471627b99b612b7669"} Oct 06 10:12:55 crc kubenswrapper[4824]: I1006 10:12:55.007392 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-54876c876f-swsgf" event={"ID":"67824860-53b9-4c9c-95a1-955d9139f6e8","Type":"ContainerStarted","Data":"74f72ee4c4b5cf93271f117a499de8de6081426d690fc2a28d6e2ba5a77aa155"} Oct 06 10:12:55 crc kubenswrapper[4824]: I1006 10:12:55.011301 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-84c678c567-wzzrh" event={"ID":"7c4f21a3-07f8-46d8-bd7b-e2a24bf280d8","Type":"ContainerStarted","Data":"9252e19da24f1e3763b8d9fa01ed8e9709952091c67bf7927a7ee6626d33b4dd"} Oct 06 10:12:55 crc kubenswrapper[4824]: I1006 10:12:55.012615 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-699b87f775-8mvml" event={"ID":"09c09b76-0deb-44f1-bc93-b4479b7baa36","Type":"ContainerStarted","Data":"100db90c33ac6555165b2681373418ad2acdd3c2816b5bab600e9589425eac3e"} Oct 06 10:12:55 crc kubenswrapper[4824]: I1006 10:12:55.016971 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-6qgln" event={"ID":"755275e4-c198-49df-90b5-0688a4fb8228","Type":"ContainerStarted","Data":"1ed89bd19eb066e4ada630c0b109e9bcc22aa68bdb1e74acb5864d2b9c9b1098"} Oct 06 10:12:55 crc kubenswrapper[4824]: I1006 10:12:55.017012 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-6qgln" event={"ID":"755275e4-c198-49df-90b5-0688a4fb8228","Type":"ContainerStarted","Data":"c32600ba4c0557748cc1e564b3b5e13cbe3f3bf16cfa96e86bcb8c02ef4bef00"} Oct 06 10:12:55 crc kubenswrapper[4824]: E1006 10:12:55.019562 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:0daf76cc40ab619ae266b11defcc1b65beb22d859369e7b1b04de9169089a4cb\\\"\"" pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-6qgln" podUID="755275e4-c198-49df-90b5-0688a4fb8228" Oct 06 10:12:55 crc kubenswrapper[4824]: I1006 10:12:55.023150 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-844bbc956-tjxqg" event={"ID":"993a8c7b-2aac-4de6-934d-b591cb6bfffe","Type":"ContainerStarted","Data":"5104c62f3235b73f263b88975d25dd87aac24fd8cf13f90417cea8da3483f104"} Oct 06 10:12:55 crc kubenswrapper[4824]: I1006 10:12:55.023179 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-844bbc956-tjxqg" event={"ID":"993a8c7b-2aac-4de6-934d-b591cb6bfffe","Type":"ContainerStarted","Data":"d07b6d1df01b72e15655bf13ff025c5c6fda1aa31afc80107813c1c54c20d5b7"} Oct 06 10:12:55 crc kubenswrapper[4824]: I1006 10:12:55.027443 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-6cbc6dd547-dglbk" event={"ID":"2a48640f-172c-4e24-8c75-5c36d26ae1aa","Type":"ContainerStarted","Data":"80ef4266b9599099b978443c5048986c4aa9e8413ae2fd7864615c8c51f1f0f4"} Oct 06 10:12:55 crc kubenswrapper[4824]: I1006 10:12:55.027473 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-6cbc6dd547-dglbk" event={"ID":"2a48640f-172c-4e24-8c75-5c36d26ae1aa","Type":"ContainerStarted","Data":"69282c12582597a84f2236fab5d20583d5970c3c3255391ca8faf7e1f8f35bce"} Oct 06 10:12:55 crc kubenswrapper[4824]: E1006 10:12:55.030259 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:64f57b2b59dea2bd9fae91490c5bec2687131884a049e6579819d9f951b877c6\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-6cbc6dd547-dglbk" podUID="2a48640f-172c-4e24-8c75-5c36d26ae1aa" Oct 06 10:12:55 crc kubenswrapper[4824]: I1006 10:12:55.035766 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-54689d9f88-8qfvp" event={"ID":"1778e737-b572-417b-931d-0eec04a89ac9","Type":"ContainerStarted","Data":"00df1f8f9623759b3294d6a28d7d9526db6e4e06bd816eab84e05165e35c34a0"} Oct 06 10:12:55 crc kubenswrapper[4824]: I1006 10:12:55.035829 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-54689d9f88-8qfvp" event={"ID":"1778e737-b572-417b-931d-0eec04a89ac9","Type":"ContainerStarted","Data":"a813fdd438d104b359052352fe1b5147a1b266fc0d558e2119bce62c9154eb35"} Oct 06 10:12:55 crc kubenswrapper[4824]: E1006 10:12:55.045908 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:adc23c5fd1aece2b16dc8e22ceed628f9a719455e39d3f98c77544665c6749e1\\\"\"" pod="openstack-operators/placement-operator-controller-manager-54689d9f88-8qfvp" podUID="1778e737-b572-417b-931d-0eec04a89ac9" Oct 06 10:12:55 crc kubenswrapper[4824]: I1006 10:12:55.047930 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-7c7fc454ff-fjjsf" event={"ID":"dff987b5-f7ce-4eb5-9778-3cb34663824a","Type":"ContainerStarted","Data":"a38bc20795366e17a56115ab20810e98dd20b4ed0603ecf879ebc4e4601b0290"} Oct 06 10:12:55 crc kubenswrapper[4824]: I1006 10:12:55.057867 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-655d88ccb9-t8sjr" event={"ID":"dd382562-c508-4b17-ae00-5db8abe20596","Type":"ContainerStarted","Data":"10eb2b148f3165418e90a8485b8002dfe243fd8f8eb5fd4cb44a5cbf54ea2331"} Oct 06 10:12:55 crc kubenswrapper[4824]: I1006 10:12:55.075295 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-c8wx2" event={"ID":"1f00bc57-93de-431a-adcf-d1e1221121a1","Type":"ContainerStarted","Data":"f357e2df9bef4dfae882833ded008026e97b2b42a74cf6fcdc9d5864ecb4b02c"} Oct 06 10:12:55 crc kubenswrapper[4824]: E1006 10:12:55.078322 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-c8wx2" podUID="1f00bc57-93de-431a-adcf-d1e1221121a1" Oct 06 10:12:55 crc kubenswrapper[4824]: I1006 10:12:55.087634 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-6859f9b676-n84pw" event={"ID":"b32a09cc-901f-41f2-9912-628bbc33da7b","Type":"ContainerStarted","Data":"43e69d17976894a8fe5105882b47c94b33f8048986ffdef6c4de91b4d38f75cd"} Oct 06 10:12:55 crc kubenswrapper[4824]: I1006 10:12:55.087707 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-6859f9b676-n84pw" event={"ID":"b32a09cc-901f-41f2-9912-628bbc33da7b","Type":"ContainerStarted","Data":"fc7db450927a5b19beed78b1cd8190c6b78caf7e347f2fe74b10b12153376213"} Oct 06 10:12:55 crc kubenswrapper[4824]: E1006 10:12:55.090018 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:637bb7b9ac308bc1e323391a3593b824f688090a856c83385814c17a571b1eed\\\"\"" pod="openstack-operators/swift-operator-controller-manager-6859f9b676-n84pw" podUID="b32a09cc-901f-41f2-9912-628bbc33da7b" Oct 06 10:12:55 crc kubenswrapper[4824]: I1006 10:12:55.091343 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-658588b8c9-6884h" event={"ID":"98cddaff-3395-473c-afd4-02c849951d51","Type":"ContainerStarted","Data":"63ce7216e478407640c924f4ddb096e77e8cc7845467be69d46007d8ba44daa7"} Oct 06 10:12:55 crc kubenswrapper[4824]: I1006 10:12:55.092900 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665cxcwhp" event={"ID":"fa5da9bb-805d-420f-adc6-e948cf910b21","Type":"ContainerStarted","Data":"591cebb83393723b324479d08bddf2038e1b0d43a0c0d7403c6e0be821e181e8"} Oct 06 10:12:55 crc kubenswrapper[4824]: I1006 10:12:55.092950 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665cxcwhp" event={"ID":"fa5da9bb-805d-420f-adc6-e948cf910b21","Type":"ContainerStarted","Data":"82008f8f3498c234263f989b854988f19cbf34e9aa6f675254fc8f30e44de7a6"} Oct 06 10:12:55 crc kubenswrapper[4824]: E1006 10:12:55.094824 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:bcd1acac74e68eea5a9c3b7ba1bcb29d3a5b43423fc23c19ad4715bdac41f799\\\"\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665cxcwhp" podUID="fa5da9bb-805d-420f-adc6-e948cf910b21" Oct 06 10:12:55 crc kubenswrapper[4824]: I1006 10:12:55.102736 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-8d984cc4d-82bf8" event={"ID":"36dbefcc-0dac-4435-aaac-5248cd2eb209","Type":"ContainerStarted","Data":"701c48f548fb40854e214e14e9d5d04d03b90ab39fbbf8a2d360fe24fde1e755"} Oct 06 10:12:55 crc kubenswrapper[4824]: I1006 10:12:55.127601 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-65d89cfd9f-gts97" event={"ID":"21123da0-cde6-45dd-9f50-f3e44a1f78c2","Type":"ContainerStarted","Data":"8e94391ca0b11e34c8ac458fb62dabcd65d488e0c3fe3747de4215c66052a9bb"} Oct 06 10:12:56 crc kubenswrapper[4824]: I1006 10:12:56.144436 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-844bbc956-tjxqg" event={"ID":"993a8c7b-2aac-4de6-934d-b591cb6bfffe","Type":"ContainerStarted","Data":"a645deb34bb3dcff8ba5a2f8072a11e0d02c48b8dbf0aa72c0603fb6d7d7b36f"} Oct 06 10:12:56 crc kubenswrapper[4824]: E1006 10:12:56.146630 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:637bb7b9ac308bc1e323391a3593b824f688090a856c83385814c17a571b1eed\\\"\"" pod="openstack-operators/swift-operator-controller-manager-6859f9b676-n84pw" podUID="b32a09cc-901f-41f2-9912-628bbc33da7b" Oct 06 10:12:56 crc kubenswrapper[4824]: E1006 10:12:56.146892 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:64f57b2b59dea2bd9fae91490c5bec2687131884a049e6579819d9f951b877c6\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-6cbc6dd547-dglbk" podUID="2a48640f-172c-4e24-8c75-5c36d26ae1aa" Oct 06 10:12:56 crc kubenswrapper[4824]: E1006 10:12:56.146991 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-c8wx2" podUID="1f00bc57-93de-431a-adcf-d1e1221121a1" Oct 06 10:12:56 crc kubenswrapper[4824]: E1006 10:12:56.148099 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:0daf76cc40ab619ae266b11defcc1b65beb22d859369e7b1b04de9169089a4cb\\\"\"" pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-6qgln" podUID="755275e4-c198-49df-90b5-0688a4fb8228" Oct 06 10:12:56 crc kubenswrapper[4824]: E1006 10:12:56.148884 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:bcd1acac74e68eea5a9c3b7ba1bcb29d3a5b43423fc23c19ad4715bdac41f799\\\"\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665cxcwhp" podUID="fa5da9bb-805d-420f-adc6-e948cf910b21" Oct 06 10:12:56 crc kubenswrapper[4824]: E1006 10:12:56.148921 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:adc23c5fd1aece2b16dc8e22ceed628f9a719455e39d3f98c77544665c6749e1\\\"\"" pod="openstack-operators/placement-operator-controller-manager-54689d9f88-8qfvp" podUID="1778e737-b572-417b-931d-0eec04a89ac9" Oct 06 10:12:56 crc kubenswrapper[4824]: I1006 10:12:56.289152 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-844bbc956-tjxqg" podStartSLOduration=4.289123867 podStartE2EDuration="4.289123867s" podCreationTimestamp="2025-10-06 10:12:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:12:56.285133912 +0000 UTC m=+885.649556783" watchObservedRunningTime="2025-10-06 10:12:56.289123867 +0000 UTC m=+885.653546728" Oct 06 10:12:57 crc kubenswrapper[4824]: I1006 10:12:57.152802 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-844bbc956-tjxqg" Oct 06 10:13:04 crc kubenswrapper[4824]: I1006 10:13:04.088636 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-844bbc956-tjxqg" Oct 06 10:13:06 crc kubenswrapper[4824]: E1006 10:13:06.163131 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.174:5001/openstack-k8s-operators/telemetry-operator:62951b60f91bd212bcc73b07adf65219e4bc1cf4" Oct 06 10:13:06 crc kubenswrapper[4824]: E1006 10:13:06.163541 4824 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.174:5001/openstack-k8s-operators/telemetry-operator:62951b60f91bd212bcc73b07adf65219e4bc1cf4" Oct 06 10:13:06 crc kubenswrapper[4824]: E1006 10:13:06.163781 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:38.102.83.174:5001/openstack-k8s-operators/telemetry-operator:62951b60f91bd212bcc73b07adf65219e4bc1cf4,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-fn9xt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-84c678c567-wzzrh_openstack-operators(7c4f21a3-07f8-46d8-bd7b-e2a24bf280d8): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 06 10:13:06 crc kubenswrapper[4824]: E1006 10:13:06.972251 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/telemetry-operator-controller-manager-84c678c567-wzzrh" podUID="7c4f21a3-07f8-46d8-bd7b-e2a24bf280d8" Oct 06 10:13:07 crc kubenswrapper[4824]: I1006 10:13:07.267388 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-8f58bc9db-tp4b2" event={"ID":"76419802-f7a7-4270-9460-3cc1a9f7f667","Type":"ContainerStarted","Data":"e1ba76c9dc6c6eb0746008741126dba201069d8e13df797e5830beccc822887c"} Oct 06 10:13:07 crc kubenswrapper[4824]: I1006 10:13:07.314720 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-7d4d4f8d-4kn7z" event={"ID":"b3c0daed-de3b-415e-9166-756cb0f5cab7","Type":"ContainerStarted","Data":"0eebffbad912852b008549aecc5690932e4b4c7ded30e7d688c2ca35eb20da88"} Oct 06 10:13:07 crc kubenswrapper[4824]: I1006 10:13:07.314764 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-7d4d4f8d-4kn7z" event={"ID":"b3c0daed-de3b-415e-9166-756cb0f5cab7","Type":"ContainerStarted","Data":"a737ab66a6c76b4c4aa917b6f5ea14eb1ef46911883947c9c96572d6d4d1a59d"} Oct 06 10:13:07 crc kubenswrapper[4824]: I1006 10:13:07.314828 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-7d4d4f8d-4kn7z" Oct 06 10:13:07 crc kubenswrapper[4824]: I1006 10:13:07.338190 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-75dfd9b554-hktsj" event={"ID":"4ca61d7c-741e-4028-8a75-d26a7b4cbd3a","Type":"ContainerStarted","Data":"6054f67a1129167d48a7c39d2960d9158157e55ddec39cc4755d6e43f2d5a6ab"} Oct 06 10:13:07 crc kubenswrapper[4824]: I1006 10:13:07.362785 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-7c7fc454ff-fjjsf" event={"ID":"dff987b5-f7ce-4eb5-9778-3cb34663824a","Type":"ContainerStarted","Data":"722b91883548f42c1bb9f1a3a9e5d1fcd375670f3035af4f8879085cf13ba88b"} Oct 06 10:13:07 crc kubenswrapper[4824]: I1006 10:13:07.388284 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-655d88ccb9-t8sjr" event={"ID":"dd382562-c508-4b17-ae00-5db8abe20596","Type":"ContainerStarted","Data":"b9b4317f81e69840638eb96d787bac24d896d50dba39cf91b1983aa69ac2f22d"} Oct 06 10:13:07 crc kubenswrapper[4824]: I1006 10:13:07.396568 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-7d4d4f8d-4kn7z" podStartSLOduration=2.466836958 podStartE2EDuration="15.396547316s" podCreationTimestamp="2025-10-06 10:12:52 +0000 UTC" firstStartedPulling="2025-10-06 10:12:53.247761802 +0000 UTC m=+882.612184663" lastFinishedPulling="2025-10-06 10:13:06.17747212 +0000 UTC m=+895.541895021" observedRunningTime="2025-10-06 10:13:07.372760872 +0000 UTC m=+896.737183733" watchObservedRunningTime="2025-10-06 10:13:07.396547316 +0000 UTC m=+896.760970177" Oct 06 10:13:07 crc kubenswrapper[4824]: I1006 10:13:07.455433 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-84c678c567-wzzrh" event={"ID":"7c4f21a3-07f8-46d8-bd7b-e2a24bf280d8","Type":"ContainerStarted","Data":"875365dc819ce04fd051feaed45b30cd3e430c93e5b04d537a6d6b18337fee15"} Oct 06 10:13:07 crc kubenswrapper[4824]: E1006 10:13:07.466231 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.174:5001/openstack-k8s-operators/telemetry-operator:62951b60f91bd212bcc73b07adf65219e4bc1cf4\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-84c678c567-wzzrh" podUID="7c4f21a3-07f8-46d8-bd7b-e2a24bf280d8" Oct 06 10:13:07 crc kubenswrapper[4824]: I1006 10:13:07.467317 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-5568b5d68-r996x" event={"ID":"e8ad16ea-027c-4ca7-accd-928ba6faf830","Type":"ContainerStarted","Data":"058af4a3d37f6da4f004dca6a7efc991dfa389f76ae3499b1ad6c00c725199d0"} Oct 06 10:13:07 crc kubenswrapper[4824]: I1006 10:13:07.468820 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-699b87f775-8mvml" event={"ID":"09c09b76-0deb-44f1-bc93-b4479b7baa36","Type":"ContainerStarted","Data":"2c8239cc6c8d29856b79c605de2465cf83fec3bc6ee0f66bb1ddc41fa48dd34a"} Oct 06 10:13:07 crc kubenswrapper[4824]: I1006 10:13:07.469928 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-8d984cc4d-82bf8" event={"ID":"36dbefcc-0dac-4435-aaac-5248cd2eb209","Type":"ContainerStarted","Data":"f50a97040acb8c561d839c71e3d0e3f7e3d7db2dc1e0bf93192573675138baea"} Oct 06 10:13:07 crc kubenswrapper[4824]: I1006 10:13:07.470753 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-5f7c849b98-fn8nc" event={"ID":"e60e7a0d-ccc6-4b1a-b645-a4802b21d48e","Type":"ContainerStarted","Data":"17eb026ad26105731a22709a71246d42e79eabee43d8991dfb4f979939339d7a"} Oct 06 10:13:07 crc kubenswrapper[4824]: I1006 10:13:07.473754 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-658588b8c9-6884h" event={"ID":"98cddaff-3395-473c-afd4-02c849951d51","Type":"ContainerStarted","Data":"fed94839d6f5a8961d4afc5317a515d574bd35836343f794f6c649133782a8fd"} Oct 06 10:13:07 crc kubenswrapper[4824]: I1006 10:13:07.485420 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-6cd6d7bdf5-4zc5b" event={"ID":"84a43a6c-6456-47ed-ae93-e4476d3cc978","Type":"ContainerStarted","Data":"233df3ccc3a4b5cb9e960b83579a27f9e960476696689e5c9ae83ce807d1dc14"} Oct 06 10:13:07 crc kubenswrapper[4824]: I1006 10:13:07.486378 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-7468f855d8-kpbcv" event={"ID":"c0da600c-cfd5-405a-82ca-ebbf3c474e44","Type":"ContainerStarted","Data":"7f07540ff3c22f1763e03103f61bc593c47ee6b39a0d151931aa47a030422b52"} Oct 06 10:13:07 crc kubenswrapper[4824]: I1006 10:13:07.493492 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-65d89cfd9f-gts97" event={"ID":"21123da0-cde6-45dd-9f50-f3e44a1f78c2","Type":"ContainerStarted","Data":"a25f29e509275f4b119ef9c946ff9dc941dab08dabf6c7728b3c367aaa275c0c"} Oct 06 10:13:07 crc kubenswrapper[4824]: I1006 10:13:07.508477 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-579449c7d5-w7k8s" event={"ID":"12802c32-cb1d-452b-8240-9a5e1e40a163","Type":"ContainerStarted","Data":"16d156f55d710599da6f7fd5aac012d46bce434a3ce05cadecbd9bafab5a559a"} Oct 06 10:13:07 crc kubenswrapper[4824]: I1006 10:13:07.537877 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-54876c876f-swsgf" event={"ID":"67824860-53b9-4c9c-95a1-955d9139f6e8","Type":"ContainerStarted","Data":"82ea7325a4ed0574ca6194bdc8bfaee25f08c5cf0ebc1942b21873b5ea3282a9"} Oct 06 10:13:07 crc kubenswrapper[4824]: I1006 10:13:07.538639 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-54876c876f-swsgf" Oct 06 10:13:07 crc kubenswrapper[4824]: I1006 10:13:07.622076 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-54876c876f-swsgf" podStartSLOduration=3.486629716 podStartE2EDuration="15.622041969s" podCreationTimestamp="2025-10-06 10:12:52 +0000 UTC" firstStartedPulling="2025-10-06 10:12:54.044168302 +0000 UTC m=+883.408591163" lastFinishedPulling="2025-10-06 10:13:06.179580515 +0000 UTC m=+895.544003416" observedRunningTime="2025-10-06 10:13:07.608635324 +0000 UTC m=+896.973058185" watchObservedRunningTime="2025-10-06 10:13:07.622041969 +0000 UTC m=+896.986464830" Oct 06 10:13:08 crc kubenswrapper[4824]: I1006 10:13:08.550273 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-699b87f775-8mvml" event={"ID":"09c09b76-0deb-44f1-bc93-b4479b7baa36","Type":"ContainerStarted","Data":"e32786d564a7c31e1500175d2cf9fffd72d9a4a0d876cd25ca6cdb5ae9e8a882"} Oct 06 10:13:08 crc kubenswrapper[4824]: I1006 10:13:08.552884 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-7c7fc454ff-fjjsf" event={"ID":"dff987b5-f7ce-4eb5-9778-3cb34663824a","Type":"ContainerStarted","Data":"41ed60464759321b855e8363fb854548c238eccb13f0a4c5268d4c89e5227d72"} Oct 06 10:13:08 crc kubenswrapper[4824]: I1006 10:13:08.553867 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-7c7fc454ff-fjjsf" Oct 06 10:13:08 crc kubenswrapper[4824]: I1006 10:13:08.556094 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-655d88ccb9-t8sjr" event={"ID":"dd382562-c508-4b17-ae00-5db8abe20596","Type":"ContainerStarted","Data":"d0e76baa675c49331d6f09405d4df09e407ac3fe15820116a6acf017432de39c"} Oct 06 10:13:08 crc kubenswrapper[4824]: I1006 10:13:08.556500 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-655d88ccb9-t8sjr" Oct 06 10:13:08 crc kubenswrapper[4824]: I1006 10:13:08.559336 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-8d984cc4d-82bf8" event={"ID":"36dbefcc-0dac-4435-aaac-5248cd2eb209","Type":"ContainerStarted","Data":"a4d31b11288c6899df8300408718842e23cac2d8b8300b91a6eebe547a880950"} Oct 06 10:13:08 crc kubenswrapper[4824]: I1006 10:13:08.561246 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-658588b8c9-6884h" event={"ID":"98cddaff-3395-473c-afd4-02c849951d51","Type":"ContainerStarted","Data":"1c332d838dc0b41afd2e7a1877ab00fe600d1756acbf49ede7094e012f1a13cc"} Oct 06 10:13:08 crc kubenswrapper[4824]: I1006 10:13:08.561577 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-658588b8c9-6884h" Oct 06 10:13:08 crc kubenswrapper[4824]: I1006 10:13:08.563735 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-579449c7d5-w7k8s" event={"ID":"12802c32-cb1d-452b-8240-9a5e1e40a163","Type":"ContainerStarted","Data":"e36c431a2d8c0f2464481af54749748258af007492f2299d910801b7f1abf35d"} Oct 06 10:13:08 crc kubenswrapper[4824]: I1006 10:13:08.564095 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-579449c7d5-w7k8s" Oct 06 10:13:08 crc kubenswrapper[4824]: I1006 10:13:08.565382 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-6cd6d7bdf5-4zc5b" event={"ID":"84a43a6c-6456-47ed-ae93-e4476d3cc978","Type":"ContainerStarted","Data":"d47c4f95147ed77d26268f094fde0b62f2877599ea1b6f50d6cf5895a82e1b34"} Oct 06 10:13:08 crc kubenswrapper[4824]: I1006 10:13:08.565733 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-6cd6d7bdf5-4zc5b" Oct 06 10:13:08 crc kubenswrapper[4824]: I1006 10:13:08.569174 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-54876c876f-swsgf" event={"ID":"67824860-53b9-4c9c-95a1-955d9139f6e8","Type":"ContainerStarted","Data":"826eb50d489a303c6a5270808770046e69e5f5c53618bbbb4853c0d15dd7d1da"} Oct 06 10:13:08 crc kubenswrapper[4824]: I1006 10:13:08.573609 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-5568b5d68-r996x" event={"ID":"e8ad16ea-027c-4ca7-accd-928ba6faf830","Type":"ContainerStarted","Data":"34999216680bceaf93d0c5b97c87f4e11bf791441fb0c3460923e1e936e08e5e"} Oct 06 10:13:08 crc kubenswrapper[4824]: I1006 10:13:08.573655 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-5568b5d68-r996x" Oct 06 10:13:08 crc kubenswrapper[4824]: E1006 10:13:08.575072 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.174:5001/openstack-k8s-operators/telemetry-operator:62951b60f91bd212bcc73b07adf65219e4bc1cf4\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-84c678c567-wzzrh" podUID="7c4f21a3-07f8-46d8-bd7b-e2a24bf280d8" Oct 06 10:13:08 crc kubenswrapper[4824]: I1006 10:13:08.583173 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-7c7fc454ff-fjjsf" podStartSLOduration=4.445555004 podStartE2EDuration="16.583152773s" podCreationTimestamp="2025-10-06 10:12:52 +0000 UTC" firstStartedPulling="2025-10-06 10:12:54.10815137 +0000 UTC m=+883.472574231" lastFinishedPulling="2025-10-06 10:13:06.245749099 +0000 UTC m=+895.610172000" observedRunningTime="2025-10-06 10:13:08.582447618 +0000 UTC m=+897.946870509" watchObservedRunningTime="2025-10-06 10:13:08.583152773 +0000 UTC m=+897.947575634" Oct 06 10:13:08 crc kubenswrapper[4824]: I1006 10:13:08.632465 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-658588b8c9-6884h" podStartSLOduration=4.440007977 podStartE2EDuration="16.632446039s" podCreationTimestamp="2025-10-06 10:12:52 +0000 UTC" firstStartedPulling="2025-10-06 10:12:54.071251607 +0000 UTC m=+883.435674468" lastFinishedPulling="2025-10-06 10:13:06.263689639 +0000 UTC m=+895.628112530" observedRunningTime="2025-10-06 10:13:08.627194747 +0000 UTC m=+897.991617618" watchObservedRunningTime="2025-10-06 10:13:08.632446039 +0000 UTC m=+897.996868900" Oct 06 10:13:08 crc kubenswrapper[4824]: I1006 10:13:08.669345 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-5568b5d68-r996x" podStartSLOduration=4.06198311 podStartE2EDuration="16.669327041s" podCreationTimestamp="2025-10-06 10:12:52 +0000 UTC" firstStartedPulling="2025-10-06 10:12:53.570024417 +0000 UTC m=+882.934447278" lastFinishedPulling="2025-10-06 10:13:06.177368328 +0000 UTC m=+895.541791209" observedRunningTime="2025-10-06 10:13:08.66411274 +0000 UTC m=+898.028535621" watchObservedRunningTime="2025-10-06 10:13:08.669327041 +0000 UTC m=+898.033749902" Oct 06 10:13:08 crc kubenswrapper[4824]: I1006 10:13:08.690129 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-6cd6d7bdf5-4zc5b" podStartSLOduration=4.567382159 podStartE2EDuration="16.690109612s" podCreationTimestamp="2025-10-06 10:12:52 +0000 UTC" firstStartedPulling="2025-10-06 10:12:54.111806008 +0000 UTC m=+883.476228859" lastFinishedPulling="2025-10-06 10:13:06.234533451 +0000 UTC m=+895.598956312" observedRunningTime="2025-10-06 10:13:08.684974052 +0000 UTC m=+898.049396923" watchObservedRunningTime="2025-10-06 10:13:08.690109612 +0000 UTC m=+898.054532463" Oct 06 10:13:08 crc kubenswrapper[4824]: I1006 10:13:08.708761 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-579449c7d5-w7k8s" podStartSLOduration=4.57779869 podStartE2EDuration="16.708739837s" podCreationTimestamp="2025-10-06 10:12:52 +0000 UTC" firstStartedPulling="2025-10-06 10:12:54.104014882 +0000 UTC m=+883.468437743" lastFinishedPulling="2025-10-06 10:13:06.234956029 +0000 UTC m=+895.599378890" observedRunningTime="2025-10-06 10:13:08.707724975 +0000 UTC m=+898.072147836" watchObservedRunningTime="2025-10-06 10:13:08.708739837 +0000 UTC m=+898.073162698" Oct 06 10:13:08 crc kubenswrapper[4824]: I1006 10:13:08.723894 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-655d88ccb9-t8sjr" podStartSLOduration=4.581980408 podStartE2EDuration="16.723875338s" podCreationTimestamp="2025-10-06 10:12:52 +0000 UTC" firstStartedPulling="2025-10-06 10:12:54.034724033 +0000 UTC m=+883.399146894" lastFinishedPulling="2025-10-06 10:13:06.176618933 +0000 UTC m=+895.541041824" observedRunningTime="2025-10-06 10:13:08.721380645 +0000 UTC m=+898.085803506" watchObservedRunningTime="2025-10-06 10:13:08.723875338 +0000 UTC m=+898.088298199" Oct 06 10:13:12 crc kubenswrapper[4824]: I1006 10:13:12.519582 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-7d4d4f8d-4kn7z" Oct 06 10:13:12 crc kubenswrapper[4824]: I1006 10:13:12.593623 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-5568b5d68-r996x" Oct 06 10:13:12 crc kubenswrapper[4824]: I1006 10:13:12.618913 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-5f7c849b98-fn8nc" event={"ID":"e60e7a0d-ccc6-4b1a-b645-a4802b21d48e","Type":"ContainerStarted","Data":"beff45e1e37901fd16e69e1c0d311774dcdc32de72946eb855f220ff35df9b1c"} Oct 06 10:13:12 crc kubenswrapper[4824]: I1006 10:13:12.622223 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-65d89cfd9f-gts97" event={"ID":"21123da0-cde6-45dd-9f50-f3e44a1f78c2","Type":"ContainerStarted","Data":"61754e2b5ecf9e31cb5e3509ec97d2c2dde2f9968c1dd862e8e7bb9aa1c70bd9"} Oct 06 10:13:12 crc kubenswrapper[4824]: I1006 10:13:12.624315 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-8f58bc9db-tp4b2" event={"ID":"76419802-f7a7-4270-9460-3cc1a9f7f667","Type":"ContainerStarted","Data":"91dea04660e7c190ef9cfe5c60d4dba2f5b611b832c9bafbab60613b6eb0b4c0"} Oct 06 10:13:12 crc kubenswrapper[4824]: I1006 10:13:12.625491 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-7468f855d8-kpbcv" event={"ID":"c0da600c-cfd5-405a-82ca-ebbf3c474e44","Type":"ContainerStarted","Data":"ebae8c74b5abe18d7a0a890a35ca5fb9ce5e07708f4b4d5ab865e76a2dd4d13b"} Oct 06 10:13:12 crc kubenswrapper[4824]: I1006 10:13:12.625722 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-7468f855d8-kpbcv" Oct 06 10:13:12 crc kubenswrapper[4824]: I1006 10:13:12.629929 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-7468f855d8-kpbcv" Oct 06 10:13:12 crc kubenswrapper[4824]: I1006 10:13:12.641068 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-75dfd9b554-hktsj" event={"ID":"4ca61d7c-741e-4028-8a75-d26a7b4cbd3a","Type":"ContainerStarted","Data":"121298db79f7d7a35e7d7a1ddb65a8a1a539b6e1701f065793fd2a31290ac70c"} Oct 06 10:13:12 crc kubenswrapper[4824]: I1006 10:13:12.641405 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-8d984cc4d-82bf8" Oct 06 10:13:12 crc kubenswrapper[4824]: I1006 10:13:12.641460 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-699b87f775-8mvml" Oct 06 10:13:12 crc kubenswrapper[4824]: I1006 10:13:12.650242 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-7468f855d8-kpbcv" podStartSLOduration=8.540100256 podStartE2EDuration="20.650220592s" podCreationTimestamp="2025-10-06 10:12:52 +0000 UTC" firstStartedPulling="2025-10-06 10:12:54.124282262 +0000 UTC m=+883.488705143" lastFinishedPulling="2025-10-06 10:13:06.234402588 +0000 UTC m=+895.598825479" observedRunningTime="2025-10-06 10:13:12.646528844 +0000 UTC m=+902.010951705" watchObservedRunningTime="2025-10-06 10:13:12.650220592 +0000 UTC m=+902.014643453" Oct 06 10:13:12 crc kubenswrapper[4824]: I1006 10:13:12.651606 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-699b87f775-8mvml" Oct 06 10:13:12 crc kubenswrapper[4824]: I1006 10:13:12.653236 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-8d984cc4d-82bf8" Oct 06 10:13:12 crc kubenswrapper[4824]: I1006 10:13:12.663932 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-699b87f775-8mvml" podStartSLOduration=8.443759883 podStartE2EDuration="20.663914413s" podCreationTimestamp="2025-10-06 10:12:52 +0000 UTC" firstStartedPulling="2025-10-06 10:12:54.015331721 +0000 UTC m=+883.379754582" lastFinishedPulling="2025-10-06 10:13:06.235486221 +0000 UTC m=+895.599909112" observedRunningTime="2025-10-06 10:13:12.663386271 +0000 UTC m=+902.027809132" watchObservedRunningTime="2025-10-06 10:13:12.663914413 +0000 UTC m=+902.028337274" Oct 06 10:13:12 crc kubenswrapper[4824]: I1006 10:13:12.686392 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-658588b8c9-6884h" Oct 06 10:13:12 crc kubenswrapper[4824]: I1006 10:13:12.704053 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-54876c876f-swsgf" Oct 06 10:13:12 crc kubenswrapper[4824]: I1006 10:13:12.727400 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-8d984cc4d-82bf8" podStartSLOduration=8.657870784 podStartE2EDuration="20.727372808s" podCreationTimestamp="2025-10-06 10:12:52 +0000 UTC" firstStartedPulling="2025-10-06 10:12:54.108445966 +0000 UTC m=+883.472868827" lastFinishedPulling="2025-10-06 10:13:06.17794799 +0000 UTC m=+895.542370851" observedRunningTime="2025-10-06 10:13:12.703926701 +0000 UTC m=+902.068349562" watchObservedRunningTime="2025-10-06 10:13:12.727372808 +0000 UTC m=+902.091795669" Oct 06 10:13:12 crc kubenswrapper[4824]: I1006 10:13:12.799213 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-655d88ccb9-t8sjr" Oct 06 10:13:12 crc kubenswrapper[4824]: I1006 10:13:12.903420 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-6cd6d7bdf5-4zc5b" Oct 06 10:13:12 crc kubenswrapper[4824]: I1006 10:13:12.962895 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-7c7fc454ff-fjjsf" Oct 06 10:13:13 crc kubenswrapper[4824]: I1006 10:13:13.011506 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-579449c7d5-w7k8s" Oct 06 10:13:13 crc kubenswrapper[4824]: I1006 10:13:13.650072 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-75dfd9b554-hktsj" Oct 06 10:13:13 crc kubenswrapper[4824]: I1006 10:13:13.652689 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-65d89cfd9f-gts97" Oct 06 10:13:13 crc kubenswrapper[4824]: I1006 10:13:13.659660 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-65d89cfd9f-gts97" Oct 06 10:13:13 crc kubenswrapper[4824]: I1006 10:13:13.660078 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-75dfd9b554-hktsj" Oct 06 10:13:13 crc kubenswrapper[4824]: I1006 10:13:13.670799 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-75dfd9b554-hktsj" podStartSLOduration=9.105248103 podStartE2EDuration="21.670779698s" podCreationTimestamp="2025-10-06 10:12:52 +0000 UTC" firstStartedPulling="2025-10-06 10:12:53.584066154 +0000 UTC m=+882.948489015" lastFinishedPulling="2025-10-06 10:13:06.149597729 +0000 UTC m=+895.514020610" observedRunningTime="2025-10-06 10:13:13.666718021 +0000 UTC m=+903.031140892" watchObservedRunningTime="2025-10-06 10:13:13.670779698 +0000 UTC m=+903.035202559" Oct 06 10:13:13 crc kubenswrapper[4824]: I1006 10:13:13.688919 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-8f58bc9db-tp4b2" podStartSLOduration=9.093218698 podStartE2EDuration="21.688899212s" podCreationTimestamp="2025-10-06 10:12:52 +0000 UTC" firstStartedPulling="2025-10-06 10:12:53.584144076 +0000 UTC m=+882.948566937" lastFinishedPulling="2025-10-06 10:13:06.17982455 +0000 UTC m=+895.544247451" observedRunningTime="2025-10-06 10:13:13.682785893 +0000 UTC m=+903.047208774" watchObservedRunningTime="2025-10-06 10:13:13.688899212 +0000 UTC m=+903.053322073" Oct 06 10:13:13 crc kubenswrapper[4824]: I1006 10:13:13.720021 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-65d89cfd9f-gts97" podStartSLOduration=10.035561943 podStartE2EDuration="21.719996751s" podCreationTimestamp="2025-10-06 10:12:52 +0000 UTC" firstStartedPulling="2025-10-06 10:12:54.549278245 +0000 UTC m=+883.913701106" lastFinishedPulling="2025-10-06 10:13:06.233713053 +0000 UTC m=+895.598135914" observedRunningTime="2025-10-06 10:13:13.716145999 +0000 UTC m=+903.080568870" watchObservedRunningTime="2025-10-06 10:13:13.719996751 +0000 UTC m=+903.084419612" Oct 06 10:13:13 crc kubenswrapper[4824]: I1006 10:13:13.752814 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-5f7c849b98-fn8nc" podStartSLOduration=8.958060051 podStartE2EDuration="21.752784097s" podCreationTimestamp="2025-10-06 10:12:52 +0000 UTC" firstStartedPulling="2025-10-06 10:12:53.441123143 +0000 UTC m=+882.805546004" lastFinishedPulling="2025-10-06 10:13:06.235847189 +0000 UTC m=+895.600270050" observedRunningTime="2025-10-06 10:13:13.741471386 +0000 UTC m=+903.105894287" watchObservedRunningTime="2025-10-06 10:13:13.752784097 +0000 UTC m=+903.117206958" Oct 06 10:13:15 crc kubenswrapper[4824]: I1006 10:13:15.678428 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-6qgln" event={"ID":"755275e4-c198-49df-90b5-0688a4fb8228","Type":"ContainerStarted","Data":"e676d8bfdba30796dfa68191079955a7bfc947256cadd4e0ca3664acde2596e3"} Oct 06 10:13:15 crc kubenswrapper[4824]: I1006 10:13:15.679521 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-6qgln" Oct 06 10:13:15 crc kubenswrapper[4824]: I1006 10:13:15.703620 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-6qgln" podStartSLOduration=4.128795186 podStartE2EDuration="23.703594602s" podCreationTimestamp="2025-10-06 10:12:52 +0000 UTC" firstStartedPulling="2025-10-06 10:12:54.559452901 +0000 UTC m=+883.923875762" lastFinishedPulling="2025-10-06 10:13:14.134252317 +0000 UTC m=+903.498675178" observedRunningTime="2025-10-06 10:13:15.698520674 +0000 UTC m=+905.062943575" watchObservedRunningTime="2025-10-06 10:13:15.703594602 +0000 UTC m=+905.068017473" Oct 06 10:13:17 crc kubenswrapper[4824]: I1006 10:13:17.703570 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665cxcwhp" event={"ID":"fa5da9bb-805d-420f-adc6-e948cf910b21","Type":"ContainerStarted","Data":"8d8b6717f22e02599df69f47263b569bda64383d28ffca51d65ce72ec9e9fed6"} Oct 06 10:13:17 crc kubenswrapper[4824]: I1006 10:13:17.704206 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665cxcwhp" Oct 06 10:13:17 crc kubenswrapper[4824]: I1006 10:13:17.706721 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-6cbc6dd547-dglbk" event={"ID":"2a48640f-172c-4e24-8c75-5c36d26ae1aa","Type":"ContainerStarted","Data":"3d9114f3a49f613e40173c2649617bfaa1c838915c8a863c4684cff2696e9237"} Oct 06 10:13:17 crc kubenswrapper[4824]: I1006 10:13:17.707030 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-6cbc6dd547-dglbk" Oct 06 10:13:17 crc kubenswrapper[4824]: I1006 10:13:17.712046 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-54689d9f88-8qfvp" event={"ID":"1778e737-b572-417b-931d-0eec04a89ac9","Type":"ContainerStarted","Data":"a6000db2d571d215f6e1b40926fedd0446d6a4bfbfe49e098063e9f9cb31e582"} Oct 06 10:13:17 crc kubenswrapper[4824]: I1006 10:13:17.712311 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-54689d9f88-8qfvp" Oct 06 10:13:17 crc kubenswrapper[4824]: I1006 10:13:17.714515 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-c8wx2" event={"ID":"1f00bc57-93de-431a-adcf-d1e1221121a1","Type":"ContainerStarted","Data":"40ad33dced90a5288d08a00e02147edbbfb23bc1eecabb07cad8ce42c0d6df8c"} Oct 06 10:13:17 crc kubenswrapper[4824]: I1006 10:13:17.717349 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-6859f9b676-n84pw" event={"ID":"b32a09cc-901f-41f2-9912-628bbc33da7b","Type":"ContainerStarted","Data":"9b85ff969a582c97b70ce8fcc9818e9f7a48817823fd238c9bf92697ea61f14f"} Oct 06 10:13:17 crc kubenswrapper[4824]: I1006 10:13:17.717585 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-6859f9b676-n84pw" Oct 06 10:13:17 crc kubenswrapper[4824]: I1006 10:13:17.746404 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665cxcwhp" podStartSLOduration=3.815725447 podStartE2EDuration="25.746377468s" podCreationTimestamp="2025-10-06 10:12:52 +0000 UTC" firstStartedPulling="2025-10-06 10:12:54.607550242 +0000 UTC m=+883.971973103" lastFinishedPulling="2025-10-06 10:13:16.538202263 +0000 UTC m=+905.902625124" observedRunningTime="2025-10-06 10:13:17.742825822 +0000 UTC m=+907.107248743" watchObservedRunningTime="2025-10-06 10:13:17.746377468 +0000 UTC m=+907.110800329" Oct 06 10:13:17 crc kubenswrapper[4824]: I1006 10:13:17.770555 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-54689d9f88-8qfvp" podStartSLOduration=3.823205465 podStartE2EDuration="25.77053406s" podCreationTimestamp="2025-10-06 10:12:52 +0000 UTC" firstStartedPulling="2025-10-06 10:12:54.572322565 +0000 UTC m=+883.936745426" lastFinishedPulling="2025-10-06 10:13:16.51965116 +0000 UTC m=+905.884074021" observedRunningTime="2025-10-06 10:13:17.764050412 +0000 UTC m=+907.128473293" watchObservedRunningTime="2025-10-06 10:13:17.77053406 +0000 UTC m=+907.134956921" Oct 06 10:13:17 crc kubenswrapper[4824]: I1006 10:13:17.805306 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-c8wx2" podStartSLOduration=3.881842658 podStartE2EDuration="25.805284537s" podCreationTimestamp="2025-10-06 10:12:52 +0000 UTC" firstStartedPulling="2025-10-06 10:12:54.608103313 +0000 UTC m=+883.972526164" lastFinishedPulling="2025-10-06 10:13:16.531545182 +0000 UTC m=+905.895968043" observedRunningTime="2025-10-06 10:13:17.799426143 +0000 UTC m=+907.163849004" watchObservedRunningTime="2025-10-06 10:13:17.805284537 +0000 UTC m=+907.169707398" Oct 06 10:13:17 crc kubenswrapper[4824]: I1006 10:13:17.806515 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-6859f9b676-n84pw" podStartSLOduration=6.220836496 podStartE2EDuration="25.806507813s" podCreationTimestamp="2025-10-06 10:12:52 +0000 UTC" firstStartedPulling="2025-10-06 10:12:54.550602863 +0000 UTC m=+883.915025724" lastFinishedPulling="2025-10-06 10:13:14.13627418 +0000 UTC m=+903.500697041" observedRunningTime="2025-10-06 10:13:17.782417883 +0000 UTC m=+907.146840744" watchObservedRunningTime="2025-10-06 10:13:17.806507813 +0000 UTC m=+907.170930674" Oct 06 10:13:17 crc kubenswrapper[4824]: I1006 10:13:17.824556 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-6cbc6dd547-dglbk" podStartSLOduration=3.900222528 podStartE2EDuration="25.824534376s" podCreationTimestamp="2025-10-06 10:12:52 +0000 UTC" firstStartedPulling="2025-10-06 10:12:54.607992011 +0000 UTC m=+883.972414872" lastFinishedPulling="2025-10-06 10:13:16.532303859 +0000 UTC m=+905.896726720" observedRunningTime="2025-10-06 10:13:17.817725271 +0000 UTC m=+907.182148152" watchObservedRunningTime="2025-10-06 10:13:17.824534376 +0000 UTC m=+907.188957237" Oct 06 10:13:19 crc kubenswrapper[4824]: I1006 10:13:19.276773 4824 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 06 10:13:19 crc kubenswrapper[4824]: I1006 10:13:19.741538 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-84c678c567-wzzrh" event={"ID":"7c4f21a3-07f8-46d8-bd7b-e2a24bf280d8","Type":"ContainerStarted","Data":"bef28c80fb2b3de4ec30917cd1d4c3bf52247c1af9282234e95ae5da73dee27a"} Oct 06 10:13:19 crc kubenswrapper[4824]: I1006 10:13:19.742525 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-84c678c567-wzzrh" Oct 06 10:13:19 crc kubenswrapper[4824]: I1006 10:13:19.777030 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-84c678c567-wzzrh" podStartSLOduration=2.972105374 podStartE2EDuration="27.776974465s" podCreationTimestamp="2025-10-06 10:12:52 +0000 UTC" firstStartedPulling="2025-10-06 10:12:54.54709338 +0000 UTC m=+883.911516241" lastFinishedPulling="2025-10-06 10:13:19.351962471 +0000 UTC m=+908.716385332" observedRunningTime="2025-10-06 10:13:19.765642905 +0000 UTC m=+909.130065796" watchObservedRunningTime="2025-10-06 10:13:19.776974465 +0000 UTC m=+909.141397356" Oct 06 10:13:22 crc kubenswrapper[4824]: I1006 10:13:22.536433 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-5f7c849b98-fn8nc" Oct 06 10:13:22 crc kubenswrapper[4824]: I1006 10:13:22.543515 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-5f7c849b98-fn8nc" Oct 06 10:13:22 crc kubenswrapper[4824]: I1006 10:13:22.609636 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-8f58bc9db-tp4b2" Oct 06 10:13:22 crc kubenswrapper[4824]: I1006 10:13:22.616346 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-8f58bc9db-tp4b2" Oct 06 10:13:23 crc kubenswrapper[4824]: I1006 10:13:23.166377 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-6859f9b676-n84pw" Oct 06 10:13:23 crc kubenswrapper[4824]: I1006 10:13:23.357558 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-54689d9f88-8qfvp" Oct 06 10:13:23 crc kubenswrapper[4824]: I1006 10:13:23.434434 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-6cbc6dd547-dglbk" Oct 06 10:13:23 crc kubenswrapper[4824]: I1006 10:13:23.629160 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665cxcwhp" Oct 06 10:13:23 crc kubenswrapper[4824]: I1006 10:13:23.658703 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-6qgln" Oct 06 10:13:33 crc kubenswrapper[4824]: I1006 10:13:33.140761 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-84c678c567-wzzrh" Oct 06 10:13:49 crc kubenswrapper[4824]: I1006 10:13:49.592937 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-qf5nc"] Oct 06 10:13:49 crc kubenswrapper[4824]: I1006 10:13:49.595179 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-qf5nc" Oct 06 10:13:49 crc kubenswrapper[4824]: I1006 10:13:49.597547 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Oct 06 10:13:49 crc kubenswrapper[4824]: I1006 10:13:49.598518 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-7dzwk" Oct 06 10:13:49 crc kubenswrapper[4824]: I1006 10:13:49.598938 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Oct 06 10:13:49 crc kubenswrapper[4824]: I1006 10:13:49.601040 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Oct 06 10:13:49 crc kubenswrapper[4824]: I1006 10:13:49.614587 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-qf5nc"] Oct 06 10:13:49 crc kubenswrapper[4824]: I1006 10:13:49.657163 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-h2kwr"] Oct 06 10:13:49 crc kubenswrapper[4824]: I1006 10:13:49.682343 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-h2kwr" Oct 06 10:13:49 crc kubenswrapper[4824]: I1006 10:13:49.698656 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Oct 06 10:13:49 crc kubenswrapper[4824]: I1006 10:13:49.707872 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-h2kwr"] Oct 06 10:13:49 crc kubenswrapper[4824]: I1006 10:13:49.785641 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b9e0d836-802d-4cad-af61-dd2abae51bbc-config\") pod \"dnsmasq-dns-78dd6ddcc-h2kwr\" (UID: \"b9e0d836-802d-4cad-af61-dd2abae51bbc\") " pod="openstack/dnsmasq-dns-78dd6ddcc-h2kwr" Oct 06 10:13:49 crc kubenswrapper[4824]: I1006 10:13:49.785717 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9khjd\" (UniqueName: \"kubernetes.io/projected/59e8cadf-2f15-45c4-8d6d-3457f8291d14-kube-api-access-9khjd\") pod \"dnsmasq-dns-675f4bcbfc-qf5nc\" (UID: \"59e8cadf-2f15-45c4-8d6d-3457f8291d14\") " pod="openstack/dnsmasq-dns-675f4bcbfc-qf5nc" Oct 06 10:13:49 crc kubenswrapper[4824]: I1006 10:13:49.785841 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4l4nz\" (UniqueName: \"kubernetes.io/projected/b9e0d836-802d-4cad-af61-dd2abae51bbc-kube-api-access-4l4nz\") pod \"dnsmasq-dns-78dd6ddcc-h2kwr\" (UID: \"b9e0d836-802d-4cad-af61-dd2abae51bbc\") " pod="openstack/dnsmasq-dns-78dd6ddcc-h2kwr" Oct 06 10:13:49 crc kubenswrapper[4824]: I1006 10:13:49.785916 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b9e0d836-802d-4cad-af61-dd2abae51bbc-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-h2kwr\" (UID: \"b9e0d836-802d-4cad-af61-dd2abae51bbc\") " pod="openstack/dnsmasq-dns-78dd6ddcc-h2kwr" Oct 06 10:13:49 crc kubenswrapper[4824]: I1006 10:13:49.785941 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/59e8cadf-2f15-45c4-8d6d-3457f8291d14-config\") pod \"dnsmasq-dns-675f4bcbfc-qf5nc\" (UID: \"59e8cadf-2f15-45c4-8d6d-3457f8291d14\") " pod="openstack/dnsmasq-dns-675f4bcbfc-qf5nc" Oct 06 10:13:49 crc kubenswrapper[4824]: I1006 10:13:49.886968 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9khjd\" (UniqueName: \"kubernetes.io/projected/59e8cadf-2f15-45c4-8d6d-3457f8291d14-kube-api-access-9khjd\") pod \"dnsmasq-dns-675f4bcbfc-qf5nc\" (UID: \"59e8cadf-2f15-45c4-8d6d-3457f8291d14\") " pod="openstack/dnsmasq-dns-675f4bcbfc-qf5nc" Oct 06 10:13:49 crc kubenswrapper[4824]: I1006 10:13:49.887096 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4l4nz\" (UniqueName: \"kubernetes.io/projected/b9e0d836-802d-4cad-af61-dd2abae51bbc-kube-api-access-4l4nz\") pod \"dnsmasq-dns-78dd6ddcc-h2kwr\" (UID: \"b9e0d836-802d-4cad-af61-dd2abae51bbc\") " pod="openstack/dnsmasq-dns-78dd6ddcc-h2kwr" Oct 06 10:13:49 crc kubenswrapper[4824]: I1006 10:13:49.887149 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b9e0d836-802d-4cad-af61-dd2abae51bbc-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-h2kwr\" (UID: \"b9e0d836-802d-4cad-af61-dd2abae51bbc\") " pod="openstack/dnsmasq-dns-78dd6ddcc-h2kwr" Oct 06 10:13:49 crc kubenswrapper[4824]: I1006 10:13:49.887166 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/59e8cadf-2f15-45c4-8d6d-3457f8291d14-config\") pod \"dnsmasq-dns-675f4bcbfc-qf5nc\" (UID: \"59e8cadf-2f15-45c4-8d6d-3457f8291d14\") " pod="openstack/dnsmasq-dns-675f4bcbfc-qf5nc" Oct 06 10:13:49 crc kubenswrapper[4824]: I1006 10:13:49.888138 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b9e0d836-802d-4cad-af61-dd2abae51bbc-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-h2kwr\" (UID: \"b9e0d836-802d-4cad-af61-dd2abae51bbc\") " pod="openstack/dnsmasq-dns-78dd6ddcc-h2kwr" Oct 06 10:13:49 crc kubenswrapper[4824]: I1006 10:13:49.888464 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/59e8cadf-2f15-45c4-8d6d-3457f8291d14-config\") pod \"dnsmasq-dns-675f4bcbfc-qf5nc\" (UID: \"59e8cadf-2f15-45c4-8d6d-3457f8291d14\") " pod="openstack/dnsmasq-dns-675f4bcbfc-qf5nc" Oct 06 10:13:49 crc kubenswrapper[4824]: I1006 10:13:49.888813 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b9e0d836-802d-4cad-af61-dd2abae51bbc-config\") pod \"dnsmasq-dns-78dd6ddcc-h2kwr\" (UID: \"b9e0d836-802d-4cad-af61-dd2abae51bbc\") " pod="openstack/dnsmasq-dns-78dd6ddcc-h2kwr" Oct 06 10:13:49 crc kubenswrapper[4824]: I1006 10:13:49.888846 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b9e0d836-802d-4cad-af61-dd2abae51bbc-config\") pod \"dnsmasq-dns-78dd6ddcc-h2kwr\" (UID: \"b9e0d836-802d-4cad-af61-dd2abae51bbc\") " pod="openstack/dnsmasq-dns-78dd6ddcc-h2kwr" Oct 06 10:13:49 crc kubenswrapper[4824]: I1006 10:13:49.912034 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4l4nz\" (UniqueName: \"kubernetes.io/projected/b9e0d836-802d-4cad-af61-dd2abae51bbc-kube-api-access-4l4nz\") pod \"dnsmasq-dns-78dd6ddcc-h2kwr\" (UID: \"b9e0d836-802d-4cad-af61-dd2abae51bbc\") " pod="openstack/dnsmasq-dns-78dd6ddcc-h2kwr" Oct 06 10:13:49 crc kubenswrapper[4824]: I1006 10:13:49.918804 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9khjd\" (UniqueName: \"kubernetes.io/projected/59e8cadf-2f15-45c4-8d6d-3457f8291d14-kube-api-access-9khjd\") pod \"dnsmasq-dns-675f4bcbfc-qf5nc\" (UID: \"59e8cadf-2f15-45c4-8d6d-3457f8291d14\") " pod="openstack/dnsmasq-dns-675f4bcbfc-qf5nc" Oct 06 10:13:50 crc kubenswrapper[4824]: I1006 10:13:50.023044 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-h2kwr" Oct 06 10:13:50 crc kubenswrapper[4824]: I1006 10:13:50.216352 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-qf5nc" Oct 06 10:13:50 crc kubenswrapper[4824]: I1006 10:13:50.478221 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-h2kwr"] Oct 06 10:13:50 crc kubenswrapper[4824]: I1006 10:13:50.645210 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-qf5nc"] Oct 06 10:13:50 crc kubenswrapper[4824]: W1006 10:13:50.649780 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod59e8cadf_2f15_45c4_8d6d_3457f8291d14.slice/crio-f2f7d30332fcc1bc1c81eb9296db8fc976d575bfe7212017d3786f5bf645dc47 WatchSource:0}: Error finding container f2f7d30332fcc1bc1c81eb9296db8fc976d575bfe7212017d3786f5bf645dc47: Status 404 returned error can't find the container with id f2f7d30332fcc1bc1c81eb9296db8fc976d575bfe7212017d3786f5bf645dc47 Oct 06 10:13:51 crc kubenswrapper[4824]: I1006 10:13:51.024153 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-h2kwr" event={"ID":"b9e0d836-802d-4cad-af61-dd2abae51bbc","Type":"ContainerStarted","Data":"d71a18afba91462402ec1ef6dd857fbaa42b0f455d592cf7b622c6c61a744af0"} Oct 06 10:13:51 crc kubenswrapper[4824]: I1006 10:13:51.026161 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-qf5nc" event={"ID":"59e8cadf-2f15-45c4-8d6d-3457f8291d14","Type":"ContainerStarted","Data":"f2f7d30332fcc1bc1c81eb9296db8fc976d575bfe7212017d3786f5bf645dc47"} Oct 06 10:13:52 crc kubenswrapper[4824]: I1006 10:13:52.641267 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-qf5nc"] Oct 06 10:13:52 crc kubenswrapper[4824]: I1006 10:13:52.671253 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-bdp5v"] Oct 06 10:13:52 crc kubenswrapper[4824]: I1006 10:13:52.673644 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-bdp5v" Oct 06 10:13:52 crc kubenswrapper[4824]: I1006 10:13:52.682396 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-bdp5v"] Oct 06 10:13:52 crc kubenswrapper[4824]: I1006 10:13:52.740341 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c7eb054c-bd83-4ba1-a1ac-f8eeaf049fb7-dns-svc\") pod \"dnsmasq-dns-666b6646f7-bdp5v\" (UID: \"c7eb054c-bd83-4ba1-a1ac-f8eeaf049fb7\") " pod="openstack/dnsmasq-dns-666b6646f7-bdp5v" Oct 06 10:13:52 crc kubenswrapper[4824]: I1006 10:13:52.740513 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c7eb054c-bd83-4ba1-a1ac-f8eeaf049fb7-config\") pod \"dnsmasq-dns-666b6646f7-bdp5v\" (UID: \"c7eb054c-bd83-4ba1-a1ac-f8eeaf049fb7\") " pod="openstack/dnsmasq-dns-666b6646f7-bdp5v" Oct 06 10:13:52 crc kubenswrapper[4824]: I1006 10:13:52.740746 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gfv2m\" (UniqueName: \"kubernetes.io/projected/c7eb054c-bd83-4ba1-a1ac-f8eeaf049fb7-kube-api-access-gfv2m\") pod \"dnsmasq-dns-666b6646f7-bdp5v\" (UID: \"c7eb054c-bd83-4ba1-a1ac-f8eeaf049fb7\") " pod="openstack/dnsmasq-dns-666b6646f7-bdp5v" Oct 06 10:13:52 crc kubenswrapper[4824]: I1006 10:13:52.844051 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gfv2m\" (UniqueName: \"kubernetes.io/projected/c7eb054c-bd83-4ba1-a1ac-f8eeaf049fb7-kube-api-access-gfv2m\") pod \"dnsmasq-dns-666b6646f7-bdp5v\" (UID: \"c7eb054c-bd83-4ba1-a1ac-f8eeaf049fb7\") " pod="openstack/dnsmasq-dns-666b6646f7-bdp5v" Oct 06 10:13:52 crc kubenswrapper[4824]: I1006 10:13:52.844140 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c7eb054c-bd83-4ba1-a1ac-f8eeaf049fb7-dns-svc\") pod \"dnsmasq-dns-666b6646f7-bdp5v\" (UID: \"c7eb054c-bd83-4ba1-a1ac-f8eeaf049fb7\") " pod="openstack/dnsmasq-dns-666b6646f7-bdp5v" Oct 06 10:13:52 crc kubenswrapper[4824]: I1006 10:13:52.844190 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c7eb054c-bd83-4ba1-a1ac-f8eeaf049fb7-config\") pod \"dnsmasq-dns-666b6646f7-bdp5v\" (UID: \"c7eb054c-bd83-4ba1-a1ac-f8eeaf049fb7\") " pod="openstack/dnsmasq-dns-666b6646f7-bdp5v" Oct 06 10:13:52 crc kubenswrapper[4824]: I1006 10:13:52.845449 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c7eb054c-bd83-4ba1-a1ac-f8eeaf049fb7-config\") pod \"dnsmasq-dns-666b6646f7-bdp5v\" (UID: \"c7eb054c-bd83-4ba1-a1ac-f8eeaf049fb7\") " pod="openstack/dnsmasq-dns-666b6646f7-bdp5v" Oct 06 10:13:52 crc kubenswrapper[4824]: I1006 10:13:52.846180 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c7eb054c-bd83-4ba1-a1ac-f8eeaf049fb7-dns-svc\") pod \"dnsmasq-dns-666b6646f7-bdp5v\" (UID: \"c7eb054c-bd83-4ba1-a1ac-f8eeaf049fb7\") " pod="openstack/dnsmasq-dns-666b6646f7-bdp5v" Oct 06 10:13:52 crc kubenswrapper[4824]: I1006 10:13:52.880549 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gfv2m\" (UniqueName: \"kubernetes.io/projected/c7eb054c-bd83-4ba1-a1ac-f8eeaf049fb7-kube-api-access-gfv2m\") pod \"dnsmasq-dns-666b6646f7-bdp5v\" (UID: \"c7eb054c-bd83-4ba1-a1ac-f8eeaf049fb7\") " pod="openstack/dnsmasq-dns-666b6646f7-bdp5v" Oct 06 10:13:52 crc kubenswrapper[4824]: I1006 10:13:52.985629 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-h2kwr"] Oct 06 10:13:53 crc kubenswrapper[4824]: I1006 10:13:53.018760 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-bdp5v" Oct 06 10:13:53 crc kubenswrapper[4824]: I1006 10:13:53.022400 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-cnwhb"] Oct 06 10:13:53 crc kubenswrapper[4824]: I1006 10:13:53.023659 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-cnwhb" Oct 06 10:13:53 crc kubenswrapper[4824]: I1006 10:13:53.031097 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-cnwhb"] Oct 06 10:13:53 crc kubenswrapper[4824]: I1006 10:13:53.148619 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pbw72\" (UniqueName: \"kubernetes.io/projected/97cd45b1-470f-483b-b1ed-6dfd5fed1ca8-kube-api-access-pbw72\") pod \"dnsmasq-dns-57d769cc4f-cnwhb\" (UID: \"97cd45b1-470f-483b-b1ed-6dfd5fed1ca8\") " pod="openstack/dnsmasq-dns-57d769cc4f-cnwhb" Oct 06 10:13:53 crc kubenswrapper[4824]: I1006 10:13:53.148719 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/97cd45b1-470f-483b-b1ed-6dfd5fed1ca8-config\") pod \"dnsmasq-dns-57d769cc4f-cnwhb\" (UID: \"97cd45b1-470f-483b-b1ed-6dfd5fed1ca8\") " pod="openstack/dnsmasq-dns-57d769cc4f-cnwhb" Oct 06 10:13:53 crc kubenswrapper[4824]: I1006 10:13:53.150172 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/97cd45b1-470f-483b-b1ed-6dfd5fed1ca8-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-cnwhb\" (UID: \"97cd45b1-470f-483b-b1ed-6dfd5fed1ca8\") " pod="openstack/dnsmasq-dns-57d769cc4f-cnwhb" Oct 06 10:13:53 crc kubenswrapper[4824]: I1006 10:13:53.251397 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pbw72\" (UniqueName: \"kubernetes.io/projected/97cd45b1-470f-483b-b1ed-6dfd5fed1ca8-kube-api-access-pbw72\") pod \"dnsmasq-dns-57d769cc4f-cnwhb\" (UID: \"97cd45b1-470f-483b-b1ed-6dfd5fed1ca8\") " pod="openstack/dnsmasq-dns-57d769cc4f-cnwhb" Oct 06 10:13:53 crc kubenswrapper[4824]: I1006 10:13:53.251525 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/97cd45b1-470f-483b-b1ed-6dfd5fed1ca8-config\") pod \"dnsmasq-dns-57d769cc4f-cnwhb\" (UID: \"97cd45b1-470f-483b-b1ed-6dfd5fed1ca8\") " pod="openstack/dnsmasq-dns-57d769cc4f-cnwhb" Oct 06 10:13:53 crc kubenswrapper[4824]: I1006 10:13:53.251643 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/97cd45b1-470f-483b-b1ed-6dfd5fed1ca8-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-cnwhb\" (UID: \"97cd45b1-470f-483b-b1ed-6dfd5fed1ca8\") " pod="openstack/dnsmasq-dns-57d769cc4f-cnwhb" Oct 06 10:13:53 crc kubenswrapper[4824]: I1006 10:13:53.253111 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/97cd45b1-470f-483b-b1ed-6dfd5fed1ca8-config\") pod \"dnsmasq-dns-57d769cc4f-cnwhb\" (UID: \"97cd45b1-470f-483b-b1ed-6dfd5fed1ca8\") " pod="openstack/dnsmasq-dns-57d769cc4f-cnwhb" Oct 06 10:13:53 crc kubenswrapper[4824]: I1006 10:13:53.253329 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/97cd45b1-470f-483b-b1ed-6dfd5fed1ca8-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-cnwhb\" (UID: \"97cd45b1-470f-483b-b1ed-6dfd5fed1ca8\") " pod="openstack/dnsmasq-dns-57d769cc4f-cnwhb" Oct 06 10:13:53 crc kubenswrapper[4824]: I1006 10:13:53.269260 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pbw72\" (UniqueName: \"kubernetes.io/projected/97cd45b1-470f-483b-b1ed-6dfd5fed1ca8-kube-api-access-pbw72\") pod \"dnsmasq-dns-57d769cc4f-cnwhb\" (UID: \"97cd45b1-470f-483b-b1ed-6dfd5fed1ca8\") " pod="openstack/dnsmasq-dns-57d769cc4f-cnwhb" Oct 06 10:13:53 crc kubenswrapper[4824]: I1006 10:13:53.359795 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-cnwhb" Oct 06 10:13:53 crc kubenswrapper[4824]: I1006 10:13:53.628862 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-bdp5v"] Oct 06 10:13:53 crc kubenswrapper[4824]: I1006 10:13:53.656571 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-cnwhb"] Oct 06 10:13:53 crc kubenswrapper[4824]: W1006 10:13:53.669475 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod97cd45b1_470f_483b_b1ed_6dfd5fed1ca8.slice/crio-2c2b14f57a26c7784a998f41c494e1840f6fe44fb9ad5607e6ebc16cfcdf70fb WatchSource:0}: Error finding container 2c2b14f57a26c7784a998f41c494e1840f6fe44fb9ad5607e6ebc16cfcdf70fb: Status 404 returned error can't find the container with id 2c2b14f57a26c7784a998f41c494e1840f6fe44fb9ad5607e6ebc16cfcdf70fb Oct 06 10:13:53 crc kubenswrapper[4824]: I1006 10:13:53.845969 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Oct 06 10:13:53 crc kubenswrapper[4824]: I1006 10:13:53.850930 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 06 10:13:53 crc kubenswrapper[4824]: I1006 10:13:53.851275 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 06 10:13:53 crc kubenswrapper[4824]: I1006 10:13:53.855697 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Oct 06 10:13:53 crc kubenswrapper[4824]: I1006 10:13:53.855816 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-r49v9" Oct 06 10:13:53 crc kubenswrapper[4824]: I1006 10:13:53.855697 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Oct 06 10:13:53 crc kubenswrapper[4824]: I1006 10:13:53.858161 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Oct 06 10:13:53 crc kubenswrapper[4824]: I1006 10:13:53.858193 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Oct 06 10:13:53 crc kubenswrapper[4824]: I1006 10:13:53.858216 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Oct 06 10:13:53 crc kubenswrapper[4824]: I1006 10:13:53.858411 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Oct 06 10:13:53 crc kubenswrapper[4824]: I1006 10:13:53.961320 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/f7d047da-89e3-4cd4-bf02-e6a3e15c585f-server-conf\") pod \"rabbitmq-server-0\" (UID: \"f7d047da-89e3-4cd4-bf02-e6a3e15c585f\") " pod="openstack/rabbitmq-server-0" Oct 06 10:13:53 crc kubenswrapper[4824]: I1006 10:13:53.961376 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f7d047da-89e3-4cd4-bf02-e6a3e15c585f-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"f7d047da-89e3-4cd4-bf02-e6a3e15c585f\") " pod="openstack/rabbitmq-server-0" Oct 06 10:13:53 crc kubenswrapper[4824]: I1006 10:13:53.961398 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8vcmv\" (UniqueName: \"kubernetes.io/projected/f7d047da-89e3-4cd4-bf02-e6a3e15c585f-kube-api-access-8vcmv\") pod \"rabbitmq-server-0\" (UID: \"f7d047da-89e3-4cd4-bf02-e6a3e15c585f\") " pod="openstack/rabbitmq-server-0" Oct 06 10:13:53 crc kubenswrapper[4824]: I1006 10:13:53.961419 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f7d047da-89e3-4cd4-bf02-e6a3e15c585f-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"f7d047da-89e3-4cd4-bf02-e6a3e15c585f\") " pod="openstack/rabbitmq-server-0" Oct 06 10:13:53 crc kubenswrapper[4824]: I1006 10:13:53.961456 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f7d047da-89e3-4cd4-bf02-e6a3e15c585f-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"f7d047da-89e3-4cd4-bf02-e6a3e15c585f\") " pod="openstack/rabbitmq-server-0" Oct 06 10:13:53 crc kubenswrapper[4824]: I1006 10:13:53.961493 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f7d047da-89e3-4cd4-bf02-e6a3e15c585f-pod-info\") pod \"rabbitmq-server-0\" (UID: \"f7d047da-89e3-4cd4-bf02-e6a3e15c585f\") " pod="openstack/rabbitmq-server-0" Oct 06 10:13:53 crc kubenswrapper[4824]: I1006 10:13:53.961510 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"f7d047da-89e3-4cd4-bf02-e6a3e15c585f\") " pod="openstack/rabbitmq-server-0" Oct 06 10:13:53 crc kubenswrapper[4824]: I1006 10:13:53.961532 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f7d047da-89e3-4cd4-bf02-e6a3e15c585f-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"f7d047da-89e3-4cd4-bf02-e6a3e15c585f\") " pod="openstack/rabbitmq-server-0" Oct 06 10:13:53 crc kubenswrapper[4824]: I1006 10:13:53.961563 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f7d047da-89e3-4cd4-bf02-e6a3e15c585f-config-data\") pod \"rabbitmq-server-0\" (UID: \"f7d047da-89e3-4cd4-bf02-e6a3e15c585f\") " pod="openstack/rabbitmq-server-0" Oct 06 10:13:53 crc kubenswrapper[4824]: I1006 10:13:53.961578 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f7d047da-89e3-4cd4-bf02-e6a3e15c585f-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"f7d047da-89e3-4cd4-bf02-e6a3e15c585f\") " pod="openstack/rabbitmq-server-0" Oct 06 10:13:53 crc kubenswrapper[4824]: I1006 10:13:53.961597 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/f7d047da-89e3-4cd4-bf02-e6a3e15c585f-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"f7d047da-89e3-4cd4-bf02-e6a3e15c585f\") " pod="openstack/rabbitmq-server-0" Oct 06 10:13:54 crc kubenswrapper[4824]: I1006 10:13:54.062657 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f7d047da-89e3-4cd4-bf02-e6a3e15c585f-config-data\") pod \"rabbitmq-server-0\" (UID: \"f7d047da-89e3-4cd4-bf02-e6a3e15c585f\") " pod="openstack/rabbitmq-server-0" Oct 06 10:13:54 crc kubenswrapper[4824]: I1006 10:13:54.062895 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f7d047da-89e3-4cd4-bf02-e6a3e15c585f-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"f7d047da-89e3-4cd4-bf02-e6a3e15c585f\") " pod="openstack/rabbitmq-server-0" Oct 06 10:13:54 crc kubenswrapper[4824]: I1006 10:13:54.062926 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/f7d047da-89e3-4cd4-bf02-e6a3e15c585f-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"f7d047da-89e3-4cd4-bf02-e6a3e15c585f\") " pod="openstack/rabbitmq-server-0" Oct 06 10:13:54 crc kubenswrapper[4824]: I1006 10:13:54.062973 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/f7d047da-89e3-4cd4-bf02-e6a3e15c585f-server-conf\") pod \"rabbitmq-server-0\" (UID: \"f7d047da-89e3-4cd4-bf02-e6a3e15c585f\") " pod="openstack/rabbitmq-server-0" Oct 06 10:13:54 crc kubenswrapper[4824]: I1006 10:13:54.063017 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f7d047da-89e3-4cd4-bf02-e6a3e15c585f-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"f7d047da-89e3-4cd4-bf02-e6a3e15c585f\") " pod="openstack/rabbitmq-server-0" Oct 06 10:13:54 crc kubenswrapper[4824]: I1006 10:13:54.063041 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f7d047da-89e3-4cd4-bf02-e6a3e15c585f-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"f7d047da-89e3-4cd4-bf02-e6a3e15c585f\") " pod="openstack/rabbitmq-server-0" Oct 06 10:13:54 crc kubenswrapper[4824]: I1006 10:13:54.063058 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8vcmv\" (UniqueName: \"kubernetes.io/projected/f7d047da-89e3-4cd4-bf02-e6a3e15c585f-kube-api-access-8vcmv\") pod \"rabbitmq-server-0\" (UID: \"f7d047da-89e3-4cd4-bf02-e6a3e15c585f\") " pod="openstack/rabbitmq-server-0" Oct 06 10:13:54 crc kubenswrapper[4824]: I1006 10:13:54.063109 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f7d047da-89e3-4cd4-bf02-e6a3e15c585f-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"f7d047da-89e3-4cd4-bf02-e6a3e15c585f\") " pod="openstack/rabbitmq-server-0" Oct 06 10:13:54 crc kubenswrapper[4824]: I1006 10:13:54.063148 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f7d047da-89e3-4cd4-bf02-e6a3e15c585f-pod-info\") pod \"rabbitmq-server-0\" (UID: \"f7d047da-89e3-4cd4-bf02-e6a3e15c585f\") " pod="openstack/rabbitmq-server-0" Oct 06 10:13:54 crc kubenswrapper[4824]: I1006 10:13:54.063169 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"f7d047da-89e3-4cd4-bf02-e6a3e15c585f\") " pod="openstack/rabbitmq-server-0" Oct 06 10:13:54 crc kubenswrapper[4824]: I1006 10:13:54.063194 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f7d047da-89e3-4cd4-bf02-e6a3e15c585f-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"f7d047da-89e3-4cd4-bf02-e6a3e15c585f\") " pod="openstack/rabbitmq-server-0" Oct 06 10:13:54 crc kubenswrapper[4824]: I1006 10:13:54.064477 4824 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"f7d047da-89e3-4cd4-bf02-e6a3e15c585f\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/rabbitmq-server-0" Oct 06 10:13:54 crc kubenswrapper[4824]: I1006 10:13:54.064550 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f7d047da-89e3-4cd4-bf02-e6a3e15c585f-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"f7d047da-89e3-4cd4-bf02-e6a3e15c585f\") " pod="openstack/rabbitmq-server-0" Oct 06 10:13:54 crc kubenswrapper[4824]: I1006 10:13:54.064667 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f7d047da-89e3-4cd4-bf02-e6a3e15c585f-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"f7d047da-89e3-4cd4-bf02-e6a3e15c585f\") " pod="openstack/rabbitmq-server-0" Oct 06 10:13:54 crc kubenswrapper[4824]: I1006 10:13:54.064724 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f7d047da-89e3-4cd4-bf02-e6a3e15c585f-config-data\") pod \"rabbitmq-server-0\" (UID: \"f7d047da-89e3-4cd4-bf02-e6a3e15c585f\") " pod="openstack/rabbitmq-server-0" Oct 06 10:13:54 crc kubenswrapper[4824]: I1006 10:13:54.065123 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f7d047da-89e3-4cd4-bf02-e6a3e15c585f-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"f7d047da-89e3-4cd4-bf02-e6a3e15c585f\") " pod="openstack/rabbitmq-server-0" Oct 06 10:13:54 crc kubenswrapper[4824]: I1006 10:13:54.066356 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/f7d047da-89e3-4cd4-bf02-e6a3e15c585f-server-conf\") pod \"rabbitmq-server-0\" (UID: \"f7d047da-89e3-4cd4-bf02-e6a3e15c585f\") " pod="openstack/rabbitmq-server-0" Oct 06 10:13:54 crc kubenswrapper[4824]: I1006 10:13:54.069942 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f7d047da-89e3-4cd4-bf02-e6a3e15c585f-pod-info\") pod \"rabbitmq-server-0\" (UID: \"f7d047da-89e3-4cd4-bf02-e6a3e15c585f\") " pod="openstack/rabbitmq-server-0" Oct 06 10:13:54 crc kubenswrapper[4824]: I1006 10:13:54.072370 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f7d047da-89e3-4cd4-bf02-e6a3e15c585f-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"f7d047da-89e3-4cd4-bf02-e6a3e15c585f\") " pod="openstack/rabbitmq-server-0" Oct 06 10:13:54 crc kubenswrapper[4824]: I1006 10:13:54.073161 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/f7d047da-89e3-4cd4-bf02-e6a3e15c585f-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"f7d047da-89e3-4cd4-bf02-e6a3e15c585f\") " pod="openstack/rabbitmq-server-0" Oct 06 10:13:54 crc kubenswrapper[4824]: I1006 10:13:54.077233 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-bdp5v" event={"ID":"c7eb054c-bd83-4ba1-a1ac-f8eeaf049fb7","Type":"ContainerStarted","Data":"671fef97b2748d4c605e089dbe734c96ba2bc2ec2eab693969b46a8325677ff4"} Oct 06 10:13:54 crc kubenswrapper[4824]: I1006 10:13:54.077908 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f7d047da-89e3-4cd4-bf02-e6a3e15c585f-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"f7d047da-89e3-4cd4-bf02-e6a3e15c585f\") " pod="openstack/rabbitmq-server-0" Oct 06 10:13:54 crc kubenswrapper[4824]: I1006 10:13:54.080131 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-cnwhb" event={"ID":"97cd45b1-470f-483b-b1ed-6dfd5fed1ca8","Type":"ContainerStarted","Data":"2c2b14f57a26c7784a998f41c494e1840f6fe44fb9ad5607e6ebc16cfcdf70fb"} Oct 06 10:13:54 crc kubenswrapper[4824]: I1006 10:13:54.083450 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8vcmv\" (UniqueName: \"kubernetes.io/projected/f7d047da-89e3-4cd4-bf02-e6a3e15c585f-kube-api-access-8vcmv\") pod \"rabbitmq-server-0\" (UID: \"f7d047da-89e3-4cd4-bf02-e6a3e15c585f\") " pod="openstack/rabbitmq-server-0" Oct 06 10:13:54 crc kubenswrapper[4824]: I1006 10:13:54.100471 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"f7d047da-89e3-4cd4-bf02-e6a3e15c585f\") " pod="openstack/rabbitmq-server-0" Oct 06 10:13:54 crc kubenswrapper[4824]: I1006 10:13:54.148246 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 06 10:13:54 crc kubenswrapper[4824]: I1006 10:13:54.152259 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 06 10:13:54 crc kubenswrapper[4824]: I1006 10:13:54.158084 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Oct 06 10:13:54 crc kubenswrapper[4824]: I1006 10:13:54.158281 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Oct 06 10:13:54 crc kubenswrapper[4824]: I1006 10:13:54.158296 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Oct 06 10:13:54 crc kubenswrapper[4824]: I1006 10:13:54.158354 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-px9kh" Oct 06 10:13:54 crc kubenswrapper[4824]: I1006 10:13:54.158388 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Oct 06 10:13:54 crc kubenswrapper[4824]: I1006 10:13:54.158478 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Oct 06 10:13:54 crc kubenswrapper[4824]: I1006 10:13:54.158288 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Oct 06 10:13:54 crc kubenswrapper[4824]: I1006 10:13:54.163143 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 06 10:13:54 crc kubenswrapper[4824]: I1006 10:13:54.210971 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 06 10:13:54 crc kubenswrapper[4824]: I1006 10:13:54.266232 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"c8dbad1e-02be-43c8-8f80-ab771ee81742\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 10:13:54 crc kubenswrapper[4824]: I1006 10:13:54.266296 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c8dbad1e-02be-43c8-8f80-ab771ee81742-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"c8dbad1e-02be-43c8-8f80-ab771ee81742\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 10:13:54 crc kubenswrapper[4824]: I1006 10:13:54.266344 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c8dbad1e-02be-43c8-8f80-ab771ee81742-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"c8dbad1e-02be-43c8-8f80-ab771ee81742\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 10:13:54 crc kubenswrapper[4824]: I1006 10:13:54.266796 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c8dbad1e-02be-43c8-8f80-ab771ee81742-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"c8dbad1e-02be-43c8-8f80-ab771ee81742\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 10:13:54 crc kubenswrapper[4824]: I1006 10:13:54.266843 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c8dbad1e-02be-43c8-8f80-ab771ee81742-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"c8dbad1e-02be-43c8-8f80-ab771ee81742\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 10:13:54 crc kubenswrapper[4824]: I1006 10:13:54.266915 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c8dbad1e-02be-43c8-8f80-ab771ee81742-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"c8dbad1e-02be-43c8-8f80-ab771ee81742\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 10:13:54 crc kubenswrapper[4824]: I1006 10:13:54.266945 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c8dbad1e-02be-43c8-8f80-ab771ee81742-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"c8dbad1e-02be-43c8-8f80-ab771ee81742\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 10:13:54 crc kubenswrapper[4824]: I1006 10:13:54.267013 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c8dbad1e-02be-43c8-8f80-ab771ee81742-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"c8dbad1e-02be-43c8-8f80-ab771ee81742\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 10:13:54 crc kubenswrapper[4824]: I1006 10:13:54.267035 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c8dbad1e-02be-43c8-8f80-ab771ee81742-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"c8dbad1e-02be-43c8-8f80-ab771ee81742\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 10:13:54 crc kubenswrapper[4824]: I1006 10:13:54.267214 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c8dbad1e-02be-43c8-8f80-ab771ee81742-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"c8dbad1e-02be-43c8-8f80-ab771ee81742\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 10:13:54 crc kubenswrapper[4824]: I1006 10:13:54.267271 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sdw27\" (UniqueName: \"kubernetes.io/projected/c8dbad1e-02be-43c8-8f80-ab771ee81742-kube-api-access-sdw27\") pod \"rabbitmq-cell1-server-0\" (UID: \"c8dbad1e-02be-43c8-8f80-ab771ee81742\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 10:13:54 crc kubenswrapper[4824]: I1006 10:13:54.368703 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c8dbad1e-02be-43c8-8f80-ab771ee81742-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"c8dbad1e-02be-43c8-8f80-ab771ee81742\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 10:13:54 crc kubenswrapper[4824]: I1006 10:13:54.369325 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c8dbad1e-02be-43c8-8f80-ab771ee81742-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"c8dbad1e-02be-43c8-8f80-ab771ee81742\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 10:13:54 crc kubenswrapper[4824]: I1006 10:13:54.369353 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c8dbad1e-02be-43c8-8f80-ab771ee81742-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"c8dbad1e-02be-43c8-8f80-ab771ee81742\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 10:13:54 crc kubenswrapper[4824]: I1006 10:13:54.369403 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c8dbad1e-02be-43c8-8f80-ab771ee81742-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"c8dbad1e-02be-43c8-8f80-ab771ee81742\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 10:13:54 crc kubenswrapper[4824]: I1006 10:13:54.369433 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sdw27\" (UniqueName: \"kubernetes.io/projected/c8dbad1e-02be-43c8-8f80-ab771ee81742-kube-api-access-sdw27\") pod \"rabbitmq-cell1-server-0\" (UID: \"c8dbad1e-02be-43c8-8f80-ab771ee81742\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 10:13:54 crc kubenswrapper[4824]: I1006 10:13:54.369478 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"c8dbad1e-02be-43c8-8f80-ab771ee81742\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 10:13:54 crc kubenswrapper[4824]: I1006 10:13:54.369508 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c8dbad1e-02be-43c8-8f80-ab771ee81742-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"c8dbad1e-02be-43c8-8f80-ab771ee81742\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 10:13:54 crc kubenswrapper[4824]: I1006 10:13:54.369548 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c8dbad1e-02be-43c8-8f80-ab771ee81742-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"c8dbad1e-02be-43c8-8f80-ab771ee81742\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 10:13:54 crc kubenswrapper[4824]: I1006 10:13:54.369592 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c8dbad1e-02be-43c8-8f80-ab771ee81742-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"c8dbad1e-02be-43c8-8f80-ab771ee81742\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 10:13:54 crc kubenswrapper[4824]: I1006 10:13:54.369617 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c8dbad1e-02be-43c8-8f80-ab771ee81742-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"c8dbad1e-02be-43c8-8f80-ab771ee81742\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 10:13:54 crc kubenswrapper[4824]: I1006 10:13:54.369656 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c8dbad1e-02be-43c8-8f80-ab771ee81742-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"c8dbad1e-02be-43c8-8f80-ab771ee81742\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 10:13:54 crc kubenswrapper[4824]: I1006 10:13:54.370684 4824 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"c8dbad1e-02be-43c8-8f80-ab771ee81742\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/rabbitmq-cell1-server-0" Oct 06 10:13:54 crc kubenswrapper[4824]: I1006 10:13:54.374517 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c8dbad1e-02be-43c8-8f80-ab771ee81742-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"c8dbad1e-02be-43c8-8f80-ab771ee81742\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 10:13:54 crc kubenswrapper[4824]: I1006 10:13:54.374517 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c8dbad1e-02be-43c8-8f80-ab771ee81742-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"c8dbad1e-02be-43c8-8f80-ab771ee81742\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 10:13:54 crc kubenswrapper[4824]: I1006 10:13:54.374623 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c8dbad1e-02be-43c8-8f80-ab771ee81742-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"c8dbad1e-02be-43c8-8f80-ab771ee81742\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 10:13:54 crc kubenswrapper[4824]: I1006 10:13:54.375354 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c8dbad1e-02be-43c8-8f80-ab771ee81742-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"c8dbad1e-02be-43c8-8f80-ab771ee81742\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 10:13:54 crc kubenswrapper[4824]: I1006 10:13:54.375374 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c8dbad1e-02be-43c8-8f80-ab771ee81742-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"c8dbad1e-02be-43c8-8f80-ab771ee81742\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 10:13:54 crc kubenswrapper[4824]: I1006 10:13:54.375908 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c8dbad1e-02be-43c8-8f80-ab771ee81742-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"c8dbad1e-02be-43c8-8f80-ab771ee81742\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 10:13:54 crc kubenswrapper[4824]: I1006 10:13:54.376032 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c8dbad1e-02be-43c8-8f80-ab771ee81742-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"c8dbad1e-02be-43c8-8f80-ab771ee81742\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 10:13:54 crc kubenswrapper[4824]: I1006 10:13:54.376969 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c8dbad1e-02be-43c8-8f80-ab771ee81742-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"c8dbad1e-02be-43c8-8f80-ab771ee81742\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 10:13:54 crc kubenswrapper[4824]: I1006 10:13:54.378848 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c8dbad1e-02be-43c8-8f80-ab771ee81742-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"c8dbad1e-02be-43c8-8f80-ab771ee81742\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 10:13:54 crc kubenswrapper[4824]: I1006 10:13:54.396727 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sdw27\" (UniqueName: \"kubernetes.io/projected/c8dbad1e-02be-43c8-8f80-ab771ee81742-kube-api-access-sdw27\") pod \"rabbitmq-cell1-server-0\" (UID: \"c8dbad1e-02be-43c8-8f80-ab771ee81742\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 10:13:54 crc kubenswrapper[4824]: I1006 10:13:54.423619 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"c8dbad1e-02be-43c8-8f80-ab771ee81742\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 10:13:54 crc kubenswrapper[4824]: I1006 10:13:54.502633 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 06 10:13:54 crc kubenswrapper[4824]: I1006 10:13:54.685658 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 06 10:13:54 crc kubenswrapper[4824]: I1006 10:13:54.976274 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 06 10:13:56 crc kubenswrapper[4824]: I1006 10:13:56.745021 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Oct 06 10:13:56 crc kubenswrapper[4824]: I1006 10:13:56.747241 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 06 10:13:56 crc kubenswrapper[4824]: I1006 10:13:56.751200 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Oct 06 10:13:56 crc kubenswrapper[4824]: I1006 10:13:56.752066 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Oct 06 10:13:56 crc kubenswrapper[4824]: I1006 10:13:56.752091 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Oct 06 10:13:56 crc kubenswrapper[4824]: I1006 10:13:56.752286 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-4kr2p" Oct 06 10:13:56 crc kubenswrapper[4824]: I1006 10:13:56.752903 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Oct 06 10:13:56 crc kubenswrapper[4824]: I1006 10:13:56.757371 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Oct 06 10:13:56 crc kubenswrapper[4824]: I1006 10:13:56.759087 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Oct 06 10:13:56 crc kubenswrapper[4824]: I1006 10:13:56.858545 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 06 10:13:56 crc kubenswrapper[4824]: I1006 10:13:56.859785 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 06 10:13:56 crc kubenswrapper[4824]: I1006 10:13:56.862012 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-klp6g" Oct 06 10:13:56 crc kubenswrapper[4824]: I1006 10:13:56.865929 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Oct 06 10:13:56 crc kubenswrapper[4824]: I1006 10:13:56.866069 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Oct 06 10:13:56 crc kubenswrapper[4824]: I1006 10:13:56.867399 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Oct 06 10:13:56 crc kubenswrapper[4824]: I1006 10:13:56.870849 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 06 10:13:56 crc kubenswrapper[4824]: I1006 10:13:56.913634 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/58f6ad3c-7a20-4be6-8ffb-cadff1c6adef-operator-scripts\") pod \"openstack-galera-0\" (UID: \"58f6ad3c-7a20-4be6-8ffb-cadff1c6adef\") " pod="openstack/openstack-galera-0" Oct 06 10:13:56 crc kubenswrapper[4824]: I1006 10:13:56.913685 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58f6ad3c-7a20-4be6-8ffb-cadff1c6adef-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"58f6ad3c-7a20-4be6-8ffb-cadff1c6adef\") " pod="openstack/openstack-galera-0" Oct 06 10:13:56 crc kubenswrapper[4824]: I1006 10:13:56.913829 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/58f6ad3c-7a20-4be6-8ffb-cadff1c6adef-config-data-default\") pod \"openstack-galera-0\" (UID: \"58f6ad3c-7a20-4be6-8ffb-cadff1c6adef\") " pod="openstack/openstack-galera-0" Oct 06 10:13:56 crc kubenswrapper[4824]: I1006 10:13:56.913942 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/58f6ad3c-7a20-4be6-8ffb-cadff1c6adef-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"58f6ad3c-7a20-4be6-8ffb-cadff1c6adef\") " pod="openstack/openstack-galera-0" Oct 06 10:13:56 crc kubenswrapper[4824]: I1006 10:13:56.913965 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zpwtq\" (UniqueName: \"kubernetes.io/projected/58f6ad3c-7a20-4be6-8ffb-cadff1c6adef-kube-api-access-zpwtq\") pod \"openstack-galera-0\" (UID: \"58f6ad3c-7a20-4be6-8ffb-cadff1c6adef\") " pod="openstack/openstack-galera-0" Oct 06 10:13:56 crc kubenswrapper[4824]: I1006 10:13:56.914061 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/58f6ad3c-7a20-4be6-8ffb-cadff1c6adef-kolla-config\") pod \"openstack-galera-0\" (UID: \"58f6ad3c-7a20-4be6-8ffb-cadff1c6adef\") " pod="openstack/openstack-galera-0" Oct 06 10:13:56 crc kubenswrapper[4824]: I1006 10:13:56.914106 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/58f6ad3c-7a20-4be6-8ffb-cadff1c6adef-config-data-generated\") pod \"openstack-galera-0\" (UID: \"58f6ad3c-7a20-4be6-8ffb-cadff1c6adef\") " pod="openstack/openstack-galera-0" Oct 06 10:13:56 crc kubenswrapper[4824]: I1006 10:13:56.914128 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"openstack-galera-0\" (UID: \"58f6ad3c-7a20-4be6-8ffb-cadff1c6adef\") " pod="openstack/openstack-galera-0" Oct 06 10:13:56 crc kubenswrapper[4824]: I1006 10:13:56.914193 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/58f6ad3c-7a20-4be6-8ffb-cadff1c6adef-secrets\") pod \"openstack-galera-0\" (UID: \"58f6ad3c-7a20-4be6-8ffb-cadff1c6adef\") " pod="openstack/openstack-galera-0" Oct 06 10:13:57 crc kubenswrapper[4824]: I1006 10:13:57.016494 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/1510d553-3505-4bd2-9666-c95cecc43d01-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"1510d553-3505-4bd2-9666-c95cecc43d01\") " pod="openstack/openstack-cell1-galera-0" Oct 06 10:13:57 crc kubenswrapper[4824]: I1006 10:13:57.016551 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-cell1-galera-0\" (UID: \"1510d553-3505-4bd2-9666-c95cecc43d01\") " pod="openstack/openstack-cell1-galera-0" Oct 06 10:13:57 crc kubenswrapper[4824]: I1006 10:13:57.016580 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/58f6ad3c-7a20-4be6-8ffb-cadff1c6adef-operator-scripts\") pod \"openstack-galera-0\" (UID: \"58f6ad3c-7a20-4be6-8ffb-cadff1c6adef\") " pod="openstack/openstack-galera-0" Oct 06 10:13:57 crc kubenswrapper[4824]: I1006 10:13:57.016635 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58f6ad3c-7a20-4be6-8ffb-cadff1c6adef-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"58f6ad3c-7a20-4be6-8ffb-cadff1c6adef\") " pod="openstack/openstack-galera-0" Oct 06 10:13:57 crc kubenswrapper[4824]: I1006 10:13:57.016664 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/58f6ad3c-7a20-4be6-8ffb-cadff1c6adef-config-data-default\") pod \"openstack-galera-0\" (UID: \"58f6ad3c-7a20-4be6-8ffb-cadff1c6adef\") " pod="openstack/openstack-galera-0" Oct 06 10:13:57 crc kubenswrapper[4824]: I1006 10:13:57.016682 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1510d553-3505-4bd2-9666-c95cecc43d01-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"1510d553-3505-4bd2-9666-c95cecc43d01\") " pod="openstack/openstack-cell1-galera-0" Oct 06 10:13:57 crc kubenswrapper[4824]: I1006 10:13:57.016709 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/1510d553-3505-4bd2-9666-c95cecc43d01-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"1510d553-3505-4bd2-9666-c95cecc43d01\") " pod="openstack/openstack-cell1-galera-0" Oct 06 10:13:57 crc kubenswrapper[4824]: I1006 10:13:57.016733 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/58f6ad3c-7a20-4be6-8ffb-cadff1c6adef-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"58f6ad3c-7a20-4be6-8ffb-cadff1c6adef\") " pod="openstack/openstack-galera-0" Oct 06 10:13:57 crc kubenswrapper[4824]: I1006 10:13:57.016747 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zpwtq\" (UniqueName: \"kubernetes.io/projected/58f6ad3c-7a20-4be6-8ffb-cadff1c6adef-kube-api-access-zpwtq\") pod \"openstack-galera-0\" (UID: \"58f6ad3c-7a20-4be6-8ffb-cadff1c6adef\") " pod="openstack/openstack-galera-0" Oct 06 10:13:57 crc kubenswrapper[4824]: I1006 10:13:57.016763 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/1510d553-3505-4bd2-9666-c95cecc43d01-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"1510d553-3505-4bd2-9666-c95cecc43d01\") " pod="openstack/openstack-cell1-galera-0" Oct 06 10:13:57 crc kubenswrapper[4824]: I1006 10:13:57.016779 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1510d553-3505-4bd2-9666-c95cecc43d01-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"1510d553-3505-4bd2-9666-c95cecc43d01\") " pod="openstack/openstack-cell1-galera-0" Oct 06 10:13:57 crc kubenswrapper[4824]: I1006 10:13:57.016796 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/1510d553-3505-4bd2-9666-c95cecc43d01-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"1510d553-3505-4bd2-9666-c95cecc43d01\") " pod="openstack/openstack-cell1-galera-0" Oct 06 10:13:57 crc kubenswrapper[4824]: I1006 10:13:57.016819 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/58f6ad3c-7a20-4be6-8ffb-cadff1c6adef-kolla-config\") pod \"openstack-galera-0\" (UID: \"58f6ad3c-7a20-4be6-8ffb-cadff1c6adef\") " pod="openstack/openstack-galera-0" Oct 06 10:13:57 crc kubenswrapper[4824]: I1006 10:13:57.016833 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/1510d553-3505-4bd2-9666-c95cecc43d01-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"1510d553-3505-4bd2-9666-c95cecc43d01\") " pod="openstack/openstack-cell1-galera-0" Oct 06 10:13:57 crc kubenswrapper[4824]: I1006 10:13:57.016854 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/58f6ad3c-7a20-4be6-8ffb-cadff1c6adef-config-data-generated\") pod \"openstack-galera-0\" (UID: \"58f6ad3c-7a20-4be6-8ffb-cadff1c6adef\") " pod="openstack/openstack-galera-0" Oct 06 10:13:57 crc kubenswrapper[4824]: I1006 10:13:57.016874 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"openstack-galera-0\" (UID: \"58f6ad3c-7a20-4be6-8ffb-cadff1c6adef\") " pod="openstack/openstack-galera-0" Oct 06 10:13:57 crc kubenswrapper[4824]: I1006 10:13:57.016901 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z522x\" (UniqueName: \"kubernetes.io/projected/1510d553-3505-4bd2-9666-c95cecc43d01-kube-api-access-z522x\") pod \"openstack-cell1-galera-0\" (UID: \"1510d553-3505-4bd2-9666-c95cecc43d01\") " pod="openstack/openstack-cell1-galera-0" Oct 06 10:13:57 crc kubenswrapper[4824]: I1006 10:13:57.016918 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/58f6ad3c-7a20-4be6-8ffb-cadff1c6adef-secrets\") pod \"openstack-galera-0\" (UID: \"58f6ad3c-7a20-4be6-8ffb-cadff1c6adef\") " pod="openstack/openstack-galera-0" Oct 06 10:13:57 crc kubenswrapper[4824]: I1006 10:13:57.017457 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/58f6ad3c-7a20-4be6-8ffb-cadff1c6adef-config-data-generated\") pod \"openstack-galera-0\" (UID: \"58f6ad3c-7a20-4be6-8ffb-cadff1c6adef\") " pod="openstack/openstack-galera-0" Oct 06 10:13:57 crc kubenswrapper[4824]: I1006 10:13:57.017722 4824 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"openstack-galera-0\" (UID: \"58f6ad3c-7a20-4be6-8ffb-cadff1c6adef\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/openstack-galera-0" Oct 06 10:13:57 crc kubenswrapper[4824]: I1006 10:13:57.018330 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/58f6ad3c-7a20-4be6-8ffb-cadff1c6adef-config-data-default\") pod \"openstack-galera-0\" (UID: \"58f6ad3c-7a20-4be6-8ffb-cadff1c6adef\") " pod="openstack/openstack-galera-0" Oct 06 10:13:57 crc kubenswrapper[4824]: I1006 10:13:57.018423 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/58f6ad3c-7a20-4be6-8ffb-cadff1c6adef-operator-scripts\") pod \"openstack-galera-0\" (UID: \"58f6ad3c-7a20-4be6-8ffb-cadff1c6adef\") " pod="openstack/openstack-galera-0" Oct 06 10:13:57 crc kubenswrapper[4824]: I1006 10:13:57.019037 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/58f6ad3c-7a20-4be6-8ffb-cadff1c6adef-kolla-config\") pod \"openstack-galera-0\" (UID: \"58f6ad3c-7a20-4be6-8ffb-cadff1c6adef\") " pod="openstack/openstack-galera-0" Oct 06 10:13:57 crc kubenswrapper[4824]: I1006 10:13:57.042192 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/58f6ad3c-7a20-4be6-8ffb-cadff1c6adef-secrets\") pod \"openstack-galera-0\" (UID: \"58f6ad3c-7a20-4be6-8ffb-cadff1c6adef\") " pod="openstack/openstack-galera-0" Oct 06 10:13:57 crc kubenswrapper[4824]: I1006 10:13:57.044300 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58f6ad3c-7a20-4be6-8ffb-cadff1c6adef-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"58f6ad3c-7a20-4be6-8ffb-cadff1c6adef\") " pod="openstack/openstack-galera-0" Oct 06 10:13:57 crc kubenswrapper[4824]: I1006 10:13:57.045661 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/58f6ad3c-7a20-4be6-8ffb-cadff1c6adef-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"58f6ad3c-7a20-4be6-8ffb-cadff1c6adef\") " pod="openstack/openstack-galera-0" Oct 06 10:13:57 crc kubenswrapper[4824]: I1006 10:13:57.055442 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"openstack-galera-0\" (UID: \"58f6ad3c-7a20-4be6-8ffb-cadff1c6adef\") " pod="openstack/openstack-galera-0" Oct 06 10:13:57 crc kubenswrapper[4824]: I1006 10:13:57.058396 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zpwtq\" (UniqueName: \"kubernetes.io/projected/58f6ad3c-7a20-4be6-8ffb-cadff1c6adef-kube-api-access-zpwtq\") pod \"openstack-galera-0\" (UID: \"58f6ad3c-7a20-4be6-8ffb-cadff1c6adef\") " pod="openstack/openstack-galera-0" Oct 06 10:13:57 crc kubenswrapper[4824]: I1006 10:13:57.070236 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 06 10:13:57 crc kubenswrapper[4824]: I1006 10:13:57.118031 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1510d553-3505-4bd2-9666-c95cecc43d01-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"1510d553-3505-4bd2-9666-c95cecc43d01\") " pod="openstack/openstack-cell1-galera-0" Oct 06 10:13:57 crc kubenswrapper[4824]: I1006 10:13:57.118103 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/1510d553-3505-4bd2-9666-c95cecc43d01-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"1510d553-3505-4bd2-9666-c95cecc43d01\") " pod="openstack/openstack-cell1-galera-0" Oct 06 10:13:57 crc kubenswrapper[4824]: I1006 10:13:57.118161 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/1510d553-3505-4bd2-9666-c95cecc43d01-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"1510d553-3505-4bd2-9666-c95cecc43d01\") " pod="openstack/openstack-cell1-galera-0" Oct 06 10:13:57 crc kubenswrapper[4824]: I1006 10:13:57.118190 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1510d553-3505-4bd2-9666-c95cecc43d01-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"1510d553-3505-4bd2-9666-c95cecc43d01\") " pod="openstack/openstack-cell1-galera-0" Oct 06 10:13:57 crc kubenswrapper[4824]: I1006 10:13:57.118212 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/1510d553-3505-4bd2-9666-c95cecc43d01-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"1510d553-3505-4bd2-9666-c95cecc43d01\") " pod="openstack/openstack-cell1-galera-0" Oct 06 10:13:57 crc kubenswrapper[4824]: I1006 10:13:57.118277 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/1510d553-3505-4bd2-9666-c95cecc43d01-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"1510d553-3505-4bd2-9666-c95cecc43d01\") " pod="openstack/openstack-cell1-galera-0" Oct 06 10:13:57 crc kubenswrapper[4824]: I1006 10:13:57.118319 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z522x\" (UniqueName: \"kubernetes.io/projected/1510d553-3505-4bd2-9666-c95cecc43d01-kube-api-access-z522x\") pod \"openstack-cell1-galera-0\" (UID: \"1510d553-3505-4bd2-9666-c95cecc43d01\") " pod="openstack/openstack-cell1-galera-0" Oct 06 10:13:57 crc kubenswrapper[4824]: I1006 10:13:57.118368 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/1510d553-3505-4bd2-9666-c95cecc43d01-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"1510d553-3505-4bd2-9666-c95cecc43d01\") " pod="openstack/openstack-cell1-galera-0" Oct 06 10:13:57 crc kubenswrapper[4824]: I1006 10:13:57.118395 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-cell1-galera-0\" (UID: \"1510d553-3505-4bd2-9666-c95cecc43d01\") " pod="openstack/openstack-cell1-galera-0" Oct 06 10:13:57 crc kubenswrapper[4824]: I1006 10:13:57.118574 4824 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-cell1-galera-0\" (UID: \"1510d553-3505-4bd2-9666-c95cecc43d01\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/openstack-cell1-galera-0" Oct 06 10:13:57 crc kubenswrapper[4824]: I1006 10:13:57.120704 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/1510d553-3505-4bd2-9666-c95cecc43d01-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"1510d553-3505-4bd2-9666-c95cecc43d01\") " pod="openstack/openstack-cell1-galera-0" Oct 06 10:13:57 crc kubenswrapper[4824]: I1006 10:13:57.121027 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/1510d553-3505-4bd2-9666-c95cecc43d01-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"1510d553-3505-4bd2-9666-c95cecc43d01\") " pod="openstack/openstack-cell1-galera-0" Oct 06 10:13:57 crc kubenswrapper[4824]: I1006 10:13:57.121732 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1510d553-3505-4bd2-9666-c95cecc43d01-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"1510d553-3505-4bd2-9666-c95cecc43d01\") " pod="openstack/openstack-cell1-galera-0" Oct 06 10:13:57 crc kubenswrapper[4824]: I1006 10:13:57.123154 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/1510d553-3505-4bd2-9666-c95cecc43d01-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"1510d553-3505-4bd2-9666-c95cecc43d01\") " pod="openstack/openstack-cell1-galera-0" Oct 06 10:13:57 crc kubenswrapper[4824]: I1006 10:13:57.127795 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/1510d553-3505-4bd2-9666-c95cecc43d01-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"1510d553-3505-4bd2-9666-c95cecc43d01\") " pod="openstack/openstack-cell1-galera-0" Oct 06 10:13:57 crc kubenswrapper[4824]: I1006 10:13:57.128328 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1510d553-3505-4bd2-9666-c95cecc43d01-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"1510d553-3505-4bd2-9666-c95cecc43d01\") " pod="openstack/openstack-cell1-galera-0" Oct 06 10:13:57 crc kubenswrapper[4824]: I1006 10:13:57.128424 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/1510d553-3505-4bd2-9666-c95cecc43d01-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"1510d553-3505-4bd2-9666-c95cecc43d01\") " pod="openstack/openstack-cell1-galera-0" Oct 06 10:13:57 crc kubenswrapper[4824]: I1006 10:13:57.147705 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z522x\" (UniqueName: \"kubernetes.io/projected/1510d553-3505-4bd2-9666-c95cecc43d01-kube-api-access-z522x\") pod \"openstack-cell1-galera-0\" (UID: \"1510d553-3505-4bd2-9666-c95cecc43d01\") " pod="openstack/openstack-cell1-galera-0" Oct 06 10:13:57 crc kubenswrapper[4824]: I1006 10:13:57.224017 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-cell1-galera-0\" (UID: \"1510d553-3505-4bd2-9666-c95cecc43d01\") " pod="openstack/openstack-cell1-galera-0" Oct 06 10:13:57 crc kubenswrapper[4824]: I1006 10:13:57.347682 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Oct 06 10:13:57 crc kubenswrapper[4824]: I1006 10:13:57.348646 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 06 10:13:57 crc kubenswrapper[4824]: I1006 10:13:57.353192 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Oct 06 10:13:57 crc kubenswrapper[4824]: I1006 10:13:57.357636 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-wmmjk" Oct 06 10:13:57 crc kubenswrapper[4824]: I1006 10:13:57.360170 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Oct 06 10:13:57 crc kubenswrapper[4824]: I1006 10:13:57.363121 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Oct 06 10:13:57 crc kubenswrapper[4824]: I1006 10:13:57.423734 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3a12653c-c0d0-46e3-b909-77bd2a7f6c4a-config-data\") pod \"memcached-0\" (UID: \"3a12653c-c0d0-46e3-b909-77bd2a7f6c4a\") " pod="openstack/memcached-0" Oct 06 10:13:57 crc kubenswrapper[4824]: I1006 10:13:57.423789 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/3a12653c-c0d0-46e3-b909-77bd2a7f6c4a-memcached-tls-certs\") pod \"memcached-0\" (UID: \"3a12653c-c0d0-46e3-b909-77bd2a7f6c4a\") " pod="openstack/memcached-0" Oct 06 10:13:57 crc kubenswrapper[4824]: I1006 10:13:57.423814 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g9pzx\" (UniqueName: \"kubernetes.io/projected/3a12653c-c0d0-46e3-b909-77bd2a7f6c4a-kube-api-access-g9pzx\") pod \"memcached-0\" (UID: \"3a12653c-c0d0-46e3-b909-77bd2a7f6c4a\") " pod="openstack/memcached-0" Oct 06 10:13:57 crc kubenswrapper[4824]: I1006 10:13:57.423853 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a12653c-c0d0-46e3-b909-77bd2a7f6c4a-combined-ca-bundle\") pod \"memcached-0\" (UID: \"3a12653c-c0d0-46e3-b909-77bd2a7f6c4a\") " pod="openstack/memcached-0" Oct 06 10:13:57 crc kubenswrapper[4824]: I1006 10:13:57.423898 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/3a12653c-c0d0-46e3-b909-77bd2a7f6c4a-kolla-config\") pod \"memcached-0\" (UID: \"3a12653c-c0d0-46e3-b909-77bd2a7f6c4a\") " pod="openstack/memcached-0" Oct 06 10:13:57 crc kubenswrapper[4824]: I1006 10:13:57.487479 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 06 10:13:57 crc kubenswrapper[4824]: I1006 10:13:57.525798 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a12653c-c0d0-46e3-b909-77bd2a7f6c4a-combined-ca-bundle\") pod \"memcached-0\" (UID: \"3a12653c-c0d0-46e3-b909-77bd2a7f6c4a\") " pod="openstack/memcached-0" Oct 06 10:13:57 crc kubenswrapper[4824]: I1006 10:13:57.525870 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/3a12653c-c0d0-46e3-b909-77bd2a7f6c4a-kolla-config\") pod \"memcached-0\" (UID: \"3a12653c-c0d0-46e3-b909-77bd2a7f6c4a\") " pod="openstack/memcached-0" Oct 06 10:13:57 crc kubenswrapper[4824]: I1006 10:13:57.525930 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3a12653c-c0d0-46e3-b909-77bd2a7f6c4a-config-data\") pod \"memcached-0\" (UID: \"3a12653c-c0d0-46e3-b909-77bd2a7f6c4a\") " pod="openstack/memcached-0" Oct 06 10:13:57 crc kubenswrapper[4824]: I1006 10:13:57.525955 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/3a12653c-c0d0-46e3-b909-77bd2a7f6c4a-memcached-tls-certs\") pod \"memcached-0\" (UID: \"3a12653c-c0d0-46e3-b909-77bd2a7f6c4a\") " pod="openstack/memcached-0" Oct 06 10:13:57 crc kubenswrapper[4824]: I1006 10:13:57.525978 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g9pzx\" (UniqueName: \"kubernetes.io/projected/3a12653c-c0d0-46e3-b909-77bd2a7f6c4a-kube-api-access-g9pzx\") pod \"memcached-0\" (UID: \"3a12653c-c0d0-46e3-b909-77bd2a7f6c4a\") " pod="openstack/memcached-0" Oct 06 10:13:57 crc kubenswrapper[4824]: I1006 10:13:57.527402 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3a12653c-c0d0-46e3-b909-77bd2a7f6c4a-config-data\") pod \"memcached-0\" (UID: \"3a12653c-c0d0-46e3-b909-77bd2a7f6c4a\") " pod="openstack/memcached-0" Oct 06 10:13:57 crc kubenswrapper[4824]: I1006 10:13:57.527665 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/3a12653c-c0d0-46e3-b909-77bd2a7f6c4a-kolla-config\") pod \"memcached-0\" (UID: \"3a12653c-c0d0-46e3-b909-77bd2a7f6c4a\") " pod="openstack/memcached-0" Oct 06 10:13:57 crc kubenswrapper[4824]: I1006 10:13:57.531386 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/3a12653c-c0d0-46e3-b909-77bd2a7f6c4a-memcached-tls-certs\") pod \"memcached-0\" (UID: \"3a12653c-c0d0-46e3-b909-77bd2a7f6c4a\") " pod="openstack/memcached-0" Oct 06 10:13:57 crc kubenswrapper[4824]: I1006 10:13:57.531707 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a12653c-c0d0-46e3-b909-77bd2a7f6c4a-combined-ca-bundle\") pod \"memcached-0\" (UID: \"3a12653c-c0d0-46e3-b909-77bd2a7f6c4a\") " pod="openstack/memcached-0" Oct 06 10:13:57 crc kubenswrapper[4824]: I1006 10:13:57.540939 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g9pzx\" (UniqueName: \"kubernetes.io/projected/3a12653c-c0d0-46e3-b909-77bd2a7f6c4a-kube-api-access-g9pzx\") pod \"memcached-0\" (UID: \"3a12653c-c0d0-46e3-b909-77bd2a7f6c4a\") " pod="openstack/memcached-0" Oct 06 10:13:57 crc kubenswrapper[4824]: I1006 10:13:57.673433 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 06 10:13:59 crc kubenswrapper[4824]: W1006 10:13:59.004660 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc8dbad1e_02be_43c8_8f80_ab771ee81742.slice/crio-96dcce3401c3ff21f9fd96a662d06af63598676c2a64ef986e5f4bb7c1d4315d WatchSource:0}: Error finding container 96dcce3401c3ff21f9fd96a662d06af63598676c2a64ef986e5f4bb7c1d4315d: Status 404 returned error can't find the container with id 96dcce3401c3ff21f9fd96a662d06af63598676c2a64ef986e5f4bb7c1d4315d Oct 06 10:13:59 crc kubenswrapper[4824]: I1006 10:13:59.059726 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Oct 06 10:13:59 crc kubenswrapper[4824]: I1006 10:13:59.060944 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 06 10:13:59 crc kubenswrapper[4824]: I1006 10:13:59.065932 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-7gmqh" Oct 06 10:13:59 crc kubenswrapper[4824]: I1006 10:13:59.073695 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 06 10:13:59 crc kubenswrapper[4824]: I1006 10:13:59.147442 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"f7d047da-89e3-4cd4-bf02-e6a3e15c585f","Type":"ContainerStarted","Data":"ae8f6eccb88670a7b087c7a985bb0482f09a525c27e50b6cba6d9b58e14310d6"} Oct 06 10:13:59 crc kubenswrapper[4824]: I1006 10:13:59.149387 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"c8dbad1e-02be-43c8-8f80-ab771ee81742","Type":"ContainerStarted","Data":"96dcce3401c3ff21f9fd96a662d06af63598676c2a64ef986e5f4bb7c1d4315d"} Oct 06 10:13:59 crc kubenswrapper[4824]: I1006 10:13:59.162002 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mplf5\" (UniqueName: \"kubernetes.io/projected/1d365fcf-17fb-411e-ab2c-3c5dc8384b21-kube-api-access-mplf5\") pod \"kube-state-metrics-0\" (UID: \"1d365fcf-17fb-411e-ab2c-3c5dc8384b21\") " pod="openstack/kube-state-metrics-0" Oct 06 10:13:59 crc kubenswrapper[4824]: I1006 10:13:59.263507 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mplf5\" (UniqueName: \"kubernetes.io/projected/1d365fcf-17fb-411e-ab2c-3c5dc8384b21-kube-api-access-mplf5\") pod \"kube-state-metrics-0\" (UID: \"1d365fcf-17fb-411e-ab2c-3c5dc8384b21\") " pod="openstack/kube-state-metrics-0" Oct 06 10:13:59 crc kubenswrapper[4824]: I1006 10:13:59.285612 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mplf5\" (UniqueName: \"kubernetes.io/projected/1d365fcf-17fb-411e-ab2c-3c5dc8384b21-kube-api-access-mplf5\") pod \"kube-state-metrics-0\" (UID: \"1d365fcf-17fb-411e-ab2c-3c5dc8384b21\") " pod="openstack/kube-state-metrics-0" Oct 06 10:13:59 crc kubenswrapper[4824]: I1006 10:13:59.397570 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 06 10:14:02 crc kubenswrapper[4824]: I1006 10:14:02.498599 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-48qsh"] Oct 06 10:14:02 crc kubenswrapper[4824]: I1006 10:14:02.504422 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-48qsh" Oct 06 10:14:02 crc kubenswrapper[4824]: I1006 10:14:02.508661 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-l2vk4" Oct 06 10:14:02 crc kubenswrapper[4824]: I1006 10:14:02.509533 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Oct 06 10:14:02 crc kubenswrapper[4824]: I1006 10:14:02.511357 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Oct 06 10:14:02 crc kubenswrapper[4824]: I1006 10:14:02.536657 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-48qsh"] Oct 06 10:14:02 crc kubenswrapper[4824]: I1006 10:14:02.566982 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-v22g5"] Oct 06 10:14:02 crc kubenswrapper[4824]: I1006 10:14:02.569367 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-v22g5" Oct 06 10:14:02 crc kubenswrapper[4824]: I1006 10:14:02.589852 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-v22g5"] Oct 06 10:14:02 crc kubenswrapper[4824]: I1006 10:14:02.617342 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/03d76a8a-df87-4a48-8fb9-e6a502b37ae8-var-run\") pod \"ovn-controller-48qsh\" (UID: \"03d76a8a-df87-4a48-8fb9-e6a502b37ae8\") " pod="openstack/ovn-controller-48qsh" Oct 06 10:14:02 crc kubenswrapper[4824]: I1006 10:14:02.617390 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/03d76a8a-df87-4a48-8fb9-e6a502b37ae8-scripts\") pod \"ovn-controller-48qsh\" (UID: \"03d76a8a-df87-4a48-8fb9-e6a502b37ae8\") " pod="openstack/ovn-controller-48qsh" Oct 06 10:14:02 crc kubenswrapper[4824]: I1006 10:14:02.617413 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dk685\" (UniqueName: \"kubernetes.io/projected/59148e21-e79a-466f-bed5-24671942c24f-kube-api-access-dk685\") pod \"ovn-controller-ovs-v22g5\" (UID: \"59148e21-e79a-466f-bed5-24671942c24f\") " pod="openstack/ovn-controller-ovs-v22g5" Oct 06 10:14:02 crc kubenswrapper[4824]: I1006 10:14:02.617445 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/59148e21-e79a-466f-bed5-24671942c24f-scripts\") pod \"ovn-controller-ovs-v22g5\" (UID: \"59148e21-e79a-466f-bed5-24671942c24f\") " pod="openstack/ovn-controller-ovs-v22g5" Oct 06 10:14:02 crc kubenswrapper[4824]: I1006 10:14:02.617466 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/03d76a8a-df87-4a48-8fb9-e6a502b37ae8-var-log-ovn\") pod \"ovn-controller-48qsh\" (UID: \"03d76a8a-df87-4a48-8fb9-e6a502b37ae8\") " pod="openstack/ovn-controller-48qsh" Oct 06 10:14:02 crc kubenswrapper[4824]: I1006 10:14:02.617602 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/03d76a8a-df87-4a48-8fb9-e6a502b37ae8-var-run-ovn\") pod \"ovn-controller-48qsh\" (UID: \"03d76a8a-df87-4a48-8fb9-e6a502b37ae8\") " pod="openstack/ovn-controller-48qsh" Oct 06 10:14:02 crc kubenswrapper[4824]: I1006 10:14:02.617704 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/59148e21-e79a-466f-bed5-24671942c24f-var-lib\") pod \"ovn-controller-ovs-v22g5\" (UID: \"59148e21-e79a-466f-bed5-24671942c24f\") " pod="openstack/ovn-controller-ovs-v22g5" Oct 06 10:14:02 crc kubenswrapper[4824]: I1006 10:14:02.617802 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03d76a8a-df87-4a48-8fb9-e6a502b37ae8-combined-ca-bundle\") pod \"ovn-controller-48qsh\" (UID: \"03d76a8a-df87-4a48-8fb9-e6a502b37ae8\") " pod="openstack/ovn-controller-48qsh" Oct 06 10:14:02 crc kubenswrapper[4824]: I1006 10:14:02.617879 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/03d76a8a-df87-4a48-8fb9-e6a502b37ae8-ovn-controller-tls-certs\") pod \"ovn-controller-48qsh\" (UID: \"03d76a8a-df87-4a48-8fb9-e6a502b37ae8\") " pod="openstack/ovn-controller-48qsh" Oct 06 10:14:02 crc kubenswrapper[4824]: I1006 10:14:02.617916 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/59148e21-e79a-466f-bed5-24671942c24f-var-log\") pod \"ovn-controller-ovs-v22g5\" (UID: \"59148e21-e79a-466f-bed5-24671942c24f\") " pod="openstack/ovn-controller-ovs-v22g5" Oct 06 10:14:02 crc kubenswrapper[4824]: I1006 10:14:02.617946 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bnn5t\" (UniqueName: \"kubernetes.io/projected/03d76a8a-df87-4a48-8fb9-e6a502b37ae8-kube-api-access-bnn5t\") pod \"ovn-controller-48qsh\" (UID: \"03d76a8a-df87-4a48-8fb9-e6a502b37ae8\") " pod="openstack/ovn-controller-48qsh" Oct 06 10:14:02 crc kubenswrapper[4824]: I1006 10:14:02.617998 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/59148e21-e79a-466f-bed5-24671942c24f-etc-ovs\") pod \"ovn-controller-ovs-v22g5\" (UID: \"59148e21-e79a-466f-bed5-24671942c24f\") " pod="openstack/ovn-controller-ovs-v22g5" Oct 06 10:14:02 crc kubenswrapper[4824]: I1006 10:14:02.618076 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/59148e21-e79a-466f-bed5-24671942c24f-var-run\") pod \"ovn-controller-ovs-v22g5\" (UID: \"59148e21-e79a-466f-bed5-24671942c24f\") " pod="openstack/ovn-controller-ovs-v22g5" Oct 06 10:14:02 crc kubenswrapper[4824]: I1006 10:14:02.720866 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03d76a8a-df87-4a48-8fb9-e6a502b37ae8-combined-ca-bundle\") pod \"ovn-controller-48qsh\" (UID: \"03d76a8a-df87-4a48-8fb9-e6a502b37ae8\") " pod="openstack/ovn-controller-48qsh" Oct 06 10:14:02 crc kubenswrapper[4824]: I1006 10:14:02.720927 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/03d76a8a-df87-4a48-8fb9-e6a502b37ae8-ovn-controller-tls-certs\") pod \"ovn-controller-48qsh\" (UID: \"03d76a8a-df87-4a48-8fb9-e6a502b37ae8\") " pod="openstack/ovn-controller-48qsh" Oct 06 10:14:02 crc kubenswrapper[4824]: I1006 10:14:02.720949 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/59148e21-e79a-466f-bed5-24671942c24f-var-log\") pod \"ovn-controller-ovs-v22g5\" (UID: \"59148e21-e79a-466f-bed5-24671942c24f\") " pod="openstack/ovn-controller-ovs-v22g5" Oct 06 10:14:02 crc kubenswrapper[4824]: I1006 10:14:02.720972 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bnn5t\" (UniqueName: \"kubernetes.io/projected/03d76a8a-df87-4a48-8fb9-e6a502b37ae8-kube-api-access-bnn5t\") pod \"ovn-controller-48qsh\" (UID: \"03d76a8a-df87-4a48-8fb9-e6a502b37ae8\") " pod="openstack/ovn-controller-48qsh" Oct 06 10:14:02 crc kubenswrapper[4824]: I1006 10:14:02.721008 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/59148e21-e79a-466f-bed5-24671942c24f-etc-ovs\") pod \"ovn-controller-ovs-v22g5\" (UID: \"59148e21-e79a-466f-bed5-24671942c24f\") " pod="openstack/ovn-controller-ovs-v22g5" Oct 06 10:14:02 crc kubenswrapper[4824]: I1006 10:14:02.721043 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/59148e21-e79a-466f-bed5-24671942c24f-var-run\") pod \"ovn-controller-ovs-v22g5\" (UID: \"59148e21-e79a-466f-bed5-24671942c24f\") " pod="openstack/ovn-controller-ovs-v22g5" Oct 06 10:14:02 crc kubenswrapper[4824]: I1006 10:14:02.721082 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/03d76a8a-df87-4a48-8fb9-e6a502b37ae8-var-run\") pod \"ovn-controller-48qsh\" (UID: \"03d76a8a-df87-4a48-8fb9-e6a502b37ae8\") " pod="openstack/ovn-controller-48qsh" Oct 06 10:14:02 crc kubenswrapper[4824]: I1006 10:14:02.721097 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/03d76a8a-df87-4a48-8fb9-e6a502b37ae8-scripts\") pod \"ovn-controller-48qsh\" (UID: \"03d76a8a-df87-4a48-8fb9-e6a502b37ae8\") " pod="openstack/ovn-controller-48qsh" Oct 06 10:14:02 crc kubenswrapper[4824]: I1006 10:14:02.721114 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dk685\" (UniqueName: \"kubernetes.io/projected/59148e21-e79a-466f-bed5-24671942c24f-kube-api-access-dk685\") pod \"ovn-controller-ovs-v22g5\" (UID: \"59148e21-e79a-466f-bed5-24671942c24f\") " pod="openstack/ovn-controller-ovs-v22g5" Oct 06 10:14:02 crc kubenswrapper[4824]: I1006 10:14:02.721138 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/59148e21-e79a-466f-bed5-24671942c24f-scripts\") pod \"ovn-controller-ovs-v22g5\" (UID: \"59148e21-e79a-466f-bed5-24671942c24f\") " pod="openstack/ovn-controller-ovs-v22g5" Oct 06 10:14:02 crc kubenswrapper[4824]: I1006 10:14:02.721156 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/03d76a8a-df87-4a48-8fb9-e6a502b37ae8-var-log-ovn\") pod \"ovn-controller-48qsh\" (UID: \"03d76a8a-df87-4a48-8fb9-e6a502b37ae8\") " pod="openstack/ovn-controller-48qsh" Oct 06 10:14:02 crc kubenswrapper[4824]: I1006 10:14:02.721184 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/03d76a8a-df87-4a48-8fb9-e6a502b37ae8-var-run-ovn\") pod \"ovn-controller-48qsh\" (UID: \"03d76a8a-df87-4a48-8fb9-e6a502b37ae8\") " pod="openstack/ovn-controller-48qsh" Oct 06 10:14:02 crc kubenswrapper[4824]: I1006 10:14:02.721213 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/59148e21-e79a-466f-bed5-24671942c24f-var-lib\") pod \"ovn-controller-ovs-v22g5\" (UID: \"59148e21-e79a-466f-bed5-24671942c24f\") " pod="openstack/ovn-controller-ovs-v22g5" Oct 06 10:14:02 crc kubenswrapper[4824]: I1006 10:14:02.721621 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/59148e21-e79a-466f-bed5-24671942c24f-var-log\") pod \"ovn-controller-ovs-v22g5\" (UID: \"59148e21-e79a-466f-bed5-24671942c24f\") " pod="openstack/ovn-controller-ovs-v22g5" Oct 06 10:14:02 crc kubenswrapper[4824]: I1006 10:14:02.721677 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/59148e21-e79a-466f-bed5-24671942c24f-var-lib\") pod \"ovn-controller-ovs-v22g5\" (UID: \"59148e21-e79a-466f-bed5-24671942c24f\") " pod="openstack/ovn-controller-ovs-v22g5" Oct 06 10:14:02 crc kubenswrapper[4824]: I1006 10:14:02.721901 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/03d76a8a-df87-4a48-8fb9-e6a502b37ae8-var-run\") pod \"ovn-controller-48qsh\" (UID: \"03d76a8a-df87-4a48-8fb9-e6a502b37ae8\") " pod="openstack/ovn-controller-48qsh" Oct 06 10:14:02 crc kubenswrapper[4824]: I1006 10:14:02.721929 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/59148e21-e79a-466f-bed5-24671942c24f-var-run\") pod \"ovn-controller-ovs-v22g5\" (UID: \"59148e21-e79a-466f-bed5-24671942c24f\") " pod="openstack/ovn-controller-ovs-v22g5" Oct 06 10:14:02 crc kubenswrapper[4824]: I1006 10:14:02.722036 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/03d76a8a-df87-4a48-8fb9-e6a502b37ae8-var-log-ovn\") pod \"ovn-controller-48qsh\" (UID: \"03d76a8a-df87-4a48-8fb9-e6a502b37ae8\") " pod="openstack/ovn-controller-48qsh" Oct 06 10:14:02 crc kubenswrapper[4824]: I1006 10:14:02.722110 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/03d76a8a-df87-4a48-8fb9-e6a502b37ae8-var-run-ovn\") pod \"ovn-controller-48qsh\" (UID: \"03d76a8a-df87-4a48-8fb9-e6a502b37ae8\") " pod="openstack/ovn-controller-48qsh" Oct 06 10:14:02 crc kubenswrapper[4824]: I1006 10:14:02.725957 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/59148e21-e79a-466f-bed5-24671942c24f-scripts\") pod \"ovn-controller-ovs-v22g5\" (UID: \"59148e21-e79a-466f-bed5-24671942c24f\") " pod="openstack/ovn-controller-ovs-v22g5" Oct 06 10:14:02 crc kubenswrapper[4824]: I1006 10:14:02.727397 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/03d76a8a-df87-4a48-8fb9-e6a502b37ae8-scripts\") pod \"ovn-controller-48qsh\" (UID: \"03d76a8a-df87-4a48-8fb9-e6a502b37ae8\") " pod="openstack/ovn-controller-48qsh" Oct 06 10:14:02 crc kubenswrapper[4824]: I1006 10:14:02.727598 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/59148e21-e79a-466f-bed5-24671942c24f-etc-ovs\") pod \"ovn-controller-ovs-v22g5\" (UID: \"59148e21-e79a-466f-bed5-24671942c24f\") " pod="openstack/ovn-controller-ovs-v22g5" Oct 06 10:14:02 crc kubenswrapper[4824]: I1006 10:14:02.728188 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/03d76a8a-df87-4a48-8fb9-e6a502b37ae8-ovn-controller-tls-certs\") pod \"ovn-controller-48qsh\" (UID: \"03d76a8a-df87-4a48-8fb9-e6a502b37ae8\") " pod="openstack/ovn-controller-48qsh" Oct 06 10:14:02 crc kubenswrapper[4824]: I1006 10:14:02.730464 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03d76a8a-df87-4a48-8fb9-e6a502b37ae8-combined-ca-bundle\") pod \"ovn-controller-48qsh\" (UID: \"03d76a8a-df87-4a48-8fb9-e6a502b37ae8\") " pod="openstack/ovn-controller-48qsh" Oct 06 10:14:02 crc kubenswrapper[4824]: I1006 10:14:02.738889 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bnn5t\" (UniqueName: \"kubernetes.io/projected/03d76a8a-df87-4a48-8fb9-e6a502b37ae8-kube-api-access-bnn5t\") pod \"ovn-controller-48qsh\" (UID: \"03d76a8a-df87-4a48-8fb9-e6a502b37ae8\") " pod="openstack/ovn-controller-48qsh" Oct 06 10:14:02 crc kubenswrapper[4824]: I1006 10:14:02.745413 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dk685\" (UniqueName: \"kubernetes.io/projected/59148e21-e79a-466f-bed5-24671942c24f-kube-api-access-dk685\") pod \"ovn-controller-ovs-v22g5\" (UID: \"59148e21-e79a-466f-bed5-24671942c24f\") " pod="openstack/ovn-controller-ovs-v22g5" Oct 06 10:14:02 crc kubenswrapper[4824]: I1006 10:14:02.873094 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-48qsh" Oct 06 10:14:02 crc kubenswrapper[4824]: I1006 10:14:02.889227 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-v22g5" Oct 06 10:14:03 crc kubenswrapper[4824]: I1006 10:14:03.387374 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 06 10:14:03 crc kubenswrapper[4824]: I1006 10:14:03.398580 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 06 10:14:03 crc kubenswrapper[4824]: I1006 10:14:03.398708 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 06 10:14:03 crc kubenswrapper[4824]: I1006 10:14:03.401041 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Oct 06 10:14:03 crc kubenswrapper[4824]: I1006 10:14:03.401317 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Oct 06 10:14:03 crc kubenswrapper[4824]: I1006 10:14:03.401461 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Oct 06 10:14:03 crc kubenswrapper[4824]: I1006 10:14:03.403047 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-wnq9r" Oct 06 10:14:03 crc kubenswrapper[4824]: I1006 10:14:03.403174 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Oct 06 10:14:03 crc kubenswrapper[4824]: I1006 10:14:03.433209 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a39bac9c-ff87-4d4a-ad96-fa5618ca4e40-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"a39bac9c-ff87-4d4a-ad96-fa5618ca4e40\") " pod="openstack/ovsdbserver-nb-0" Oct 06 10:14:03 crc kubenswrapper[4824]: I1006 10:14:03.433259 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/a39bac9c-ff87-4d4a-ad96-fa5618ca4e40-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"a39bac9c-ff87-4d4a-ad96-fa5618ca4e40\") " pod="openstack/ovsdbserver-nb-0" Oct 06 10:14:03 crc kubenswrapper[4824]: I1006 10:14:03.433297 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/a39bac9c-ff87-4d4a-ad96-fa5618ca4e40-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"a39bac9c-ff87-4d4a-ad96-fa5618ca4e40\") " pod="openstack/ovsdbserver-nb-0" Oct 06 10:14:03 crc kubenswrapper[4824]: I1006 10:14:03.433329 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a39bac9c-ff87-4d4a-ad96-fa5618ca4e40-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"a39bac9c-ff87-4d4a-ad96-fa5618ca4e40\") " pod="openstack/ovsdbserver-nb-0" Oct 06 10:14:03 crc kubenswrapper[4824]: I1006 10:14:03.433343 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a39bac9c-ff87-4d4a-ad96-fa5618ca4e40-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"a39bac9c-ff87-4d4a-ad96-fa5618ca4e40\") " pod="openstack/ovsdbserver-nb-0" Oct 06 10:14:03 crc kubenswrapper[4824]: I1006 10:14:03.433386 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a39bac9c-ff87-4d4a-ad96-fa5618ca4e40-config\") pod \"ovsdbserver-nb-0\" (UID: \"a39bac9c-ff87-4d4a-ad96-fa5618ca4e40\") " pod="openstack/ovsdbserver-nb-0" Oct 06 10:14:03 crc kubenswrapper[4824]: I1006 10:14:03.433425 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cncj2\" (UniqueName: \"kubernetes.io/projected/a39bac9c-ff87-4d4a-ad96-fa5618ca4e40-kube-api-access-cncj2\") pod \"ovsdbserver-nb-0\" (UID: \"a39bac9c-ff87-4d4a-ad96-fa5618ca4e40\") " pod="openstack/ovsdbserver-nb-0" Oct 06 10:14:03 crc kubenswrapper[4824]: I1006 10:14:03.433443 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"ovsdbserver-nb-0\" (UID: \"a39bac9c-ff87-4d4a-ad96-fa5618ca4e40\") " pod="openstack/ovsdbserver-nb-0" Oct 06 10:14:03 crc kubenswrapper[4824]: I1006 10:14:03.535260 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a39bac9c-ff87-4d4a-ad96-fa5618ca4e40-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"a39bac9c-ff87-4d4a-ad96-fa5618ca4e40\") " pod="openstack/ovsdbserver-nb-0" Oct 06 10:14:03 crc kubenswrapper[4824]: I1006 10:14:03.535301 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a39bac9c-ff87-4d4a-ad96-fa5618ca4e40-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"a39bac9c-ff87-4d4a-ad96-fa5618ca4e40\") " pod="openstack/ovsdbserver-nb-0" Oct 06 10:14:03 crc kubenswrapper[4824]: I1006 10:14:03.535364 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a39bac9c-ff87-4d4a-ad96-fa5618ca4e40-config\") pod \"ovsdbserver-nb-0\" (UID: \"a39bac9c-ff87-4d4a-ad96-fa5618ca4e40\") " pod="openstack/ovsdbserver-nb-0" Oct 06 10:14:03 crc kubenswrapper[4824]: I1006 10:14:03.535410 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cncj2\" (UniqueName: \"kubernetes.io/projected/a39bac9c-ff87-4d4a-ad96-fa5618ca4e40-kube-api-access-cncj2\") pod \"ovsdbserver-nb-0\" (UID: \"a39bac9c-ff87-4d4a-ad96-fa5618ca4e40\") " pod="openstack/ovsdbserver-nb-0" Oct 06 10:14:03 crc kubenswrapper[4824]: I1006 10:14:03.535430 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"ovsdbserver-nb-0\" (UID: \"a39bac9c-ff87-4d4a-ad96-fa5618ca4e40\") " pod="openstack/ovsdbserver-nb-0" Oct 06 10:14:03 crc kubenswrapper[4824]: I1006 10:14:03.535462 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a39bac9c-ff87-4d4a-ad96-fa5618ca4e40-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"a39bac9c-ff87-4d4a-ad96-fa5618ca4e40\") " pod="openstack/ovsdbserver-nb-0" Oct 06 10:14:03 crc kubenswrapper[4824]: I1006 10:14:03.535491 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/a39bac9c-ff87-4d4a-ad96-fa5618ca4e40-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"a39bac9c-ff87-4d4a-ad96-fa5618ca4e40\") " pod="openstack/ovsdbserver-nb-0" Oct 06 10:14:03 crc kubenswrapper[4824]: I1006 10:14:03.535524 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/a39bac9c-ff87-4d4a-ad96-fa5618ca4e40-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"a39bac9c-ff87-4d4a-ad96-fa5618ca4e40\") " pod="openstack/ovsdbserver-nb-0" Oct 06 10:14:03 crc kubenswrapper[4824]: I1006 10:14:03.536803 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a39bac9c-ff87-4d4a-ad96-fa5618ca4e40-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"a39bac9c-ff87-4d4a-ad96-fa5618ca4e40\") " pod="openstack/ovsdbserver-nb-0" Oct 06 10:14:03 crc kubenswrapper[4824]: I1006 10:14:03.536964 4824 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"ovsdbserver-nb-0\" (UID: \"a39bac9c-ff87-4d4a-ad96-fa5618ca4e40\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/ovsdbserver-nb-0" Oct 06 10:14:03 crc kubenswrapper[4824]: I1006 10:14:03.537991 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a39bac9c-ff87-4d4a-ad96-fa5618ca4e40-config\") pod \"ovsdbserver-nb-0\" (UID: \"a39bac9c-ff87-4d4a-ad96-fa5618ca4e40\") " pod="openstack/ovsdbserver-nb-0" Oct 06 10:14:03 crc kubenswrapper[4824]: I1006 10:14:03.538643 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/a39bac9c-ff87-4d4a-ad96-fa5618ca4e40-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"a39bac9c-ff87-4d4a-ad96-fa5618ca4e40\") " pod="openstack/ovsdbserver-nb-0" Oct 06 10:14:03 crc kubenswrapper[4824]: I1006 10:14:03.540773 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/a39bac9c-ff87-4d4a-ad96-fa5618ca4e40-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"a39bac9c-ff87-4d4a-ad96-fa5618ca4e40\") " pod="openstack/ovsdbserver-nb-0" Oct 06 10:14:03 crc kubenswrapper[4824]: I1006 10:14:03.544203 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a39bac9c-ff87-4d4a-ad96-fa5618ca4e40-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"a39bac9c-ff87-4d4a-ad96-fa5618ca4e40\") " pod="openstack/ovsdbserver-nb-0" Oct 06 10:14:03 crc kubenswrapper[4824]: I1006 10:14:03.545423 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a39bac9c-ff87-4d4a-ad96-fa5618ca4e40-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"a39bac9c-ff87-4d4a-ad96-fa5618ca4e40\") " pod="openstack/ovsdbserver-nb-0" Oct 06 10:14:03 crc kubenswrapper[4824]: I1006 10:14:03.553887 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cncj2\" (UniqueName: \"kubernetes.io/projected/a39bac9c-ff87-4d4a-ad96-fa5618ca4e40-kube-api-access-cncj2\") pod \"ovsdbserver-nb-0\" (UID: \"a39bac9c-ff87-4d4a-ad96-fa5618ca4e40\") " pod="openstack/ovsdbserver-nb-0" Oct 06 10:14:03 crc kubenswrapper[4824]: I1006 10:14:03.563805 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"ovsdbserver-nb-0\" (UID: \"a39bac9c-ff87-4d4a-ad96-fa5618ca4e40\") " pod="openstack/ovsdbserver-nb-0" Oct 06 10:14:03 crc kubenswrapper[4824]: I1006 10:14:03.723619 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 06 10:14:06 crc kubenswrapper[4824]: I1006 10:14:06.835792 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 06 10:14:06 crc kubenswrapper[4824]: I1006 10:14:06.838674 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 06 10:14:06 crc kubenswrapper[4824]: I1006 10:14:06.846843 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Oct 06 10:14:06 crc kubenswrapper[4824]: I1006 10:14:06.847290 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-kfzgr" Oct 06 10:14:06 crc kubenswrapper[4824]: I1006 10:14:06.847357 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Oct 06 10:14:06 crc kubenswrapper[4824]: I1006 10:14:06.847394 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Oct 06 10:14:06 crc kubenswrapper[4824]: I1006 10:14:06.867692 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 06 10:14:06 crc kubenswrapper[4824]: I1006 10:14:06.900373 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/b6303c51-6db3-4ab6-aeab-edf39e88fdec-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"b6303c51-6db3-4ab6-aeab-edf39e88fdec\") " pod="openstack/ovsdbserver-sb-0" Oct 06 10:14:06 crc kubenswrapper[4824]: I1006 10:14:06.900440 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b6303c51-6db3-4ab6-aeab-edf39e88fdec-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"b6303c51-6db3-4ab6-aeab-edf39e88fdec\") " pod="openstack/ovsdbserver-sb-0" Oct 06 10:14:06 crc kubenswrapper[4824]: I1006 10:14:06.900670 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6303c51-6db3-4ab6-aeab-edf39e88fdec-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"b6303c51-6db3-4ab6-aeab-edf39e88fdec\") " pod="openstack/ovsdbserver-sb-0" Oct 06 10:14:06 crc kubenswrapper[4824]: I1006 10:14:06.900712 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b6303c51-6db3-4ab6-aeab-edf39e88fdec-config\") pod \"ovsdbserver-sb-0\" (UID: \"b6303c51-6db3-4ab6-aeab-edf39e88fdec\") " pod="openstack/ovsdbserver-sb-0" Oct 06 10:14:06 crc kubenswrapper[4824]: I1006 10:14:06.900783 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/b6303c51-6db3-4ab6-aeab-edf39e88fdec-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"b6303c51-6db3-4ab6-aeab-edf39e88fdec\") " pod="openstack/ovsdbserver-sb-0" Oct 06 10:14:06 crc kubenswrapper[4824]: I1006 10:14:06.900835 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mbw65\" (UniqueName: \"kubernetes.io/projected/b6303c51-6db3-4ab6-aeab-edf39e88fdec-kube-api-access-mbw65\") pod \"ovsdbserver-sb-0\" (UID: \"b6303c51-6db3-4ab6-aeab-edf39e88fdec\") " pod="openstack/ovsdbserver-sb-0" Oct 06 10:14:06 crc kubenswrapper[4824]: I1006 10:14:06.900891 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/b6303c51-6db3-4ab6-aeab-edf39e88fdec-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"b6303c51-6db3-4ab6-aeab-edf39e88fdec\") " pod="openstack/ovsdbserver-sb-0" Oct 06 10:14:06 crc kubenswrapper[4824]: I1006 10:14:06.900963 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ovsdbserver-sb-0\" (UID: \"b6303c51-6db3-4ab6-aeab-edf39e88fdec\") " pod="openstack/ovsdbserver-sb-0" Oct 06 10:14:07 crc kubenswrapper[4824]: I1006 10:14:07.002045 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/b6303c51-6db3-4ab6-aeab-edf39e88fdec-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"b6303c51-6db3-4ab6-aeab-edf39e88fdec\") " pod="openstack/ovsdbserver-sb-0" Oct 06 10:14:07 crc kubenswrapper[4824]: I1006 10:14:07.002104 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b6303c51-6db3-4ab6-aeab-edf39e88fdec-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"b6303c51-6db3-4ab6-aeab-edf39e88fdec\") " pod="openstack/ovsdbserver-sb-0" Oct 06 10:14:07 crc kubenswrapper[4824]: I1006 10:14:07.002177 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6303c51-6db3-4ab6-aeab-edf39e88fdec-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"b6303c51-6db3-4ab6-aeab-edf39e88fdec\") " pod="openstack/ovsdbserver-sb-0" Oct 06 10:14:07 crc kubenswrapper[4824]: I1006 10:14:07.002197 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b6303c51-6db3-4ab6-aeab-edf39e88fdec-config\") pod \"ovsdbserver-sb-0\" (UID: \"b6303c51-6db3-4ab6-aeab-edf39e88fdec\") " pod="openstack/ovsdbserver-sb-0" Oct 06 10:14:07 crc kubenswrapper[4824]: I1006 10:14:07.002224 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/b6303c51-6db3-4ab6-aeab-edf39e88fdec-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"b6303c51-6db3-4ab6-aeab-edf39e88fdec\") " pod="openstack/ovsdbserver-sb-0" Oct 06 10:14:07 crc kubenswrapper[4824]: I1006 10:14:07.002247 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mbw65\" (UniqueName: \"kubernetes.io/projected/b6303c51-6db3-4ab6-aeab-edf39e88fdec-kube-api-access-mbw65\") pod \"ovsdbserver-sb-0\" (UID: \"b6303c51-6db3-4ab6-aeab-edf39e88fdec\") " pod="openstack/ovsdbserver-sb-0" Oct 06 10:14:07 crc kubenswrapper[4824]: I1006 10:14:07.002270 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/b6303c51-6db3-4ab6-aeab-edf39e88fdec-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"b6303c51-6db3-4ab6-aeab-edf39e88fdec\") " pod="openstack/ovsdbserver-sb-0" Oct 06 10:14:07 crc kubenswrapper[4824]: I1006 10:14:07.002302 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ovsdbserver-sb-0\" (UID: \"b6303c51-6db3-4ab6-aeab-edf39e88fdec\") " pod="openstack/ovsdbserver-sb-0" Oct 06 10:14:07 crc kubenswrapper[4824]: I1006 10:14:07.003681 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/b6303c51-6db3-4ab6-aeab-edf39e88fdec-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"b6303c51-6db3-4ab6-aeab-edf39e88fdec\") " pod="openstack/ovsdbserver-sb-0" Oct 06 10:14:07 crc kubenswrapper[4824]: I1006 10:14:07.005544 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b6303c51-6db3-4ab6-aeab-edf39e88fdec-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"b6303c51-6db3-4ab6-aeab-edf39e88fdec\") " pod="openstack/ovsdbserver-sb-0" Oct 06 10:14:07 crc kubenswrapper[4824]: I1006 10:14:07.006229 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b6303c51-6db3-4ab6-aeab-edf39e88fdec-config\") pod \"ovsdbserver-sb-0\" (UID: \"b6303c51-6db3-4ab6-aeab-edf39e88fdec\") " pod="openstack/ovsdbserver-sb-0" Oct 06 10:14:07 crc kubenswrapper[4824]: I1006 10:14:07.006612 4824 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ovsdbserver-sb-0\" (UID: \"b6303c51-6db3-4ab6-aeab-edf39e88fdec\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/ovsdbserver-sb-0" Oct 06 10:14:07 crc kubenswrapper[4824]: I1006 10:14:07.009941 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/b6303c51-6db3-4ab6-aeab-edf39e88fdec-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"b6303c51-6db3-4ab6-aeab-edf39e88fdec\") " pod="openstack/ovsdbserver-sb-0" Oct 06 10:14:07 crc kubenswrapper[4824]: I1006 10:14:07.010442 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6303c51-6db3-4ab6-aeab-edf39e88fdec-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"b6303c51-6db3-4ab6-aeab-edf39e88fdec\") " pod="openstack/ovsdbserver-sb-0" Oct 06 10:14:07 crc kubenswrapper[4824]: I1006 10:14:07.011342 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/b6303c51-6db3-4ab6-aeab-edf39e88fdec-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"b6303c51-6db3-4ab6-aeab-edf39e88fdec\") " pod="openstack/ovsdbserver-sb-0" Oct 06 10:14:07 crc kubenswrapper[4824]: I1006 10:14:07.020819 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mbw65\" (UniqueName: \"kubernetes.io/projected/b6303c51-6db3-4ab6-aeab-edf39e88fdec-kube-api-access-mbw65\") pod \"ovsdbserver-sb-0\" (UID: \"b6303c51-6db3-4ab6-aeab-edf39e88fdec\") " pod="openstack/ovsdbserver-sb-0" Oct 06 10:14:07 crc kubenswrapper[4824]: I1006 10:14:07.029241 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ovsdbserver-sb-0\" (UID: \"b6303c51-6db3-4ab6-aeab-edf39e88fdec\") " pod="openstack/ovsdbserver-sb-0" Oct 06 10:14:07 crc kubenswrapper[4824]: I1006 10:14:07.180814 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 06 10:14:07 crc kubenswrapper[4824]: I1006 10:14:07.202363 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Oct 06 10:14:07 crc kubenswrapper[4824]: E1006 10:14:07.938037 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Oct 06 10:14:07 crc kubenswrapper[4824]: E1006 10:14:07.940074 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n68chd6h679hbfh55fhc6h5ffh5d8h94h56ch589hb4hc5h57bh677hcdh655h8dh667h675h654h66ch567h8fh659h5b4h675h566h55bh54h67dh6dq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-gfv2m,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-666b6646f7-bdp5v_openstack(c7eb054c-bd83-4ba1-a1ac-f8eeaf049fb7): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 06 10:14:07 crc kubenswrapper[4824]: E1006 10:14:07.941308 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-666b6646f7-bdp5v" podUID="c7eb054c-bd83-4ba1-a1ac-f8eeaf049fb7" Oct 06 10:14:07 crc kubenswrapper[4824]: E1006 10:14:07.957945 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Oct 06 10:14:07 crc kubenswrapper[4824]: E1006 10:14:07.958317 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-4l4nz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-78dd6ddcc-h2kwr_openstack(b9e0d836-802d-4cad-af61-dd2abae51bbc): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 06 10:14:07 crc kubenswrapper[4824]: E1006 10:14:07.961122 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-78dd6ddcc-h2kwr" podUID="b9e0d836-802d-4cad-af61-dd2abae51bbc" Oct 06 10:14:07 crc kubenswrapper[4824]: E1006 10:14:07.968435 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Oct 06 10:14:07 crc kubenswrapper[4824]: E1006 10:14:07.968605 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9khjd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-675f4bcbfc-qf5nc_openstack(59e8cadf-2f15-45c4-8d6d-3457f8291d14): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 06 10:14:07 crc kubenswrapper[4824]: E1006 10:14:07.969760 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-675f4bcbfc-qf5nc" podUID="59e8cadf-2f15-45c4-8d6d-3457f8291d14" Oct 06 10:14:08 crc kubenswrapper[4824]: E1006 10:14:08.243823 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-666b6646f7-bdp5v" podUID="c7eb054c-bd83-4ba1-a1ac-f8eeaf049fb7" Oct 06 10:14:08 crc kubenswrapper[4824]: I1006 10:14:08.326364 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 06 10:14:09 crc kubenswrapper[4824]: I1006 10:14:09.273564 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-qf5nc" event={"ID":"59e8cadf-2f15-45c4-8d6d-3457f8291d14","Type":"ContainerDied","Data":"f2f7d30332fcc1bc1c81eb9296db8fc976d575bfe7212017d3786f5bf645dc47"} Oct 06 10:14:09 crc kubenswrapper[4824]: I1006 10:14:09.273622 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f2f7d30332fcc1bc1c81eb9296db8fc976d575bfe7212017d3786f5bf645dc47" Oct 06 10:14:09 crc kubenswrapper[4824]: I1006 10:14:09.291617 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"3a12653c-c0d0-46e3-b909-77bd2a7f6c4a","Type":"ContainerStarted","Data":"4190a2121dc1f0ce8a483c7a2e8c31e3b608a0ebf6ee4368ff175e9fb5671802"} Oct 06 10:14:09 crc kubenswrapper[4824]: I1006 10:14:09.291724 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"1510d553-3505-4bd2-9666-c95cecc43d01","Type":"ContainerStarted","Data":"b617f9f4a0f8ab5830122250bafc9b50d1b4a88566b49e37eb8f950c6c90ee92"} Oct 06 10:14:09 crc kubenswrapper[4824]: I1006 10:14:09.291787 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-h2kwr" event={"ID":"b9e0d836-802d-4cad-af61-dd2abae51bbc","Type":"ContainerDied","Data":"d71a18afba91462402ec1ef6dd857fbaa42b0f455d592cf7b622c6c61a744af0"} Oct 06 10:14:09 crc kubenswrapper[4824]: I1006 10:14:09.291812 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d71a18afba91462402ec1ef6dd857fbaa42b0f455d592cf7b622c6c61a744af0" Oct 06 10:14:09 crc kubenswrapper[4824]: I1006 10:14:09.360661 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-qf5nc" Oct 06 10:14:09 crc kubenswrapper[4824]: I1006 10:14:09.373878 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-h2kwr" Oct 06 10:14:09 crc kubenswrapper[4824]: I1006 10:14:09.557766 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b9e0d836-802d-4cad-af61-dd2abae51bbc-config\") pod \"b9e0d836-802d-4cad-af61-dd2abae51bbc\" (UID: \"b9e0d836-802d-4cad-af61-dd2abae51bbc\") " Oct 06 10:14:09 crc kubenswrapper[4824]: I1006 10:14:09.558225 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b9e0d836-802d-4cad-af61-dd2abae51bbc-dns-svc\") pod \"b9e0d836-802d-4cad-af61-dd2abae51bbc\" (UID: \"b9e0d836-802d-4cad-af61-dd2abae51bbc\") " Oct 06 10:14:09 crc kubenswrapper[4824]: I1006 10:14:09.558292 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9khjd\" (UniqueName: \"kubernetes.io/projected/59e8cadf-2f15-45c4-8d6d-3457f8291d14-kube-api-access-9khjd\") pod \"59e8cadf-2f15-45c4-8d6d-3457f8291d14\" (UID: \"59e8cadf-2f15-45c4-8d6d-3457f8291d14\") " Oct 06 10:14:09 crc kubenswrapper[4824]: I1006 10:14:09.558358 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4l4nz\" (UniqueName: \"kubernetes.io/projected/b9e0d836-802d-4cad-af61-dd2abae51bbc-kube-api-access-4l4nz\") pod \"b9e0d836-802d-4cad-af61-dd2abae51bbc\" (UID: \"b9e0d836-802d-4cad-af61-dd2abae51bbc\") " Oct 06 10:14:09 crc kubenswrapper[4824]: I1006 10:14:09.558393 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/59e8cadf-2f15-45c4-8d6d-3457f8291d14-config\") pod \"59e8cadf-2f15-45c4-8d6d-3457f8291d14\" (UID: \"59e8cadf-2f15-45c4-8d6d-3457f8291d14\") " Oct 06 10:14:09 crc kubenswrapper[4824]: I1006 10:14:09.558427 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b9e0d836-802d-4cad-af61-dd2abae51bbc-config" (OuterVolumeSpecName: "config") pod "b9e0d836-802d-4cad-af61-dd2abae51bbc" (UID: "b9e0d836-802d-4cad-af61-dd2abae51bbc"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:14:09 crc kubenswrapper[4824]: I1006 10:14:09.558838 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b9e0d836-802d-4cad-af61-dd2abae51bbc-config\") on node \"crc\" DevicePath \"\"" Oct 06 10:14:09 crc kubenswrapper[4824]: I1006 10:14:09.558921 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b9e0d836-802d-4cad-af61-dd2abae51bbc-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b9e0d836-802d-4cad-af61-dd2abae51bbc" (UID: "b9e0d836-802d-4cad-af61-dd2abae51bbc"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:14:09 crc kubenswrapper[4824]: I1006 10:14:09.563595 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/59e8cadf-2f15-45c4-8d6d-3457f8291d14-config" (OuterVolumeSpecName: "config") pod "59e8cadf-2f15-45c4-8d6d-3457f8291d14" (UID: "59e8cadf-2f15-45c4-8d6d-3457f8291d14"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:14:09 crc kubenswrapper[4824]: I1006 10:14:09.564460 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/59e8cadf-2f15-45c4-8d6d-3457f8291d14-kube-api-access-9khjd" (OuterVolumeSpecName: "kube-api-access-9khjd") pod "59e8cadf-2f15-45c4-8d6d-3457f8291d14" (UID: "59e8cadf-2f15-45c4-8d6d-3457f8291d14"). InnerVolumeSpecName "kube-api-access-9khjd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:14:09 crc kubenswrapper[4824]: I1006 10:14:09.581508 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b9e0d836-802d-4cad-af61-dd2abae51bbc-kube-api-access-4l4nz" (OuterVolumeSpecName: "kube-api-access-4l4nz") pod "b9e0d836-802d-4cad-af61-dd2abae51bbc" (UID: "b9e0d836-802d-4cad-af61-dd2abae51bbc"). InnerVolumeSpecName "kube-api-access-4l4nz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:14:09 crc kubenswrapper[4824]: I1006 10:14:09.660348 4824 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b9e0d836-802d-4cad-af61-dd2abae51bbc-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 06 10:14:09 crc kubenswrapper[4824]: I1006 10:14:09.660379 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9khjd\" (UniqueName: \"kubernetes.io/projected/59e8cadf-2f15-45c4-8d6d-3457f8291d14-kube-api-access-9khjd\") on node \"crc\" DevicePath \"\"" Oct 06 10:14:09 crc kubenswrapper[4824]: I1006 10:14:09.660390 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4l4nz\" (UniqueName: \"kubernetes.io/projected/b9e0d836-802d-4cad-af61-dd2abae51bbc-kube-api-access-4l4nz\") on node \"crc\" DevicePath \"\"" Oct 06 10:14:09 crc kubenswrapper[4824]: I1006 10:14:09.660402 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/59e8cadf-2f15-45c4-8d6d-3457f8291d14-config\") on node \"crc\" DevicePath \"\"" Oct 06 10:14:09 crc kubenswrapper[4824]: I1006 10:14:09.713359 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 06 10:14:09 crc kubenswrapper[4824]: W1006 10:14:09.723489 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1d365fcf_17fb_411e_ab2c_3c5dc8384b21.slice/crio-72614a4f22524423b9852d1c0767b06c5bb22df6e098392a586f481e71f95774 WatchSource:0}: Error finding container 72614a4f22524423b9852d1c0767b06c5bb22df6e098392a586f481e71f95774: Status 404 returned error can't find the container with id 72614a4f22524423b9852d1c0767b06c5bb22df6e098392a586f481e71f95774 Oct 06 10:14:09 crc kubenswrapper[4824]: W1006 10:14:09.728662 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod58f6ad3c_7a20_4be6_8ffb_cadff1c6adef.slice/crio-e52e451ae3dd7302ee9af1f1094d29da59d305f186e0ef605834117428006c38 WatchSource:0}: Error finding container e52e451ae3dd7302ee9af1f1094d29da59d305f186e0ef605834117428006c38: Status 404 returned error can't find the container with id e52e451ae3dd7302ee9af1f1094d29da59d305f186e0ef605834117428006c38 Oct 06 10:14:09 crc kubenswrapper[4824]: I1006 10:14:09.731243 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Oct 06 10:14:09 crc kubenswrapper[4824]: I1006 10:14:09.854818 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-48qsh"] Oct 06 10:14:09 crc kubenswrapper[4824]: I1006 10:14:09.986151 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-v22g5"] Oct 06 10:14:10 crc kubenswrapper[4824]: I1006 10:14:10.087620 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 06 10:14:10 crc kubenswrapper[4824]: W1006 10:14:10.090356 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda39bac9c_ff87_4d4a_ad96_fa5618ca4e40.slice/crio-57826f32f56fff4d4a65cfd59124b490ac21cbe1d8847767fad8fcda3667ef4b WatchSource:0}: Error finding container 57826f32f56fff4d4a65cfd59124b490ac21cbe1d8847767fad8fcda3667ef4b: Status 404 returned error can't find the container with id 57826f32f56fff4d4a65cfd59124b490ac21cbe1d8847767fad8fcda3667ef4b Oct 06 10:14:10 crc kubenswrapper[4824]: I1006 10:14:10.188170 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 06 10:14:10 crc kubenswrapper[4824]: W1006 10:14:10.221391 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb6303c51_6db3_4ab6_aeab_edf39e88fdec.slice/crio-fa57cfdbd3697f6143b58764f5cd1c96849076630a005471d9b491b605925c18 WatchSource:0}: Error finding container fa57cfdbd3697f6143b58764f5cd1c96849076630a005471d9b491b605925c18: Status 404 returned error can't find the container with id fa57cfdbd3697f6143b58764f5cd1c96849076630a005471d9b491b605925c18 Oct 06 10:14:10 crc kubenswrapper[4824]: I1006 10:14:10.290097 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"58f6ad3c-7a20-4be6-8ffb-cadff1c6adef","Type":"ContainerStarted","Data":"e52e451ae3dd7302ee9af1f1094d29da59d305f186e0ef605834117428006c38"} Oct 06 10:14:10 crc kubenswrapper[4824]: I1006 10:14:10.291599 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"1d365fcf-17fb-411e-ab2c-3c5dc8384b21","Type":"ContainerStarted","Data":"72614a4f22524423b9852d1c0767b06c5bb22df6e098392a586f481e71f95774"} Oct 06 10:14:10 crc kubenswrapper[4824]: I1006 10:14:10.293649 4824 generic.go:334] "Generic (PLEG): container finished" podID="97cd45b1-470f-483b-b1ed-6dfd5fed1ca8" containerID="949861b1f09b2a44df67727602bedc146340d3bf143c1435c6da5c97af7e69bf" exitCode=0 Oct 06 10:14:10 crc kubenswrapper[4824]: I1006 10:14:10.293718 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-cnwhb" event={"ID":"97cd45b1-470f-483b-b1ed-6dfd5fed1ca8","Type":"ContainerDied","Data":"949861b1f09b2a44df67727602bedc146340d3bf143c1435c6da5c97af7e69bf"} Oct 06 10:14:10 crc kubenswrapper[4824]: I1006 10:14:10.295151 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-v22g5" event={"ID":"59148e21-e79a-466f-bed5-24671942c24f","Type":"ContainerStarted","Data":"e56276dbcfac0060274e8203e11b28f48abf24dfd148dd032b2fd76557d384b7"} Oct 06 10:14:10 crc kubenswrapper[4824]: I1006 10:14:10.297437 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"b6303c51-6db3-4ab6-aeab-edf39e88fdec","Type":"ContainerStarted","Data":"fa57cfdbd3697f6143b58764f5cd1c96849076630a005471d9b491b605925c18"} Oct 06 10:14:10 crc kubenswrapper[4824]: I1006 10:14:10.299337 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"a39bac9c-ff87-4d4a-ad96-fa5618ca4e40","Type":"ContainerStarted","Data":"57826f32f56fff4d4a65cfd59124b490ac21cbe1d8847767fad8fcda3667ef4b"} Oct 06 10:14:10 crc kubenswrapper[4824]: I1006 10:14:10.300499 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-h2kwr" Oct 06 10:14:10 crc kubenswrapper[4824]: I1006 10:14:10.302206 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-48qsh" event={"ID":"03d76a8a-df87-4a48-8fb9-e6a502b37ae8","Type":"ContainerStarted","Data":"3d2a7f73b21e4ce3bd777e121321ce63f2ea2e5edca7a1a7273b498a98df53fd"} Oct 06 10:14:10 crc kubenswrapper[4824]: I1006 10:14:10.302293 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-qf5nc" Oct 06 10:14:10 crc kubenswrapper[4824]: I1006 10:14:10.370521 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-qf5nc"] Oct 06 10:14:10 crc kubenswrapper[4824]: I1006 10:14:10.378883 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-qf5nc"] Oct 06 10:14:10 crc kubenswrapper[4824]: I1006 10:14:10.403634 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-h2kwr"] Oct 06 10:14:10 crc kubenswrapper[4824]: I1006 10:14:10.411430 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-h2kwr"] Oct 06 10:14:11 crc kubenswrapper[4824]: I1006 10:14:11.293675 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="59e8cadf-2f15-45c4-8d6d-3457f8291d14" path="/var/lib/kubelet/pods/59e8cadf-2f15-45c4-8d6d-3457f8291d14/volumes" Oct 06 10:14:11 crc kubenswrapper[4824]: I1006 10:14:11.294277 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b9e0d836-802d-4cad-af61-dd2abae51bbc" path="/var/lib/kubelet/pods/b9e0d836-802d-4cad-af61-dd2abae51bbc/volumes" Oct 06 10:14:11 crc kubenswrapper[4824]: I1006 10:14:11.321342 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"c8dbad1e-02be-43c8-8f80-ab771ee81742","Type":"ContainerStarted","Data":"9e782d23005bd2ab15abfb2afc9b231130f78b0f01da4901f27b58d3b1eb931d"} Oct 06 10:14:11 crc kubenswrapper[4824]: I1006 10:14:11.323215 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"f7d047da-89e3-4cd4-bf02-e6a3e15c585f","Type":"ContainerStarted","Data":"2512559a4666f95711ee0e9d31717a5aa74ce89aab85c9eec580e792833cd962"} Oct 06 10:14:17 crc kubenswrapper[4824]: I1006 10:14:17.365541 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-cnwhb" event={"ID":"97cd45b1-470f-483b-b1ed-6dfd5fed1ca8","Type":"ContainerStarted","Data":"c55a2252f37111bdab065db4f6503c909cd8a7fb4769f9893846fd731cd879e7"} Oct 06 10:14:17 crc kubenswrapper[4824]: I1006 10:14:17.367033 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-57d769cc4f-cnwhb" Oct 06 10:14:17 crc kubenswrapper[4824]: I1006 10:14:17.393097 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-57d769cc4f-cnwhb" podStartSLOduration=9.554090747 podStartE2EDuration="25.393073555s" podCreationTimestamp="2025-10-06 10:13:52 +0000 UTC" firstStartedPulling="2025-10-06 10:13:53.677313064 +0000 UTC m=+943.041735915" lastFinishedPulling="2025-10-06 10:14:09.516295872 +0000 UTC m=+958.880718723" observedRunningTime="2025-10-06 10:14:17.386886222 +0000 UTC m=+966.751309083" watchObservedRunningTime="2025-10-06 10:14:17.393073555 +0000 UTC m=+966.757496416" Oct 06 10:14:18 crc kubenswrapper[4824]: I1006 10:14:18.375141 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"58f6ad3c-7a20-4be6-8ffb-cadff1c6adef","Type":"ContainerStarted","Data":"2efed31dce26d2ac9a6ea070fbe6c1b3defd24ca940731942145cf69f7840691"} Oct 06 10:14:18 crc kubenswrapper[4824]: I1006 10:14:18.378709 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"1d365fcf-17fb-411e-ab2c-3c5dc8384b21","Type":"ContainerStarted","Data":"5bbaaf09ee506a794b2f4f3000b75a5e8afaa034fab8e585936a61b6d460e76d"} Oct 06 10:14:18 crc kubenswrapper[4824]: I1006 10:14:18.378855 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Oct 06 10:14:18 crc kubenswrapper[4824]: I1006 10:14:18.380689 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-v22g5" event={"ID":"59148e21-e79a-466f-bed5-24671942c24f","Type":"ContainerStarted","Data":"a3fd1b67f07ede7a000caf4385a37aca7a27ee10366b0247d6e5c0dca75ed294"} Oct 06 10:14:18 crc kubenswrapper[4824]: I1006 10:14:18.383078 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"b6303c51-6db3-4ab6-aeab-edf39e88fdec","Type":"ContainerStarted","Data":"883445e45a67560d57ba473ff338e60e5b7d6d78368e0498a45d8801a1d14817"} Oct 06 10:14:18 crc kubenswrapper[4824]: I1006 10:14:18.385019 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"3a12653c-c0d0-46e3-b909-77bd2a7f6c4a","Type":"ContainerStarted","Data":"fceab564f3523f55944067edab1c7b4de74728084f0dc31932b61bc07b69f084"} Oct 06 10:14:18 crc kubenswrapper[4824]: I1006 10:14:18.386059 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Oct 06 10:14:18 crc kubenswrapper[4824]: I1006 10:14:18.387356 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"a39bac9c-ff87-4d4a-ad96-fa5618ca4e40","Type":"ContainerStarted","Data":"11d658f9e875603479768d14195c8ad0d810e50b43b85494fcef128f442ab8d1"} Oct 06 10:14:18 crc kubenswrapper[4824]: I1006 10:14:18.389544 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"1510d553-3505-4bd2-9666-c95cecc43d01","Type":"ContainerStarted","Data":"31d68ffb4c622e96c3eeaeeaac04859ee823eedf70782100091a205006934f0f"} Oct 06 10:14:18 crc kubenswrapper[4824]: I1006 10:14:18.392878 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-48qsh" event={"ID":"03d76a8a-df87-4a48-8fb9-e6a502b37ae8","Type":"ContainerStarted","Data":"f0d817a7861dfce691c492c0547f661a4d89c07018b70324742e89a4d67b2354"} Oct 06 10:14:18 crc kubenswrapper[4824]: I1006 10:14:18.392920 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-48qsh" Oct 06 10:14:18 crc kubenswrapper[4824]: I1006 10:14:18.430774 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=11.353472857 podStartE2EDuration="19.430744856s" podCreationTimestamp="2025-10-06 10:13:59 +0000 UTC" firstStartedPulling="2025-10-06 10:14:09.728961439 +0000 UTC m=+959.093384300" lastFinishedPulling="2025-10-06 10:14:17.806233438 +0000 UTC m=+967.170656299" observedRunningTime="2025-10-06 10:14:18.425173938 +0000 UTC m=+967.789596809" watchObservedRunningTime="2025-10-06 10:14:18.430744856 +0000 UTC m=+967.795167717" Oct 06 10:14:18 crc kubenswrapper[4824]: I1006 10:14:18.478776 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-48qsh" podStartSLOduration=8.802677683 podStartE2EDuration="16.478750285s" podCreationTimestamp="2025-10-06 10:14:02 +0000 UTC" firstStartedPulling="2025-10-06 10:14:09.8735434 +0000 UTC m=+959.237966261" lastFinishedPulling="2025-10-06 10:14:17.549616002 +0000 UTC m=+966.914038863" observedRunningTime="2025-10-06 10:14:18.473793151 +0000 UTC m=+967.838216032" watchObservedRunningTime="2025-10-06 10:14:18.478750285 +0000 UTC m=+967.843173146" Oct 06 10:14:18 crc kubenswrapper[4824]: I1006 10:14:18.523075 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=14.774957271 podStartE2EDuration="21.523049251s" podCreationTimestamp="2025-10-06 10:13:57 +0000 UTC" firstStartedPulling="2025-10-06 10:14:09.200743046 +0000 UTC m=+958.565165947" lastFinishedPulling="2025-10-06 10:14:15.948835066 +0000 UTC m=+965.313257927" observedRunningTime="2025-10-06 10:14:18.516304935 +0000 UTC m=+967.880727796" watchObservedRunningTime="2025-10-06 10:14:18.523049251 +0000 UTC m=+967.887472122" Oct 06 10:14:19 crc kubenswrapper[4824]: I1006 10:14:19.409659 4824 generic.go:334] "Generic (PLEG): container finished" podID="59148e21-e79a-466f-bed5-24671942c24f" containerID="a3fd1b67f07ede7a000caf4385a37aca7a27ee10366b0247d6e5c0dca75ed294" exitCode=0 Oct 06 10:14:19 crc kubenswrapper[4824]: I1006 10:14:19.409818 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-v22g5" event={"ID":"59148e21-e79a-466f-bed5-24671942c24f","Type":"ContainerDied","Data":"a3fd1b67f07ede7a000caf4385a37aca7a27ee10366b0247d6e5c0dca75ed294"} Oct 06 10:14:20 crc kubenswrapper[4824]: I1006 10:14:20.425867 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-v22g5" event={"ID":"59148e21-e79a-466f-bed5-24671942c24f","Type":"ContainerStarted","Data":"b54ce0af227e3df9eb29ba89f761f9e9c1b94560ae4c47883e637a7c10efe7d1"} Oct 06 10:14:22 crc kubenswrapper[4824]: I1006 10:14:22.445903 4824 generic.go:334] "Generic (PLEG): container finished" podID="1510d553-3505-4bd2-9666-c95cecc43d01" containerID="31d68ffb4c622e96c3eeaeeaac04859ee823eedf70782100091a205006934f0f" exitCode=0 Oct 06 10:14:22 crc kubenswrapper[4824]: I1006 10:14:22.446046 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"1510d553-3505-4bd2-9666-c95cecc43d01","Type":"ContainerDied","Data":"31d68ffb4c622e96c3eeaeeaac04859ee823eedf70782100091a205006934f0f"} Oct 06 10:14:22 crc kubenswrapper[4824]: I1006 10:14:22.448726 4824 generic.go:334] "Generic (PLEG): container finished" podID="58f6ad3c-7a20-4be6-8ffb-cadff1c6adef" containerID="2efed31dce26d2ac9a6ea070fbe6c1b3defd24ca940731942145cf69f7840691" exitCode=0 Oct 06 10:14:22 crc kubenswrapper[4824]: I1006 10:14:22.448770 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"58f6ad3c-7a20-4be6-8ffb-cadff1c6adef","Type":"ContainerDied","Data":"2efed31dce26d2ac9a6ea070fbe6c1b3defd24ca940731942145cf69f7840691"} Oct 06 10:14:22 crc kubenswrapper[4824]: I1006 10:14:22.452039 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-v22g5" event={"ID":"59148e21-e79a-466f-bed5-24671942c24f","Type":"ContainerStarted","Data":"4a1a4ccfff66ea4a04962a960ff0acce6581547f962c137d97664e99cda34775"} Oct 06 10:14:22 crc kubenswrapper[4824]: I1006 10:14:22.452296 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-v22g5" Oct 06 10:14:22 crc kubenswrapper[4824]: I1006 10:14:22.461081 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"b6303c51-6db3-4ab6-aeab-edf39e88fdec","Type":"ContainerStarted","Data":"e7b001cff33521fa71c34e1ff741b023f983943b01972e56b5f4828824d1bec0"} Oct 06 10:14:22 crc kubenswrapper[4824]: I1006 10:14:22.466662 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"a39bac9c-ff87-4d4a-ad96-fa5618ca4e40","Type":"ContainerStarted","Data":"02efd33f5747689c3effecf6d0a7fffe065bf1735670f32226e18af4953ab5a6"} Oct 06 10:14:22 crc kubenswrapper[4824]: I1006 10:14:22.517172 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=5.8219977499999995 podStartE2EDuration="17.517148953s" podCreationTimestamp="2025-10-06 10:14:05 +0000 UTC" firstStartedPulling="2025-10-06 10:14:10.224592764 +0000 UTC m=+959.589015625" lastFinishedPulling="2025-10-06 10:14:21.919743967 +0000 UTC m=+971.284166828" observedRunningTime="2025-10-06 10:14:22.508597542 +0000 UTC m=+971.873020403" watchObservedRunningTime="2025-10-06 10:14:22.517148953 +0000 UTC m=+971.881571804" Oct 06 10:14:22 crc kubenswrapper[4824]: I1006 10:14:22.533327 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-v22g5" podStartSLOduration=13.718374499 podStartE2EDuration="20.533307084s" podCreationTimestamp="2025-10-06 10:14:02 +0000 UTC" firstStartedPulling="2025-10-06 10:14:09.999996311 +0000 UTC m=+959.364419172" lastFinishedPulling="2025-10-06 10:14:16.814928896 +0000 UTC m=+966.179351757" observedRunningTime="2025-10-06 10:14:22.531236992 +0000 UTC m=+971.895659843" watchObservedRunningTime="2025-10-06 10:14:22.533307084 +0000 UTC m=+971.897729965" Oct 06 10:14:22 crc kubenswrapper[4824]: I1006 10:14:22.558012 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=8.737376074 podStartE2EDuration="20.557994695s" podCreationTimestamp="2025-10-06 10:14:02 +0000 UTC" firstStartedPulling="2025-10-06 10:14:10.09281151 +0000 UTC m=+959.457234371" lastFinishedPulling="2025-10-06 10:14:21.913430091 +0000 UTC m=+971.277852992" observedRunningTime="2025-10-06 10:14:22.554444827 +0000 UTC m=+971.918867688" watchObservedRunningTime="2025-10-06 10:14:22.557994695 +0000 UTC m=+971.922417556" Oct 06 10:14:22 crc kubenswrapper[4824]: I1006 10:14:22.675803 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Oct 06 10:14:22 crc kubenswrapper[4824]: I1006 10:14:22.890405 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-v22g5" Oct 06 10:14:23 crc kubenswrapper[4824]: I1006 10:14:23.370543 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-57d769cc4f-cnwhb" Oct 06 10:14:23 crc kubenswrapper[4824]: I1006 10:14:23.433575 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-bdp5v"] Oct 06 10:14:23 crc kubenswrapper[4824]: I1006 10:14:23.496232 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"58f6ad3c-7a20-4be6-8ffb-cadff1c6adef","Type":"ContainerStarted","Data":"0c8f63e4dee10d1798e2e917e9a756f8c2af845ca951d24747d690280ddd8002"} Oct 06 10:14:23 crc kubenswrapper[4824]: I1006 10:14:23.498615 4824 generic.go:334] "Generic (PLEG): container finished" podID="c7eb054c-bd83-4ba1-a1ac-f8eeaf049fb7" containerID="146faf8c0c4082afdc61e51b86773d8ac758115121f4273bf674fc8724947637" exitCode=0 Oct 06 10:14:23 crc kubenswrapper[4824]: I1006 10:14:23.498689 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-bdp5v" event={"ID":"c7eb054c-bd83-4ba1-a1ac-f8eeaf049fb7","Type":"ContainerDied","Data":"146faf8c0c4082afdc61e51b86773d8ac758115121f4273bf674fc8724947637"} Oct 06 10:14:23 crc kubenswrapper[4824]: I1006 10:14:23.508539 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"1510d553-3505-4bd2-9666-c95cecc43d01","Type":"ContainerStarted","Data":"f9e48ac8a9ecd19d28a7acc1b337e95016af54bbe7539a73ae200602792aef6f"} Oct 06 10:14:23 crc kubenswrapper[4824]: I1006 10:14:23.527543 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=21.443880757 podStartE2EDuration="28.527518148s" podCreationTimestamp="2025-10-06 10:13:55 +0000 UTC" firstStartedPulling="2025-10-06 10:14:09.731420569 +0000 UTC m=+959.095843430" lastFinishedPulling="2025-10-06 10:14:16.81505792 +0000 UTC m=+966.179480821" observedRunningTime="2025-10-06 10:14:23.521097228 +0000 UTC m=+972.885520089" watchObservedRunningTime="2025-10-06 10:14:23.527518148 +0000 UTC m=+972.891941009" Oct 06 10:14:23 crc kubenswrapper[4824]: I1006 10:14:23.581100 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=21.83443533 podStartE2EDuration="28.581069844s" podCreationTimestamp="2025-10-06 10:13:55 +0000 UTC" firstStartedPulling="2025-10-06 10:14:09.203066444 +0000 UTC m=+958.567489305" lastFinishedPulling="2025-10-06 10:14:15.949700958 +0000 UTC m=+965.314123819" observedRunningTime="2025-10-06 10:14:23.570171813 +0000 UTC m=+972.934594674" watchObservedRunningTime="2025-10-06 10:14:23.581069844 +0000 UTC m=+972.945492695" Oct 06 10:14:23 crc kubenswrapper[4824]: I1006 10:14:23.728126 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Oct 06 10:14:23 crc kubenswrapper[4824]: I1006 10:14:23.847371 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-bdp5v" Oct 06 10:14:23 crc kubenswrapper[4824]: I1006 10:14:23.965354 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c7eb054c-bd83-4ba1-a1ac-f8eeaf049fb7-dns-svc\") pod \"c7eb054c-bd83-4ba1-a1ac-f8eeaf049fb7\" (UID: \"c7eb054c-bd83-4ba1-a1ac-f8eeaf049fb7\") " Oct 06 10:14:23 crc kubenswrapper[4824]: I1006 10:14:23.965436 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c7eb054c-bd83-4ba1-a1ac-f8eeaf049fb7-config\") pod \"c7eb054c-bd83-4ba1-a1ac-f8eeaf049fb7\" (UID: \"c7eb054c-bd83-4ba1-a1ac-f8eeaf049fb7\") " Oct 06 10:14:23 crc kubenswrapper[4824]: I1006 10:14:23.965540 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gfv2m\" (UniqueName: \"kubernetes.io/projected/c7eb054c-bd83-4ba1-a1ac-f8eeaf049fb7-kube-api-access-gfv2m\") pod \"c7eb054c-bd83-4ba1-a1ac-f8eeaf049fb7\" (UID: \"c7eb054c-bd83-4ba1-a1ac-f8eeaf049fb7\") " Oct 06 10:14:23 crc kubenswrapper[4824]: I1006 10:14:23.972331 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c7eb054c-bd83-4ba1-a1ac-f8eeaf049fb7-kube-api-access-gfv2m" (OuterVolumeSpecName: "kube-api-access-gfv2m") pod "c7eb054c-bd83-4ba1-a1ac-f8eeaf049fb7" (UID: "c7eb054c-bd83-4ba1-a1ac-f8eeaf049fb7"). InnerVolumeSpecName "kube-api-access-gfv2m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:14:23 crc kubenswrapper[4824]: I1006 10:14:23.989572 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c7eb054c-bd83-4ba1-a1ac-f8eeaf049fb7-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "c7eb054c-bd83-4ba1-a1ac-f8eeaf049fb7" (UID: "c7eb054c-bd83-4ba1-a1ac-f8eeaf049fb7"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:14:23 crc kubenswrapper[4824]: I1006 10:14:23.994581 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c7eb054c-bd83-4ba1-a1ac-f8eeaf049fb7-config" (OuterVolumeSpecName: "config") pod "c7eb054c-bd83-4ba1-a1ac-f8eeaf049fb7" (UID: "c7eb054c-bd83-4ba1-a1ac-f8eeaf049fb7"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:14:24 crc kubenswrapper[4824]: I1006 10:14:24.067493 4824 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c7eb054c-bd83-4ba1-a1ac-f8eeaf049fb7-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 06 10:14:24 crc kubenswrapper[4824]: I1006 10:14:24.067543 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c7eb054c-bd83-4ba1-a1ac-f8eeaf049fb7-config\") on node \"crc\" DevicePath \"\"" Oct 06 10:14:24 crc kubenswrapper[4824]: I1006 10:14:24.067559 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gfv2m\" (UniqueName: \"kubernetes.io/projected/c7eb054c-bd83-4ba1-a1ac-f8eeaf049fb7-kube-api-access-gfv2m\") on node \"crc\" DevicePath \"\"" Oct 06 10:14:24 crc kubenswrapper[4824]: I1006 10:14:24.519799 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-bdp5v" Oct 06 10:14:24 crc kubenswrapper[4824]: I1006 10:14:24.522114 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-bdp5v" event={"ID":"c7eb054c-bd83-4ba1-a1ac-f8eeaf049fb7","Type":"ContainerDied","Data":"671fef97b2748d4c605e089dbe734c96ba2bc2ec2eab693969b46a8325677ff4"} Oct 06 10:14:24 crc kubenswrapper[4824]: I1006 10:14:24.522174 4824 scope.go:117] "RemoveContainer" containerID="146faf8c0c4082afdc61e51b86773d8ac758115121f4273bf674fc8724947637" Oct 06 10:14:24 crc kubenswrapper[4824]: I1006 10:14:24.627984 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-bdp5v"] Oct 06 10:14:24 crc kubenswrapper[4824]: I1006 10:14:24.631951 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-bdp5v"] Oct 06 10:14:24 crc kubenswrapper[4824]: I1006 10:14:24.725275 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Oct 06 10:14:24 crc kubenswrapper[4824]: I1006 10:14:24.772641 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Oct 06 10:14:25 crc kubenswrapper[4824]: I1006 10:14:25.181645 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Oct 06 10:14:25 crc kubenswrapper[4824]: I1006 10:14:25.239540 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Oct 06 10:14:25 crc kubenswrapper[4824]: I1006 10:14:25.286697 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c7eb054c-bd83-4ba1-a1ac-f8eeaf049fb7" path="/var/lib/kubelet/pods/c7eb054c-bd83-4ba1-a1ac-f8eeaf049fb7/volumes" Oct 06 10:14:25 crc kubenswrapper[4824]: I1006 10:14:25.529119 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Oct 06 10:14:25 crc kubenswrapper[4824]: I1006 10:14:25.583464 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Oct 06 10:14:25 crc kubenswrapper[4824]: I1006 10:14:25.600401 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Oct 06 10:14:25 crc kubenswrapper[4824]: I1006 10:14:25.871405 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-7md89"] Oct 06 10:14:25 crc kubenswrapper[4824]: E1006 10:14:25.871831 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c7eb054c-bd83-4ba1-a1ac-f8eeaf049fb7" containerName="init" Oct 06 10:14:25 crc kubenswrapper[4824]: I1006 10:14:25.871852 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="c7eb054c-bd83-4ba1-a1ac-f8eeaf049fb7" containerName="init" Oct 06 10:14:25 crc kubenswrapper[4824]: I1006 10:14:25.872119 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="c7eb054c-bd83-4ba1-a1ac-f8eeaf049fb7" containerName="init" Oct 06 10:14:25 crc kubenswrapper[4824]: I1006 10:14:25.873231 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-7md89" Oct 06 10:14:25 crc kubenswrapper[4824]: I1006 10:14:25.875424 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Oct 06 10:14:25 crc kubenswrapper[4824]: I1006 10:14:25.904409 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-7md89"] Oct 06 10:14:25 crc kubenswrapper[4824]: I1006 10:14:25.933952 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-77kzd"] Oct 06 10:14:25 crc kubenswrapper[4824]: I1006 10:14:25.940091 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-77kzd" Oct 06 10:14:25 crc kubenswrapper[4824]: I1006 10:14:25.948508 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Oct 06 10:14:26 crc kubenswrapper[4824]: I1006 10:14:26.017437 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-77kzd"] Oct 06 10:14:26 crc kubenswrapper[4824]: I1006 10:14:26.043914 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/419276d8-4ecf-4133-a98e-6e9373438668-config\") pod \"ovn-controller-metrics-77kzd\" (UID: \"419276d8-4ecf-4133-a98e-6e9373438668\") " pod="openstack/ovn-controller-metrics-77kzd" Oct 06 10:14:26 crc kubenswrapper[4824]: I1006 10:14:26.043994 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e02ad6fa-12f7-4c9e-897f-e6fa39d41372-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-7md89\" (UID: \"e02ad6fa-12f7-4c9e-897f-e6fa39d41372\") " pod="openstack/dnsmasq-dns-7fd796d7df-7md89" Oct 06 10:14:26 crc kubenswrapper[4824]: I1006 10:14:26.044035 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e02ad6fa-12f7-4c9e-897f-e6fa39d41372-config\") pod \"dnsmasq-dns-7fd796d7df-7md89\" (UID: \"e02ad6fa-12f7-4c9e-897f-e6fa39d41372\") " pod="openstack/dnsmasq-dns-7fd796d7df-7md89" Oct 06 10:14:26 crc kubenswrapper[4824]: I1006 10:14:26.044056 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/419276d8-4ecf-4133-a98e-6e9373438668-ovs-rundir\") pod \"ovn-controller-metrics-77kzd\" (UID: \"419276d8-4ecf-4133-a98e-6e9373438668\") " pod="openstack/ovn-controller-metrics-77kzd" Oct 06 10:14:26 crc kubenswrapper[4824]: I1006 10:14:26.044082 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/419276d8-4ecf-4133-a98e-6e9373438668-combined-ca-bundle\") pod \"ovn-controller-metrics-77kzd\" (UID: \"419276d8-4ecf-4133-a98e-6e9373438668\") " pod="openstack/ovn-controller-metrics-77kzd" Oct 06 10:14:26 crc kubenswrapper[4824]: I1006 10:14:26.044101 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tgfn6\" (UniqueName: \"kubernetes.io/projected/419276d8-4ecf-4133-a98e-6e9373438668-kube-api-access-tgfn6\") pod \"ovn-controller-metrics-77kzd\" (UID: \"419276d8-4ecf-4133-a98e-6e9373438668\") " pod="openstack/ovn-controller-metrics-77kzd" Oct 06 10:14:26 crc kubenswrapper[4824]: I1006 10:14:26.044134 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/419276d8-4ecf-4133-a98e-6e9373438668-ovn-rundir\") pod \"ovn-controller-metrics-77kzd\" (UID: \"419276d8-4ecf-4133-a98e-6e9373438668\") " pod="openstack/ovn-controller-metrics-77kzd" Oct 06 10:14:26 crc kubenswrapper[4824]: I1006 10:14:26.044163 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/419276d8-4ecf-4133-a98e-6e9373438668-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-77kzd\" (UID: \"419276d8-4ecf-4133-a98e-6e9373438668\") " pod="openstack/ovn-controller-metrics-77kzd" Oct 06 10:14:26 crc kubenswrapper[4824]: I1006 10:14:26.044183 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e02ad6fa-12f7-4c9e-897f-e6fa39d41372-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-7md89\" (UID: \"e02ad6fa-12f7-4c9e-897f-e6fa39d41372\") " pod="openstack/dnsmasq-dns-7fd796d7df-7md89" Oct 06 10:14:26 crc kubenswrapper[4824]: I1006 10:14:26.044218 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4fk8r\" (UniqueName: \"kubernetes.io/projected/e02ad6fa-12f7-4c9e-897f-e6fa39d41372-kube-api-access-4fk8r\") pod \"dnsmasq-dns-7fd796d7df-7md89\" (UID: \"e02ad6fa-12f7-4c9e-897f-e6fa39d41372\") " pod="openstack/dnsmasq-dns-7fd796d7df-7md89" Oct 06 10:14:26 crc kubenswrapper[4824]: I1006 10:14:26.145808 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/419276d8-4ecf-4133-a98e-6e9373438668-ovn-rundir\") pod \"ovn-controller-metrics-77kzd\" (UID: \"419276d8-4ecf-4133-a98e-6e9373438668\") " pod="openstack/ovn-controller-metrics-77kzd" Oct 06 10:14:26 crc kubenswrapper[4824]: I1006 10:14:26.145874 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/419276d8-4ecf-4133-a98e-6e9373438668-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-77kzd\" (UID: \"419276d8-4ecf-4133-a98e-6e9373438668\") " pod="openstack/ovn-controller-metrics-77kzd" Oct 06 10:14:26 crc kubenswrapper[4824]: I1006 10:14:26.145895 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e02ad6fa-12f7-4c9e-897f-e6fa39d41372-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-7md89\" (UID: \"e02ad6fa-12f7-4c9e-897f-e6fa39d41372\") " pod="openstack/dnsmasq-dns-7fd796d7df-7md89" Oct 06 10:14:26 crc kubenswrapper[4824]: I1006 10:14:26.145934 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4fk8r\" (UniqueName: \"kubernetes.io/projected/e02ad6fa-12f7-4c9e-897f-e6fa39d41372-kube-api-access-4fk8r\") pod \"dnsmasq-dns-7fd796d7df-7md89\" (UID: \"e02ad6fa-12f7-4c9e-897f-e6fa39d41372\") " pod="openstack/dnsmasq-dns-7fd796d7df-7md89" Oct 06 10:14:26 crc kubenswrapper[4824]: I1006 10:14:26.145968 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/419276d8-4ecf-4133-a98e-6e9373438668-config\") pod \"ovn-controller-metrics-77kzd\" (UID: \"419276d8-4ecf-4133-a98e-6e9373438668\") " pod="openstack/ovn-controller-metrics-77kzd" Oct 06 10:14:26 crc kubenswrapper[4824]: I1006 10:14:26.146138 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e02ad6fa-12f7-4c9e-897f-e6fa39d41372-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-7md89\" (UID: \"e02ad6fa-12f7-4c9e-897f-e6fa39d41372\") " pod="openstack/dnsmasq-dns-7fd796d7df-7md89" Oct 06 10:14:26 crc kubenswrapper[4824]: I1006 10:14:26.146171 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e02ad6fa-12f7-4c9e-897f-e6fa39d41372-config\") pod \"dnsmasq-dns-7fd796d7df-7md89\" (UID: \"e02ad6fa-12f7-4c9e-897f-e6fa39d41372\") " pod="openstack/dnsmasq-dns-7fd796d7df-7md89" Oct 06 10:14:26 crc kubenswrapper[4824]: I1006 10:14:26.146193 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/419276d8-4ecf-4133-a98e-6e9373438668-ovs-rundir\") pod \"ovn-controller-metrics-77kzd\" (UID: \"419276d8-4ecf-4133-a98e-6e9373438668\") " pod="openstack/ovn-controller-metrics-77kzd" Oct 06 10:14:26 crc kubenswrapper[4824]: I1006 10:14:26.146211 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/419276d8-4ecf-4133-a98e-6e9373438668-combined-ca-bundle\") pod \"ovn-controller-metrics-77kzd\" (UID: \"419276d8-4ecf-4133-a98e-6e9373438668\") " pod="openstack/ovn-controller-metrics-77kzd" Oct 06 10:14:26 crc kubenswrapper[4824]: I1006 10:14:26.146229 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tgfn6\" (UniqueName: \"kubernetes.io/projected/419276d8-4ecf-4133-a98e-6e9373438668-kube-api-access-tgfn6\") pod \"ovn-controller-metrics-77kzd\" (UID: \"419276d8-4ecf-4133-a98e-6e9373438668\") " pod="openstack/ovn-controller-metrics-77kzd" Oct 06 10:14:26 crc kubenswrapper[4824]: I1006 10:14:26.146904 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/419276d8-4ecf-4133-a98e-6e9373438668-ovn-rundir\") pod \"ovn-controller-metrics-77kzd\" (UID: \"419276d8-4ecf-4133-a98e-6e9373438668\") " pod="openstack/ovn-controller-metrics-77kzd" Oct 06 10:14:26 crc kubenswrapper[4824]: I1006 10:14:26.147470 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/419276d8-4ecf-4133-a98e-6e9373438668-config\") pod \"ovn-controller-metrics-77kzd\" (UID: \"419276d8-4ecf-4133-a98e-6e9373438668\") " pod="openstack/ovn-controller-metrics-77kzd" Oct 06 10:14:26 crc kubenswrapper[4824]: I1006 10:14:26.147796 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e02ad6fa-12f7-4c9e-897f-e6fa39d41372-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-7md89\" (UID: \"e02ad6fa-12f7-4c9e-897f-e6fa39d41372\") " pod="openstack/dnsmasq-dns-7fd796d7df-7md89" Oct 06 10:14:26 crc kubenswrapper[4824]: I1006 10:14:26.148563 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e02ad6fa-12f7-4c9e-897f-e6fa39d41372-config\") pod \"dnsmasq-dns-7fd796d7df-7md89\" (UID: \"e02ad6fa-12f7-4c9e-897f-e6fa39d41372\") " pod="openstack/dnsmasq-dns-7fd796d7df-7md89" Oct 06 10:14:26 crc kubenswrapper[4824]: I1006 10:14:26.148648 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/419276d8-4ecf-4133-a98e-6e9373438668-ovs-rundir\") pod \"ovn-controller-metrics-77kzd\" (UID: \"419276d8-4ecf-4133-a98e-6e9373438668\") " pod="openstack/ovn-controller-metrics-77kzd" Oct 06 10:14:26 crc kubenswrapper[4824]: I1006 10:14:26.148867 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e02ad6fa-12f7-4c9e-897f-e6fa39d41372-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-7md89\" (UID: \"e02ad6fa-12f7-4c9e-897f-e6fa39d41372\") " pod="openstack/dnsmasq-dns-7fd796d7df-7md89" Oct 06 10:14:26 crc kubenswrapper[4824]: I1006 10:14:26.161076 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/419276d8-4ecf-4133-a98e-6e9373438668-combined-ca-bundle\") pod \"ovn-controller-metrics-77kzd\" (UID: \"419276d8-4ecf-4133-a98e-6e9373438668\") " pod="openstack/ovn-controller-metrics-77kzd" Oct 06 10:14:26 crc kubenswrapper[4824]: I1006 10:14:26.175059 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Oct 06 10:14:26 crc kubenswrapper[4824]: I1006 10:14:26.176429 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 06 10:14:26 crc kubenswrapper[4824]: I1006 10:14:26.181672 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/419276d8-4ecf-4133-a98e-6e9373438668-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-77kzd\" (UID: \"419276d8-4ecf-4133-a98e-6e9373438668\") " pod="openstack/ovn-controller-metrics-77kzd" Oct 06 10:14:26 crc kubenswrapper[4824]: I1006 10:14:26.188115 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-7md89"] Oct 06 10:14:26 crc kubenswrapper[4824]: E1006 10:14:26.188880 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[kube-api-access-4fk8r], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/dnsmasq-dns-7fd796d7df-7md89" podUID="e02ad6fa-12f7-4c9e-897f-e6fa39d41372" Oct 06 10:14:26 crc kubenswrapper[4824]: I1006 10:14:26.194302 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Oct 06 10:14:26 crc kubenswrapper[4824]: I1006 10:14:26.194392 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Oct 06 10:14:26 crc kubenswrapper[4824]: I1006 10:14:26.194806 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Oct 06 10:14:26 crc kubenswrapper[4824]: I1006 10:14:26.194866 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tgfn6\" (UniqueName: \"kubernetes.io/projected/419276d8-4ecf-4133-a98e-6e9373438668-kube-api-access-tgfn6\") pod \"ovn-controller-metrics-77kzd\" (UID: \"419276d8-4ecf-4133-a98e-6e9373438668\") " pod="openstack/ovn-controller-metrics-77kzd" Oct 06 10:14:26 crc kubenswrapper[4824]: I1006 10:14:26.200880 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4fk8r\" (UniqueName: \"kubernetes.io/projected/e02ad6fa-12f7-4c9e-897f-e6fa39d41372-kube-api-access-4fk8r\") pod \"dnsmasq-dns-7fd796d7df-7md89\" (UID: \"e02ad6fa-12f7-4c9e-897f-e6fa39d41372\") " pod="openstack/dnsmasq-dns-7fd796d7df-7md89" Oct 06 10:14:26 crc kubenswrapper[4824]: I1006 10:14:26.202703 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-8m656" Oct 06 10:14:26 crc kubenswrapper[4824]: I1006 10:14:26.202973 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Oct 06 10:14:26 crc kubenswrapper[4824]: I1006 10:14:26.268580 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-cbp2v"] Oct 06 10:14:26 crc kubenswrapper[4824]: I1006 10:14:26.270464 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-cbp2v" Oct 06 10:14:26 crc kubenswrapper[4824]: I1006 10:14:26.279432 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-77kzd" Oct 06 10:14:26 crc kubenswrapper[4824]: I1006 10:14:26.279449 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Oct 06 10:14:26 crc kubenswrapper[4824]: I1006 10:14:26.316828 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-cbp2v"] Oct 06 10:14:26 crc kubenswrapper[4824]: I1006 10:14:26.351762 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/43002b2d-db5d-488e-9791-56388700cf9f-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"43002b2d-db5d-488e-9791-56388700cf9f\") " pod="openstack/ovn-northd-0" Oct 06 10:14:26 crc kubenswrapper[4824]: I1006 10:14:26.351857 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tz6gw\" (UniqueName: \"kubernetes.io/projected/43002b2d-db5d-488e-9791-56388700cf9f-kube-api-access-tz6gw\") pod \"ovn-northd-0\" (UID: \"43002b2d-db5d-488e-9791-56388700cf9f\") " pod="openstack/ovn-northd-0" Oct 06 10:14:26 crc kubenswrapper[4824]: I1006 10:14:26.351909 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c6359046-e4df-4c51-b082-85650b021d4e-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-cbp2v\" (UID: \"c6359046-e4df-4c51-b082-85650b021d4e\") " pod="openstack/dnsmasq-dns-86db49b7ff-cbp2v" Oct 06 10:14:26 crc kubenswrapper[4824]: I1006 10:14:26.351955 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-54jlr\" (UniqueName: \"kubernetes.io/projected/c6359046-e4df-4c51-b082-85650b021d4e-kube-api-access-54jlr\") pod \"dnsmasq-dns-86db49b7ff-cbp2v\" (UID: \"c6359046-e4df-4c51-b082-85650b021d4e\") " pod="openstack/dnsmasq-dns-86db49b7ff-cbp2v" Oct 06 10:14:26 crc kubenswrapper[4824]: I1006 10:14:26.351991 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c6359046-e4df-4c51-b082-85650b021d4e-config\") pod \"dnsmasq-dns-86db49b7ff-cbp2v\" (UID: \"c6359046-e4df-4c51-b082-85650b021d4e\") " pod="openstack/dnsmasq-dns-86db49b7ff-cbp2v" Oct 06 10:14:26 crc kubenswrapper[4824]: I1006 10:14:26.352082 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c6359046-e4df-4c51-b082-85650b021d4e-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-cbp2v\" (UID: \"c6359046-e4df-4c51-b082-85650b021d4e\") " pod="openstack/dnsmasq-dns-86db49b7ff-cbp2v" Oct 06 10:14:26 crc kubenswrapper[4824]: I1006 10:14:26.352121 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/43002b2d-db5d-488e-9791-56388700cf9f-config\") pod \"ovn-northd-0\" (UID: \"43002b2d-db5d-488e-9791-56388700cf9f\") " pod="openstack/ovn-northd-0" Oct 06 10:14:26 crc kubenswrapper[4824]: I1006 10:14:26.352156 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43002b2d-db5d-488e-9791-56388700cf9f-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"43002b2d-db5d-488e-9791-56388700cf9f\") " pod="openstack/ovn-northd-0" Oct 06 10:14:26 crc kubenswrapper[4824]: I1006 10:14:26.352181 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/43002b2d-db5d-488e-9791-56388700cf9f-scripts\") pod \"ovn-northd-0\" (UID: \"43002b2d-db5d-488e-9791-56388700cf9f\") " pod="openstack/ovn-northd-0" Oct 06 10:14:26 crc kubenswrapper[4824]: I1006 10:14:26.352203 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/43002b2d-db5d-488e-9791-56388700cf9f-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"43002b2d-db5d-488e-9791-56388700cf9f\") " pod="openstack/ovn-northd-0" Oct 06 10:14:26 crc kubenswrapper[4824]: I1006 10:14:26.352222 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c6359046-e4df-4c51-b082-85650b021d4e-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-cbp2v\" (UID: \"c6359046-e4df-4c51-b082-85650b021d4e\") " pod="openstack/dnsmasq-dns-86db49b7ff-cbp2v" Oct 06 10:14:26 crc kubenswrapper[4824]: I1006 10:14:26.352259 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/43002b2d-db5d-488e-9791-56388700cf9f-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"43002b2d-db5d-488e-9791-56388700cf9f\") " pod="openstack/ovn-northd-0" Oct 06 10:14:26 crc kubenswrapper[4824]: I1006 10:14:26.453277 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/43002b2d-db5d-488e-9791-56388700cf9f-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"43002b2d-db5d-488e-9791-56388700cf9f\") " pod="openstack/ovn-northd-0" Oct 06 10:14:26 crc kubenswrapper[4824]: I1006 10:14:26.453458 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/43002b2d-db5d-488e-9791-56388700cf9f-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"43002b2d-db5d-488e-9791-56388700cf9f\") " pod="openstack/ovn-northd-0" Oct 06 10:14:26 crc kubenswrapper[4824]: I1006 10:14:26.453487 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tz6gw\" (UniqueName: \"kubernetes.io/projected/43002b2d-db5d-488e-9791-56388700cf9f-kube-api-access-tz6gw\") pod \"ovn-northd-0\" (UID: \"43002b2d-db5d-488e-9791-56388700cf9f\") " pod="openstack/ovn-northd-0" Oct 06 10:14:26 crc kubenswrapper[4824]: I1006 10:14:26.453516 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c6359046-e4df-4c51-b082-85650b021d4e-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-cbp2v\" (UID: \"c6359046-e4df-4c51-b082-85650b021d4e\") " pod="openstack/dnsmasq-dns-86db49b7ff-cbp2v" Oct 06 10:14:26 crc kubenswrapper[4824]: I1006 10:14:26.453549 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-54jlr\" (UniqueName: \"kubernetes.io/projected/c6359046-e4df-4c51-b082-85650b021d4e-kube-api-access-54jlr\") pod \"dnsmasq-dns-86db49b7ff-cbp2v\" (UID: \"c6359046-e4df-4c51-b082-85650b021d4e\") " pod="openstack/dnsmasq-dns-86db49b7ff-cbp2v" Oct 06 10:14:26 crc kubenswrapper[4824]: I1006 10:14:26.453570 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c6359046-e4df-4c51-b082-85650b021d4e-config\") pod \"dnsmasq-dns-86db49b7ff-cbp2v\" (UID: \"c6359046-e4df-4c51-b082-85650b021d4e\") " pod="openstack/dnsmasq-dns-86db49b7ff-cbp2v" Oct 06 10:14:26 crc kubenswrapper[4824]: I1006 10:14:26.453612 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c6359046-e4df-4c51-b082-85650b021d4e-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-cbp2v\" (UID: \"c6359046-e4df-4c51-b082-85650b021d4e\") " pod="openstack/dnsmasq-dns-86db49b7ff-cbp2v" Oct 06 10:14:26 crc kubenswrapper[4824]: I1006 10:14:26.453634 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/43002b2d-db5d-488e-9791-56388700cf9f-config\") pod \"ovn-northd-0\" (UID: \"43002b2d-db5d-488e-9791-56388700cf9f\") " pod="openstack/ovn-northd-0" Oct 06 10:14:26 crc kubenswrapper[4824]: I1006 10:14:26.453656 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43002b2d-db5d-488e-9791-56388700cf9f-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"43002b2d-db5d-488e-9791-56388700cf9f\") " pod="openstack/ovn-northd-0" Oct 06 10:14:26 crc kubenswrapper[4824]: I1006 10:14:26.453677 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/43002b2d-db5d-488e-9791-56388700cf9f-scripts\") pod \"ovn-northd-0\" (UID: \"43002b2d-db5d-488e-9791-56388700cf9f\") " pod="openstack/ovn-northd-0" Oct 06 10:14:26 crc kubenswrapper[4824]: I1006 10:14:26.453695 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c6359046-e4df-4c51-b082-85650b021d4e-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-cbp2v\" (UID: \"c6359046-e4df-4c51-b082-85650b021d4e\") " pod="openstack/dnsmasq-dns-86db49b7ff-cbp2v" Oct 06 10:14:26 crc kubenswrapper[4824]: I1006 10:14:26.453720 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/43002b2d-db5d-488e-9791-56388700cf9f-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"43002b2d-db5d-488e-9791-56388700cf9f\") " pod="openstack/ovn-northd-0" Oct 06 10:14:26 crc kubenswrapper[4824]: I1006 10:14:26.454922 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c6359046-e4df-4c51-b082-85650b021d4e-config\") pod \"dnsmasq-dns-86db49b7ff-cbp2v\" (UID: \"c6359046-e4df-4c51-b082-85650b021d4e\") " pod="openstack/dnsmasq-dns-86db49b7ff-cbp2v" Oct 06 10:14:26 crc kubenswrapper[4824]: I1006 10:14:26.455527 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c6359046-e4df-4c51-b082-85650b021d4e-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-cbp2v\" (UID: \"c6359046-e4df-4c51-b082-85650b021d4e\") " pod="openstack/dnsmasq-dns-86db49b7ff-cbp2v" Oct 06 10:14:26 crc kubenswrapper[4824]: I1006 10:14:26.455631 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/43002b2d-db5d-488e-9791-56388700cf9f-scripts\") pod \"ovn-northd-0\" (UID: \"43002b2d-db5d-488e-9791-56388700cf9f\") " pod="openstack/ovn-northd-0" Oct 06 10:14:26 crc kubenswrapper[4824]: I1006 10:14:26.456174 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/43002b2d-db5d-488e-9791-56388700cf9f-config\") pod \"ovn-northd-0\" (UID: \"43002b2d-db5d-488e-9791-56388700cf9f\") " pod="openstack/ovn-northd-0" Oct 06 10:14:26 crc kubenswrapper[4824]: I1006 10:14:26.458795 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c6359046-e4df-4c51-b082-85650b021d4e-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-cbp2v\" (UID: \"c6359046-e4df-4c51-b082-85650b021d4e\") " pod="openstack/dnsmasq-dns-86db49b7ff-cbp2v" Oct 06 10:14:26 crc kubenswrapper[4824]: I1006 10:14:26.458827 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c6359046-e4df-4c51-b082-85650b021d4e-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-cbp2v\" (UID: \"c6359046-e4df-4c51-b082-85650b021d4e\") " pod="openstack/dnsmasq-dns-86db49b7ff-cbp2v" Oct 06 10:14:26 crc kubenswrapper[4824]: I1006 10:14:26.459163 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/43002b2d-db5d-488e-9791-56388700cf9f-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"43002b2d-db5d-488e-9791-56388700cf9f\") " pod="openstack/ovn-northd-0" Oct 06 10:14:26 crc kubenswrapper[4824]: I1006 10:14:26.461459 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43002b2d-db5d-488e-9791-56388700cf9f-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"43002b2d-db5d-488e-9791-56388700cf9f\") " pod="openstack/ovn-northd-0" Oct 06 10:14:26 crc kubenswrapper[4824]: I1006 10:14:26.464169 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/43002b2d-db5d-488e-9791-56388700cf9f-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"43002b2d-db5d-488e-9791-56388700cf9f\") " pod="openstack/ovn-northd-0" Oct 06 10:14:26 crc kubenswrapper[4824]: I1006 10:14:26.465672 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/43002b2d-db5d-488e-9791-56388700cf9f-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"43002b2d-db5d-488e-9791-56388700cf9f\") " pod="openstack/ovn-northd-0" Oct 06 10:14:26 crc kubenswrapper[4824]: I1006 10:14:26.471666 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tz6gw\" (UniqueName: \"kubernetes.io/projected/43002b2d-db5d-488e-9791-56388700cf9f-kube-api-access-tz6gw\") pod \"ovn-northd-0\" (UID: \"43002b2d-db5d-488e-9791-56388700cf9f\") " pod="openstack/ovn-northd-0" Oct 06 10:14:26 crc kubenswrapper[4824]: I1006 10:14:26.474674 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-54jlr\" (UniqueName: \"kubernetes.io/projected/c6359046-e4df-4c51-b082-85650b021d4e-kube-api-access-54jlr\") pod \"dnsmasq-dns-86db49b7ff-cbp2v\" (UID: \"c6359046-e4df-4c51-b082-85650b021d4e\") " pod="openstack/dnsmasq-dns-86db49b7ff-cbp2v" Oct 06 10:14:26 crc kubenswrapper[4824]: I1006 10:14:26.542362 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-7md89" Oct 06 10:14:26 crc kubenswrapper[4824]: I1006 10:14:26.553797 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-7md89" Oct 06 10:14:26 crc kubenswrapper[4824]: I1006 10:14:26.620449 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 06 10:14:26 crc kubenswrapper[4824]: I1006 10:14:26.656799 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e02ad6fa-12f7-4c9e-897f-e6fa39d41372-dns-svc\") pod \"e02ad6fa-12f7-4c9e-897f-e6fa39d41372\" (UID: \"e02ad6fa-12f7-4c9e-897f-e6fa39d41372\") " Oct 06 10:14:26 crc kubenswrapper[4824]: I1006 10:14:26.656949 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4fk8r\" (UniqueName: \"kubernetes.io/projected/e02ad6fa-12f7-4c9e-897f-e6fa39d41372-kube-api-access-4fk8r\") pod \"e02ad6fa-12f7-4c9e-897f-e6fa39d41372\" (UID: \"e02ad6fa-12f7-4c9e-897f-e6fa39d41372\") " Oct 06 10:14:26 crc kubenswrapper[4824]: I1006 10:14:26.657193 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e02ad6fa-12f7-4c9e-897f-e6fa39d41372-ovsdbserver-nb\") pod \"e02ad6fa-12f7-4c9e-897f-e6fa39d41372\" (UID: \"e02ad6fa-12f7-4c9e-897f-e6fa39d41372\") " Oct 06 10:14:26 crc kubenswrapper[4824]: I1006 10:14:26.657257 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e02ad6fa-12f7-4c9e-897f-e6fa39d41372-config\") pod \"e02ad6fa-12f7-4c9e-897f-e6fa39d41372\" (UID: \"e02ad6fa-12f7-4c9e-897f-e6fa39d41372\") " Oct 06 10:14:26 crc kubenswrapper[4824]: I1006 10:14:26.657458 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e02ad6fa-12f7-4c9e-897f-e6fa39d41372-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "e02ad6fa-12f7-4c9e-897f-e6fa39d41372" (UID: "e02ad6fa-12f7-4c9e-897f-e6fa39d41372"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:14:26 crc kubenswrapper[4824]: I1006 10:14:26.657806 4824 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e02ad6fa-12f7-4c9e-897f-e6fa39d41372-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 06 10:14:26 crc kubenswrapper[4824]: I1006 10:14:26.659512 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e02ad6fa-12f7-4c9e-897f-e6fa39d41372-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "e02ad6fa-12f7-4c9e-897f-e6fa39d41372" (UID: "e02ad6fa-12f7-4c9e-897f-e6fa39d41372"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:14:26 crc kubenswrapper[4824]: I1006 10:14:26.659980 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e02ad6fa-12f7-4c9e-897f-e6fa39d41372-config" (OuterVolumeSpecName: "config") pod "e02ad6fa-12f7-4c9e-897f-e6fa39d41372" (UID: "e02ad6fa-12f7-4c9e-897f-e6fa39d41372"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:14:26 crc kubenswrapper[4824]: I1006 10:14:26.668108 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e02ad6fa-12f7-4c9e-897f-e6fa39d41372-kube-api-access-4fk8r" (OuterVolumeSpecName: "kube-api-access-4fk8r") pod "e02ad6fa-12f7-4c9e-897f-e6fa39d41372" (UID: "e02ad6fa-12f7-4c9e-897f-e6fa39d41372"). InnerVolumeSpecName "kube-api-access-4fk8r". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:14:26 crc kubenswrapper[4824]: I1006 10:14:26.707511 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-cbp2v" Oct 06 10:14:26 crc kubenswrapper[4824]: I1006 10:14:26.759463 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e02ad6fa-12f7-4c9e-897f-e6fa39d41372-config\") on node \"crc\" DevicePath \"\"" Oct 06 10:14:26 crc kubenswrapper[4824]: I1006 10:14:26.759498 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4fk8r\" (UniqueName: \"kubernetes.io/projected/e02ad6fa-12f7-4c9e-897f-e6fa39d41372-kube-api-access-4fk8r\") on node \"crc\" DevicePath \"\"" Oct 06 10:14:26 crc kubenswrapper[4824]: I1006 10:14:26.759509 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e02ad6fa-12f7-4c9e-897f-e6fa39d41372-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 06 10:14:26 crc kubenswrapper[4824]: I1006 10:14:26.864047 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-77kzd"] Oct 06 10:14:27 crc kubenswrapper[4824]: I1006 10:14:27.072321 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Oct 06 10:14:27 crc kubenswrapper[4824]: I1006 10:14:27.072848 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Oct 06 10:14:27 crc kubenswrapper[4824]: I1006 10:14:27.081560 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Oct 06 10:14:27 crc kubenswrapper[4824]: W1006 10:14:27.091868 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod43002b2d_db5d_488e_9791_56388700cf9f.slice/crio-aeb97aad048417e988e19a2e64882e4b608c5b8c7a9645731db2f9d0e582aa0b WatchSource:0}: Error finding container aeb97aad048417e988e19a2e64882e4b608c5b8c7a9645731db2f9d0e582aa0b: Status 404 returned error can't find the container with id aeb97aad048417e988e19a2e64882e4b608c5b8c7a9645731db2f9d0e582aa0b Oct 06 10:14:27 crc kubenswrapper[4824]: W1006 10:14:27.288700 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc6359046_e4df_4c51_b082_85650b021d4e.slice/crio-413132bee0e4b3ea37e95a833fa533ffce6da5fb0a1bd2d5004db70a87cc4baf WatchSource:0}: Error finding container 413132bee0e4b3ea37e95a833fa533ffce6da5fb0a1bd2d5004db70a87cc4baf: Status 404 returned error can't find the container with id 413132bee0e4b3ea37e95a833fa533ffce6da5fb0a1bd2d5004db70a87cc4baf Oct 06 10:14:27 crc kubenswrapper[4824]: I1006 10:14:27.291804 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-cbp2v"] Oct 06 10:14:27 crc kubenswrapper[4824]: I1006 10:14:27.488736 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Oct 06 10:14:27 crc kubenswrapper[4824]: I1006 10:14:27.488783 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Oct 06 10:14:27 crc kubenswrapper[4824]: I1006 10:14:27.552904 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"43002b2d-db5d-488e-9791-56388700cf9f","Type":"ContainerStarted","Data":"aeb97aad048417e988e19a2e64882e4b608c5b8c7a9645731db2f9d0e582aa0b"} Oct 06 10:14:27 crc kubenswrapper[4824]: I1006 10:14:27.554783 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-77kzd" event={"ID":"419276d8-4ecf-4133-a98e-6e9373438668","Type":"ContainerStarted","Data":"e1f20aae5d3251dd829f21d84ba7365c433102de8584384763a4527c3b9df445"} Oct 06 10:14:27 crc kubenswrapper[4824]: I1006 10:14:27.556184 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-7md89" Oct 06 10:14:27 crc kubenswrapper[4824]: I1006 10:14:27.557110 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-cbp2v" event={"ID":"c6359046-e4df-4c51-b082-85650b021d4e","Type":"ContainerStarted","Data":"413132bee0e4b3ea37e95a833fa533ffce6da5fb0a1bd2d5004db70a87cc4baf"} Oct 06 10:14:27 crc kubenswrapper[4824]: I1006 10:14:27.603453 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-7md89"] Oct 06 10:14:27 crc kubenswrapper[4824]: I1006 10:14:27.609329 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-7md89"] Oct 06 10:14:29 crc kubenswrapper[4824]: I1006 10:14:29.285185 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e02ad6fa-12f7-4c9e-897f-e6fa39d41372" path="/var/lib/kubelet/pods/e02ad6fa-12f7-4c9e-897f-e6fa39d41372/volumes" Oct 06 10:14:29 crc kubenswrapper[4824]: I1006 10:14:29.406276 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Oct 06 10:14:29 crc kubenswrapper[4824]: I1006 10:14:29.482102 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-cbp2v"] Oct 06 10:14:29 crc kubenswrapper[4824]: I1006 10:14:29.530538 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-698758b865-cwvfc"] Oct 06 10:14:29 crc kubenswrapper[4824]: I1006 10:14:29.531826 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-cwvfc" Oct 06 10:14:29 crc kubenswrapper[4824]: I1006 10:14:29.563399 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-cwvfc"] Oct 06 10:14:29 crc kubenswrapper[4824]: I1006 10:14:29.606498 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b4ac2488-c99e-4380-8e82-e89207990c45-config\") pod \"dnsmasq-dns-698758b865-cwvfc\" (UID: \"b4ac2488-c99e-4380-8e82-e89207990c45\") " pod="openstack/dnsmasq-dns-698758b865-cwvfc" Oct 06 10:14:29 crc kubenswrapper[4824]: I1006 10:14:29.606597 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b4ac2488-c99e-4380-8e82-e89207990c45-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-cwvfc\" (UID: \"b4ac2488-c99e-4380-8e82-e89207990c45\") " pod="openstack/dnsmasq-dns-698758b865-cwvfc" Oct 06 10:14:29 crc kubenswrapper[4824]: I1006 10:14:29.606633 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vwmp5\" (UniqueName: \"kubernetes.io/projected/b4ac2488-c99e-4380-8e82-e89207990c45-kube-api-access-vwmp5\") pod \"dnsmasq-dns-698758b865-cwvfc\" (UID: \"b4ac2488-c99e-4380-8e82-e89207990c45\") " pod="openstack/dnsmasq-dns-698758b865-cwvfc" Oct 06 10:14:29 crc kubenswrapper[4824]: I1006 10:14:29.606660 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b4ac2488-c99e-4380-8e82-e89207990c45-dns-svc\") pod \"dnsmasq-dns-698758b865-cwvfc\" (UID: \"b4ac2488-c99e-4380-8e82-e89207990c45\") " pod="openstack/dnsmasq-dns-698758b865-cwvfc" Oct 06 10:14:29 crc kubenswrapper[4824]: I1006 10:14:29.606697 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b4ac2488-c99e-4380-8e82-e89207990c45-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-cwvfc\" (UID: \"b4ac2488-c99e-4380-8e82-e89207990c45\") " pod="openstack/dnsmasq-dns-698758b865-cwvfc" Oct 06 10:14:29 crc kubenswrapper[4824]: I1006 10:14:29.708137 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vwmp5\" (UniqueName: \"kubernetes.io/projected/b4ac2488-c99e-4380-8e82-e89207990c45-kube-api-access-vwmp5\") pod \"dnsmasq-dns-698758b865-cwvfc\" (UID: \"b4ac2488-c99e-4380-8e82-e89207990c45\") " pod="openstack/dnsmasq-dns-698758b865-cwvfc" Oct 06 10:14:29 crc kubenswrapper[4824]: I1006 10:14:29.708199 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b4ac2488-c99e-4380-8e82-e89207990c45-dns-svc\") pod \"dnsmasq-dns-698758b865-cwvfc\" (UID: \"b4ac2488-c99e-4380-8e82-e89207990c45\") " pod="openstack/dnsmasq-dns-698758b865-cwvfc" Oct 06 10:14:29 crc kubenswrapper[4824]: I1006 10:14:29.708244 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b4ac2488-c99e-4380-8e82-e89207990c45-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-cwvfc\" (UID: \"b4ac2488-c99e-4380-8e82-e89207990c45\") " pod="openstack/dnsmasq-dns-698758b865-cwvfc" Oct 06 10:14:29 crc kubenswrapper[4824]: I1006 10:14:29.708306 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b4ac2488-c99e-4380-8e82-e89207990c45-config\") pod \"dnsmasq-dns-698758b865-cwvfc\" (UID: \"b4ac2488-c99e-4380-8e82-e89207990c45\") " pod="openstack/dnsmasq-dns-698758b865-cwvfc" Oct 06 10:14:29 crc kubenswrapper[4824]: I1006 10:14:29.708369 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b4ac2488-c99e-4380-8e82-e89207990c45-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-cwvfc\" (UID: \"b4ac2488-c99e-4380-8e82-e89207990c45\") " pod="openstack/dnsmasq-dns-698758b865-cwvfc" Oct 06 10:14:29 crc kubenswrapper[4824]: I1006 10:14:29.709419 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b4ac2488-c99e-4380-8e82-e89207990c45-dns-svc\") pod \"dnsmasq-dns-698758b865-cwvfc\" (UID: \"b4ac2488-c99e-4380-8e82-e89207990c45\") " pod="openstack/dnsmasq-dns-698758b865-cwvfc" Oct 06 10:14:29 crc kubenswrapper[4824]: I1006 10:14:29.709468 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b4ac2488-c99e-4380-8e82-e89207990c45-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-cwvfc\" (UID: \"b4ac2488-c99e-4380-8e82-e89207990c45\") " pod="openstack/dnsmasq-dns-698758b865-cwvfc" Oct 06 10:14:29 crc kubenswrapper[4824]: I1006 10:14:29.709476 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b4ac2488-c99e-4380-8e82-e89207990c45-config\") pod \"dnsmasq-dns-698758b865-cwvfc\" (UID: \"b4ac2488-c99e-4380-8e82-e89207990c45\") " pod="openstack/dnsmasq-dns-698758b865-cwvfc" Oct 06 10:14:29 crc kubenswrapper[4824]: I1006 10:14:29.709676 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b4ac2488-c99e-4380-8e82-e89207990c45-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-cwvfc\" (UID: \"b4ac2488-c99e-4380-8e82-e89207990c45\") " pod="openstack/dnsmasq-dns-698758b865-cwvfc" Oct 06 10:14:29 crc kubenswrapper[4824]: I1006 10:14:29.737230 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vwmp5\" (UniqueName: \"kubernetes.io/projected/b4ac2488-c99e-4380-8e82-e89207990c45-kube-api-access-vwmp5\") pod \"dnsmasq-dns-698758b865-cwvfc\" (UID: \"b4ac2488-c99e-4380-8e82-e89207990c45\") " pod="openstack/dnsmasq-dns-698758b865-cwvfc" Oct 06 10:14:29 crc kubenswrapper[4824]: I1006 10:14:29.850897 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-cwvfc" Oct 06 10:14:30 crc kubenswrapper[4824]: I1006 10:14:30.318667 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-cwvfc"] Oct 06 10:14:30 crc kubenswrapper[4824]: I1006 10:14:30.563235 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Oct 06 10:14:30 crc kubenswrapper[4824]: I1006 10:14:30.593274 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Oct 06 10:14:30 crc kubenswrapper[4824]: I1006 10:14:30.593494 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Oct 06 10:14:30 crc kubenswrapper[4824]: I1006 10:14:30.595432 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-cwvfc" event={"ID":"b4ac2488-c99e-4380-8e82-e89207990c45","Type":"ContainerStarted","Data":"6c52816fe2b22eba75635e304c08a8a12f3001fef7d495eceb80a745fc7a5a63"} Oct 06 10:14:30 crc kubenswrapper[4824]: I1006 10:14:30.597180 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Oct 06 10:14:30 crc kubenswrapper[4824]: I1006 10:14:30.597369 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Oct 06 10:14:30 crc kubenswrapper[4824]: I1006 10:14:30.597410 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-cw2lp" Oct 06 10:14:30 crc kubenswrapper[4824]: I1006 10:14:30.597828 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Oct 06 10:14:30 crc kubenswrapper[4824]: I1006 10:14:30.726337 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/6cdeba93-f1c9-487d-b776-5c2f1bb353d9-etc-swift\") pod \"swift-storage-0\" (UID: \"6cdeba93-f1c9-487d-b776-5c2f1bb353d9\") " pod="openstack/swift-storage-0" Oct 06 10:14:30 crc kubenswrapper[4824]: I1006 10:14:30.726389 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"swift-storage-0\" (UID: \"6cdeba93-f1c9-487d-b776-5c2f1bb353d9\") " pod="openstack/swift-storage-0" Oct 06 10:14:30 crc kubenswrapper[4824]: I1006 10:14:30.727115 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/6cdeba93-f1c9-487d-b776-5c2f1bb353d9-cache\") pod \"swift-storage-0\" (UID: \"6cdeba93-f1c9-487d-b776-5c2f1bb353d9\") " pod="openstack/swift-storage-0" Oct 06 10:14:30 crc kubenswrapper[4824]: I1006 10:14:30.727180 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/6cdeba93-f1c9-487d-b776-5c2f1bb353d9-lock\") pod \"swift-storage-0\" (UID: \"6cdeba93-f1c9-487d-b776-5c2f1bb353d9\") " pod="openstack/swift-storage-0" Oct 06 10:14:30 crc kubenswrapper[4824]: I1006 10:14:30.727249 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bn9ns\" (UniqueName: \"kubernetes.io/projected/6cdeba93-f1c9-487d-b776-5c2f1bb353d9-kube-api-access-bn9ns\") pod \"swift-storage-0\" (UID: \"6cdeba93-f1c9-487d-b776-5c2f1bb353d9\") " pod="openstack/swift-storage-0" Oct 06 10:14:30 crc kubenswrapper[4824]: I1006 10:14:30.829410 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/6cdeba93-f1c9-487d-b776-5c2f1bb353d9-etc-swift\") pod \"swift-storage-0\" (UID: \"6cdeba93-f1c9-487d-b776-5c2f1bb353d9\") " pod="openstack/swift-storage-0" Oct 06 10:14:30 crc kubenswrapper[4824]: I1006 10:14:30.829467 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"swift-storage-0\" (UID: \"6cdeba93-f1c9-487d-b776-5c2f1bb353d9\") " pod="openstack/swift-storage-0" Oct 06 10:14:30 crc kubenswrapper[4824]: I1006 10:14:30.829559 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/6cdeba93-f1c9-487d-b776-5c2f1bb353d9-cache\") pod \"swift-storage-0\" (UID: \"6cdeba93-f1c9-487d-b776-5c2f1bb353d9\") " pod="openstack/swift-storage-0" Oct 06 10:14:30 crc kubenswrapper[4824]: I1006 10:14:30.829578 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/6cdeba93-f1c9-487d-b776-5c2f1bb353d9-lock\") pod \"swift-storage-0\" (UID: \"6cdeba93-f1c9-487d-b776-5c2f1bb353d9\") " pod="openstack/swift-storage-0" Oct 06 10:14:30 crc kubenswrapper[4824]: I1006 10:14:30.829608 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bn9ns\" (UniqueName: \"kubernetes.io/projected/6cdeba93-f1c9-487d-b776-5c2f1bb353d9-kube-api-access-bn9ns\") pod \"swift-storage-0\" (UID: \"6cdeba93-f1c9-487d-b776-5c2f1bb353d9\") " pod="openstack/swift-storage-0" Oct 06 10:14:30 crc kubenswrapper[4824]: E1006 10:14:30.829666 4824 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 06 10:14:30 crc kubenswrapper[4824]: E1006 10:14:30.829700 4824 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 06 10:14:30 crc kubenswrapper[4824]: E1006 10:14:30.829767 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/6cdeba93-f1c9-487d-b776-5c2f1bb353d9-etc-swift podName:6cdeba93-f1c9-487d-b776-5c2f1bb353d9 nodeName:}" failed. No retries permitted until 2025-10-06 10:14:31.32974608 +0000 UTC m=+980.694168941 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/6cdeba93-f1c9-487d-b776-5c2f1bb353d9-etc-swift") pod "swift-storage-0" (UID: "6cdeba93-f1c9-487d-b776-5c2f1bb353d9") : configmap "swift-ring-files" not found Oct 06 10:14:30 crc kubenswrapper[4824]: I1006 10:14:30.829975 4824 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"swift-storage-0\" (UID: \"6cdeba93-f1c9-487d-b776-5c2f1bb353d9\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/swift-storage-0" Oct 06 10:14:30 crc kubenswrapper[4824]: I1006 10:14:30.830376 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/6cdeba93-f1c9-487d-b776-5c2f1bb353d9-cache\") pod \"swift-storage-0\" (UID: \"6cdeba93-f1c9-487d-b776-5c2f1bb353d9\") " pod="openstack/swift-storage-0" Oct 06 10:14:30 crc kubenswrapper[4824]: I1006 10:14:30.830464 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/6cdeba93-f1c9-487d-b776-5c2f1bb353d9-lock\") pod \"swift-storage-0\" (UID: \"6cdeba93-f1c9-487d-b776-5c2f1bb353d9\") " pod="openstack/swift-storage-0" Oct 06 10:14:30 crc kubenswrapper[4824]: I1006 10:14:30.851686 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bn9ns\" (UniqueName: \"kubernetes.io/projected/6cdeba93-f1c9-487d-b776-5c2f1bb353d9-kube-api-access-bn9ns\") pod \"swift-storage-0\" (UID: \"6cdeba93-f1c9-487d-b776-5c2f1bb353d9\") " pod="openstack/swift-storage-0" Oct 06 10:14:30 crc kubenswrapper[4824]: I1006 10:14:30.853571 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"swift-storage-0\" (UID: \"6cdeba93-f1c9-487d-b776-5c2f1bb353d9\") " pod="openstack/swift-storage-0" Oct 06 10:14:31 crc kubenswrapper[4824]: I1006 10:14:31.338579 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/6cdeba93-f1c9-487d-b776-5c2f1bb353d9-etc-swift\") pod \"swift-storage-0\" (UID: \"6cdeba93-f1c9-487d-b776-5c2f1bb353d9\") " pod="openstack/swift-storage-0" Oct 06 10:14:31 crc kubenswrapper[4824]: E1006 10:14:31.338842 4824 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 06 10:14:31 crc kubenswrapper[4824]: E1006 10:14:31.339302 4824 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 06 10:14:31 crc kubenswrapper[4824]: E1006 10:14:31.339383 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/6cdeba93-f1c9-487d-b776-5c2f1bb353d9-etc-swift podName:6cdeba93-f1c9-487d-b776-5c2f1bb353d9 nodeName:}" failed. No retries permitted until 2025-10-06 10:14:32.339355263 +0000 UTC m=+981.703778134 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/6cdeba93-f1c9-487d-b776-5c2f1bb353d9-etc-swift") pod "swift-storage-0" (UID: "6cdeba93-f1c9-487d-b776-5c2f1bb353d9") : configmap "swift-ring-files" not found Oct 06 10:14:32 crc kubenswrapper[4824]: I1006 10:14:32.357927 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/6cdeba93-f1c9-487d-b776-5c2f1bb353d9-etc-swift\") pod \"swift-storage-0\" (UID: \"6cdeba93-f1c9-487d-b776-5c2f1bb353d9\") " pod="openstack/swift-storage-0" Oct 06 10:14:32 crc kubenswrapper[4824]: E1006 10:14:32.358104 4824 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 06 10:14:32 crc kubenswrapper[4824]: E1006 10:14:32.358374 4824 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 06 10:14:32 crc kubenswrapper[4824]: E1006 10:14:32.358426 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/6cdeba93-f1c9-487d-b776-5c2f1bb353d9-etc-swift podName:6cdeba93-f1c9-487d-b776-5c2f1bb353d9 nodeName:}" failed. No retries permitted until 2025-10-06 10:14:34.358411963 +0000 UTC m=+983.722834824 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/6cdeba93-f1c9-487d-b776-5c2f1bb353d9-etc-swift") pod "swift-storage-0" (UID: "6cdeba93-f1c9-487d-b776-5c2f1bb353d9") : configmap "swift-ring-files" not found Oct 06 10:14:32 crc kubenswrapper[4824]: I1006 10:14:32.614560 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-77kzd" event={"ID":"419276d8-4ecf-4133-a98e-6e9373438668","Type":"ContainerStarted","Data":"777bdab515b3e3dd573807c6d4281e9c0b2c963238ff5cb9831d9d008661475b"} Oct 06 10:14:33 crc kubenswrapper[4824]: I1006 10:14:33.629601 4824 generic.go:334] "Generic (PLEG): container finished" podID="c6359046-e4df-4c51-b082-85650b021d4e" containerID="234399ce46398548b58603bd742dbf7daf9d4428f1396d60c224005eac94eaa5" exitCode=0 Oct 06 10:14:33 crc kubenswrapper[4824]: I1006 10:14:33.629706 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-cbp2v" event={"ID":"c6359046-e4df-4c51-b082-85650b021d4e","Type":"ContainerDied","Data":"234399ce46398548b58603bd742dbf7daf9d4428f1396d60c224005eac94eaa5"} Oct 06 10:14:33 crc kubenswrapper[4824]: I1006 10:14:33.633387 4824 generic.go:334] "Generic (PLEG): container finished" podID="b4ac2488-c99e-4380-8e82-e89207990c45" containerID="6019be6266c0c1b65010463c73882989c76979fbcb0b846a20dc58af346d84e8" exitCode=0 Oct 06 10:14:33 crc kubenswrapper[4824]: I1006 10:14:33.633449 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-cwvfc" event={"ID":"b4ac2488-c99e-4380-8e82-e89207990c45","Type":"ContainerDied","Data":"6019be6266c0c1b65010463c73882989c76979fbcb0b846a20dc58af346d84e8"} Oct 06 10:14:33 crc kubenswrapper[4824]: I1006 10:14:33.691105 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-77kzd" podStartSLOduration=8.691085379 podStartE2EDuration="8.691085379s" podCreationTimestamp="2025-10-06 10:14:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:14:33.675847082 +0000 UTC m=+983.040269943" watchObservedRunningTime="2025-10-06 10:14:33.691085379 +0000 UTC m=+983.055508250" Oct 06 10:14:33 crc kubenswrapper[4824]: I1006 10:14:33.940695 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-cbp2v" Oct 06 10:14:34 crc kubenswrapper[4824]: I1006 10:14:34.091783 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c6359046-e4df-4c51-b082-85650b021d4e-config\") pod \"c6359046-e4df-4c51-b082-85650b021d4e\" (UID: \"c6359046-e4df-4c51-b082-85650b021d4e\") " Oct 06 10:14:34 crc kubenswrapper[4824]: I1006 10:14:34.091831 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c6359046-e4df-4c51-b082-85650b021d4e-ovsdbserver-sb\") pod \"c6359046-e4df-4c51-b082-85650b021d4e\" (UID: \"c6359046-e4df-4c51-b082-85650b021d4e\") " Oct 06 10:14:34 crc kubenswrapper[4824]: I1006 10:14:34.091857 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-54jlr\" (UniqueName: \"kubernetes.io/projected/c6359046-e4df-4c51-b082-85650b021d4e-kube-api-access-54jlr\") pod \"c6359046-e4df-4c51-b082-85650b021d4e\" (UID: \"c6359046-e4df-4c51-b082-85650b021d4e\") " Oct 06 10:14:34 crc kubenswrapper[4824]: I1006 10:14:34.091916 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c6359046-e4df-4c51-b082-85650b021d4e-ovsdbserver-nb\") pod \"c6359046-e4df-4c51-b082-85650b021d4e\" (UID: \"c6359046-e4df-4c51-b082-85650b021d4e\") " Oct 06 10:14:34 crc kubenswrapper[4824]: I1006 10:14:34.092072 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c6359046-e4df-4c51-b082-85650b021d4e-dns-svc\") pod \"c6359046-e4df-4c51-b082-85650b021d4e\" (UID: \"c6359046-e4df-4c51-b082-85650b021d4e\") " Oct 06 10:14:34 crc kubenswrapper[4824]: I1006 10:14:34.101702 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c6359046-e4df-4c51-b082-85650b021d4e-kube-api-access-54jlr" (OuterVolumeSpecName: "kube-api-access-54jlr") pod "c6359046-e4df-4c51-b082-85650b021d4e" (UID: "c6359046-e4df-4c51-b082-85650b021d4e"). InnerVolumeSpecName "kube-api-access-54jlr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:14:34 crc kubenswrapper[4824]: I1006 10:14:34.113692 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c6359046-e4df-4c51-b082-85650b021d4e-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "c6359046-e4df-4c51-b082-85650b021d4e" (UID: "c6359046-e4df-4c51-b082-85650b021d4e"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:14:34 crc kubenswrapper[4824]: I1006 10:14:34.116507 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c6359046-e4df-4c51-b082-85650b021d4e-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "c6359046-e4df-4c51-b082-85650b021d4e" (UID: "c6359046-e4df-4c51-b082-85650b021d4e"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:14:34 crc kubenswrapper[4824]: I1006 10:14:34.117908 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c6359046-e4df-4c51-b082-85650b021d4e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "c6359046-e4df-4c51-b082-85650b021d4e" (UID: "c6359046-e4df-4c51-b082-85650b021d4e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:14:34 crc kubenswrapper[4824]: I1006 10:14:34.123080 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c6359046-e4df-4c51-b082-85650b021d4e-config" (OuterVolumeSpecName: "config") pod "c6359046-e4df-4c51-b082-85650b021d4e" (UID: "c6359046-e4df-4c51-b082-85650b021d4e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:14:34 crc kubenswrapper[4824]: I1006 10:14:34.195183 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c6359046-e4df-4c51-b082-85650b021d4e-config\") on node \"crc\" DevicePath \"\"" Oct 06 10:14:34 crc kubenswrapper[4824]: I1006 10:14:34.195234 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c6359046-e4df-4c51-b082-85650b021d4e-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 06 10:14:34 crc kubenswrapper[4824]: I1006 10:14:34.195256 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-54jlr\" (UniqueName: \"kubernetes.io/projected/c6359046-e4df-4c51-b082-85650b021d4e-kube-api-access-54jlr\") on node \"crc\" DevicePath \"\"" Oct 06 10:14:34 crc kubenswrapper[4824]: I1006 10:14:34.195269 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c6359046-e4df-4c51-b082-85650b021d4e-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 06 10:14:34 crc kubenswrapper[4824]: I1006 10:14:34.195278 4824 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c6359046-e4df-4c51-b082-85650b021d4e-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 06 10:14:34 crc kubenswrapper[4824]: I1006 10:14:34.399412 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/6cdeba93-f1c9-487d-b776-5c2f1bb353d9-etc-swift\") pod \"swift-storage-0\" (UID: \"6cdeba93-f1c9-487d-b776-5c2f1bb353d9\") " pod="openstack/swift-storage-0" Oct 06 10:14:34 crc kubenswrapper[4824]: E1006 10:14:34.399689 4824 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 06 10:14:34 crc kubenswrapper[4824]: E1006 10:14:34.399712 4824 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 06 10:14:34 crc kubenswrapper[4824]: E1006 10:14:34.399766 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/6cdeba93-f1c9-487d-b776-5c2f1bb353d9-etc-swift podName:6cdeba93-f1c9-487d-b776-5c2f1bb353d9 nodeName:}" failed. No retries permitted until 2025-10-06 10:14:38.399747733 +0000 UTC m=+987.764170604 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/6cdeba93-f1c9-487d-b776-5c2f1bb353d9-etc-swift") pod "swift-storage-0" (UID: "6cdeba93-f1c9-487d-b776-5c2f1bb353d9") : configmap "swift-ring-files" not found Oct 06 10:14:34 crc kubenswrapper[4824]: I1006 10:14:34.641025 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-z6wvs"] Oct 06 10:14:34 crc kubenswrapper[4824]: I1006 10:14:34.641252 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-cbp2v" Oct 06 10:14:34 crc kubenswrapper[4824]: E1006 10:14:34.641375 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6359046-e4df-4c51-b082-85650b021d4e" containerName="init" Oct 06 10:14:34 crc kubenswrapper[4824]: I1006 10:14:34.641390 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6359046-e4df-4c51-b082-85650b021d4e" containerName="init" Oct 06 10:14:34 crc kubenswrapper[4824]: I1006 10:14:34.641528 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="c6359046-e4df-4c51-b082-85650b021d4e" containerName="init" Oct 06 10:14:34 crc kubenswrapper[4824]: I1006 10:14:34.642080 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-cbp2v" event={"ID":"c6359046-e4df-4c51-b082-85650b021d4e","Type":"ContainerDied","Data":"413132bee0e4b3ea37e95a833fa533ffce6da5fb0a1bd2d5004db70a87cc4baf"} Oct 06 10:14:34 crc kubenswrapper[4824]: I1006 10:14:34.642116 4824 scope.go:117] "RemoveContainer" containerID="234399ce46398548b58603bd742dbf7daf9d4428f1396d60c224005eac94eaa5" Oct 06 10:14:34 crc kubenswrapper[4824]: I1006 10:14:34.642267 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-z6wvs" Oct 06 10:14:34 crc kubenswrapper[4824]: I1006 10:14:34.643668 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"43002b2d-db5d-488e-9791-56388700cf9f","Type":"ContainerStarted","Data":"33ff7652699679c749e65f759577b245cc1cfc5d1e8d959bb30ae49625311445"} Oct 06 10:14:34 crc kubenswrapper[4824]: I1006 10:14:34.643713 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"43002b2d-db5d-488e-9791-56388700cf9f","Type":"ContainerStarted","Data":"26172f15a87495d16404f968a4036e900a1d13ba7d2765938d6304274be94a59"} Oct 06 10:14:34 crc kubenswrapper[4824]: I1006 10:14:34.643882 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Oct 06 10:14:34 crc kubenswrapper[4824]: I1006 10:14:34.644122 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Oct 06 10:14:34 crc kubenswrapper[4824]: I1006 10:14:34.645089 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Oct 06 10:14:34 crc kubenswrapper[4824]: I1006 10:14:34.646965 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-cwvfc" event={"ID":"b4ac2488-c99e-4380-8e82-e89207990c45","Type":"ContainerStarted","Data":"50ee8359d95369ba4299b5af2a6db387bcddd6aee11c5ce974e82a168eb8a3a0"} Oct 06 10:14:34 crc kubenswrapper[4824]: I1006 10:14:34.647159 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-698758b865-cwvfc" Oct 06 10:14:34 crc kubenswrapper[4824]: I1006 10:14:34.649338 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Oct 06 10:14:34 crc kubenswrapper[4824]: I1006 10:14:34.673246 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-z6wvs"] Oct 06 10:14:34 crc kubenswrapper[4824]: I1006 10:14:34.683099 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=2.339890647 podStartE2EDuration="8.68308036s" podCreationTimestamp="2025-10-06 10:14:26 +0000 UTC" firstStartedPulling="2025-10-06 10:14:27.095042783 +0000 UTC m=+976.459465644" lastFinishedPulling="2025-10-06 10:14:33.438232476 +0000 UTC m=+982.802655357" observedRunningTime="2025-10-06 10:14:34.68068514 +0000 UTC m=+984.045108011" watchObservedRunningTime="2025-10-06 10:14:34.68308036 +0000 UTC m=+984.047503221" Oct 06 10:14:34 crc kubenswrapper[4824]: I1006 10:14:34.736149 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-698758b865-cwvfc" podStartSLOduration=5.736116181 podStartE2EDuration="5.736116181s" podCreationTimestamp="2025-10-06 10:14:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:14:34.724528895 +0000 UTC m=+984.088951756" watchObservedRunningTime="2025-10-06 10:14:34.736116181 +0000 UTC m=+984.100539042" Oct 06 10:14:34 crc kubenswrapper[4824]: I1006 10:14:34.760587 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-cbp2v"] Oct 06 10:14:34 crc kubenswrapper[4824]: I1006 10:14:34.765903 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-cbp2v"] Oct 06 10:14:34 crc kubenswrapper[4824]: I1006 10:14:34.806771 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f85d924-3fee-4cec-a9df-5a0e41567719-combined-ca-bundle\") pod \"swift-ring-rebalance-z6wvs\" (UID: \"7f85d924-3fee-4cec-a9df-5a0e41567719\") " pod="openstack/swift-ring-rebalance-z6wvs" Oct 06 10:14:34 crc kubenswrapper[4824]: I1006 10:14:34.806938 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/7f85d924-3fee-4cec-a9df-5a0e41567719-dispersionconf\") pod \"swift-ring-rebalance-z6wvs\" (UID: \"7f85d924-3fee-4cec-a9df-5a0e41567719\") " pod="openstack/swift-ring-rebalance-z6wvs" Oct 06 10:14:34 crc kubenswrapper[4824]: I1006 10:14:34.806966 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bxk9p\" (UniqueName: \"kubernetes.io/projected/7f85d924-3fee-4cec-a9df-5a0e41567719-kube-api-access-bxk9p\") pod \"swift-ring-rebalance-z6wvs\" (UID: \"7f85d924-3fee-4cec-a9df-5a0e41567719\") " pod="openstack/swift-ring-rebalance-z6wvs" Oct 06 10:14:34 crc kubenswrapper[4824]: I1006 10:14:34.807032 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/7f85d924-3fee-4cec-a9df-5a0e41567719-swiftconf\") pod \"swift-ring-rebalance-z6wvs\" (UID: \"7f85d924-3fee-4cec-a9df-5a0e41567719\") " pod="openstack/swift-ring-rebalance-z6wvs" Oct 06 10:14:34 crc kubenswrapper[4824]: I1006 10:14:34.807073 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7f85d924-3fee-4cec-a9df-5a0e41567719-scripts\") pod \"swift-ring-rebalance-z6wvs\" (UID: \"7f85d924-3fee-4cec-a9df-5a0e41567719\") " pod="openstack/swift-ring-rebalance-z6wvs" Oct 06 10:14:34 crc kubenswrapper[4824]: I1006 10:14:34.807225 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/7f85d924-3fee-4cec-a9df-5a0e41567719-etc-swift\") pod \"swift-ring-rebalance-z6wvs\" (UID: \"7f85d924-3fee-4cec-a9df-5a0e41567719\") " pod="openstack/swift-ring-rebalance-z6wvs" Oct 06 10:14:34 crc kubenswrapper[4824]: I1006 10:14:34.807413 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/7f85d924-3fee-4cec-a9df-5a0e41567719-ring-data-devices\") pod \"swift-ring-rebalance-z6wvs\" (UID: \"7f85d924-3fee-4cec-a9df-5a0e41567719\") " pod="openstack/swift-ring-rebalance-z6wvs" Oct 06 10:14:34 crc kubenswrapper[4824]: I1006 10:14:34.909091 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/7f85d924-3fee-4cec-a9df-5a0e41567719-swiftconf\") pod \"swift-ring-rebalance-z6wvs\" (UID: \"7f85d924-3fee-4cec-a9df-5a0e41567719\") " pod="openstack/swift-ring-rebalance-z6wvs" Oct 06 10:14:34 crc kubenswrapper[4824]: I1006 10:14:34.909182 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7f85d924-3fee-4cec-a9df-5a0e41567719-scripts\") pod \"swift-ring-rebalance-z6wvs\" (UID: \"7f85d924-3fee-4cec-a9df-5a0e41567719\") " pod="openstack/swift-ring-rebalance-z6wvs" Oct 06 10:14:34 crc kubenswrapper[4824]: I1006 10:14:34.909233 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/7f85d924-3fee-4cec-a9df-5a0e41567719-etc-swift\") pod \"swift-ring-rebalance-z6wvs\" (UID: \"7f85d924-3fee-4cec-a9df-5a0e41567719\") " pod="openstack/swift-ring-rebalance-z6wvs" Oct 06 10:14:34 crc kubenswrapper[4824]: I1006 10:14:34.909310 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/7f85d924-3fee-4cec-a9df-5a0e41567719-ring-data-devices\") pod \"swift-ring-rebalance-z6wvs\" (UID: \"7f85d924-3fee-4cec-a9df-5a0e41567719\") " pod="openstack/swift-ring-rebalance-z6wvs" Oct 06 10:14:34 crc kubenswrapper[4824]: I1006 10:14:34.909482 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f85d924-3fee-4cec-a9df-5a0e41567719-combined-ca-bundle\") pod \"swift-ring-rebalance-z6wvs\" (UID: \"7f85d924-3fee-4cec-a9df-5a0e41567719\") " pod="openstack/swift-ring-rebalance-z6wvs" Oct 06 10:14:34 crc kubenswrapper[4824]: I1006 10:14:34.909641 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/7f85d924-3fee-4cec-a9df-5a0e41567719-etc-swift\") pod \"swift-ring-rebalance-z6wvs\" (UID: \"7f85d924-3fee-4cec-a9df-5a0e41567719\") " pod="openstack/swift-ring-rebalance-z6wvs" Oct 06 10:14:34 crc kubenswrapper[4824]: I1006 10:14:34.909684 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/7f85d924-3fee-4cec-a9df-5a0e41567719-dispersionconf\") pod \"swift-ring-rebalance-z6wvs\" (UID: \"7f85d924-3fee-4cec-a9df-5a0e41567719\") " pod="openstack/swift-ring-rebalance-z6wvs" Oct 06 10:14:34 crc kubenswrapper[4824]: I1006 10:14:34.909730 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bxk9p\" (UniqueName: \"kubernetes.io/projected/7f85d924-3fee-4cec-a9df-5a0e41567719-kube-api-access-bxk9p\") pod \"swift-ring-rebalance-z6wvs\" (UID: \"7f85d924-3fee-4cec-a9df-5a0e41567719\") " pod="openstack/swift-ring-rebalance-z6wvs" Oct 06 10:14:34 crc kubenswrapper[4824]: I1006 10:14:34.910328 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7f85d924-3fee-4cec-a9df-5a0e41567719-scripts\") pod \"swift-ring-rebalance-z6wvs\" (UID: \"7f85d924-3fee-4cec-a9df-5a0e41567719\") " pod="openstack/swift-ring-rebalance-z6wvs" Oct 06 10:14:34 crc kubenswrapper[4824]: I1006 10:14:34.910560 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/7f85d924-3fee-4cec-a9df-5a0e41567719-ring-data-devices\") pod \"swift-ring-rebalance-z6wvs\" (UID: \"7f85d924-3fee-4cec-a9df-5a0e41567719\") " pod="openstack/swift-ring-rebalance-z6wvs" Oct 06 10:14:34 crc kubenswrapper[4824]: I1006 10:14:34.915580 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/7f85d924-3fee-4cec-a9df-5a0e41567719-dispersionconf\") pod \"swift-ring-rebalance-z6wvs\" (UID: \"7f85d924-3fee-4cec-a9df-5a0e41567719\") " pod="openstack/swift-ring-rebalance-z6wvs" Oct 06 10:14:34 crc kubenswrapper[4824]: I1006 10:14:34.916785 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/7f85d924-3fee-4cec-a9df-5a0e41567719-swiftconf\") pod \"swift-ring-rebalance-z6wvs\" (UID: \"7f85d924-3fee-4cec-a9df-5a0e41567719\") " pod="openstack/swift-ring-rebalance-z6wvs" Oct 06 10:14:34 crc kubenswrapper[4824]: I1006 10:14:34.927313 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bxk9p\" (UniqueName: \"kubernetes.io/projected/7f85d924-3fee-4cec-a9df-5a0e41567719-kube-api-access-bxk9p\") pod \"swift-ring-rebalance-z6wvs\" (UID: \"7f85d924-3fee-4cec-a9df-5a0e41567719\") " pod="openstack/swift-ring-rebalance-z6wvs" Oct 06 10:14:34 crc kubenswrapper[4824]: I1006 10:14:34.929495 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f85d924-3fee-4cec-a9df-5a0e41567719-combined-ca-bundle\") pod \"swift-ring-rebalance-z6wvs\" (UID: \"7f85d924-3fee-4cec-a9df-5a0e41567719\") " pod="openstack/swift-ring-rebalance-z6wvs" Oct 06 10:14:35 crc kubenswrapper[4824]: I1006 10:14:35.003458 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-z6wvs" Oct 06 10:14:35 crc kubenswrapper[4824]: I1006 10:14:35.286832 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c6359046-e4df-4c51-b082-85650b021d4e" path="/var/lib/kubelet/pods/c6359046-e4df-4c51-b082-85650b021d4e/volumes" Oct 06 10:14:35 crc kubenswrapper[4824]: I1006 10:14:35.483812 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-z6wvs"] Oct 06 10:14:35 crc kubenswrapper[4824]: W1006 10:14:35.491686 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7f85d924_3fee_4cec_a9df_5a0e41567719.slice/crio-1c8cf83603bedead0c112f5d7270c206dee2f12caca278e315d20ff3e2eb53c0 WatchSource:0}: Error finding container 1c8cf83603bedead0c112f5d7270c206dee2f12caca278e315d20ff3e2eb53c0: Status 404 returned error can't find the container with id 1c8cf83603bedead0c112f5d7270c206dee2f12caca278e315d20ff3e2eb53c0 Oct 06 10:14:35 crc kubenswrapper[4824]: I1006 10:14:35.628473 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Oct 06 10:14:35 crc kubenswrapper[4824]: I1006 10:14:35.664927 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-z6wvs" event={"ID":"7f85d924-3fee-4cec-a9df-5a0e41567719","Type":"ContainerStarted","Data":"1c8cf83603bedead0c112f5d7270c206dee2f12caca278e315d20ff3e2eb53c0"} Oct 06 10:14:35 crc kubenswrapper[4824]: I1006 10:14:35.691962 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Oct 06 10:14:38 crc kubenswrapper[4824]: I1006 10:14:38.369731 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Oct 06 10:14:38 crc kubenswrapper[4824]: I1006 10:14:38.420677 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-galera-0" podUID="58f6ad3c-7a20-4be6-8ffb-cadff1c6adef" containerName="galera" probeResult="failure" output=< Oct 06 10:14:38 crc kubenswrapper[4824]: wsrep_local_state_comment (Joined) differs from Synced Oct 06 10:14:38 crc kubenswrapper[4824]: > Oct 06 10:14:38 crc kubenswrapper[4824]: I1006 10:14:38.482866 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/6cdeba93-f1c9-487d-b776-5c2f1bb353d9-etc-swift\") pod \"swift-storage-0\" (UID: \"6cdeba93-f1c9-487d-b776-5c2f1bb353d9\") " pod="openstack/swift-storage-0" Oct 06 10:14:38 crc kubenswrapper[4824]: E1006 10:14:38.484028 4824 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 06 10:14:38 crc kubenswrapper[4824]: E1006 10:14:38.484064 4824 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 06 10:14:38 crc kubenswrapper[4824]: E1006 10:14:38.484137 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/6cdeba93-f1c9-487d-b776-5c2f1bb353d9-etc-swift podName:6cdeba93-f1c9-487d-b776-5c2f1bb353d9 nodeName:}" failed. No retries permitted until 2025-10-06 10:14:46.484107846 +0000 UTC m=+995.848530777 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/6cdeba93-f1c9-487d-b776-5c2f1bb353d9-etc-swift") pod "swift-storage-0" (UID: "6cdeba93-f1c9-487d-b776-5c2f1bb353d9") : configmap "swift-ring-files" not found Oct 06 10:14:39 crc kubenswrapper[4824]: I1006 10:14:39.853254 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-698758b865-cwvfc" Oct 06 10:14:39 crc kubenswrapper[4824]: I1006 10:14:39.939058 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-cnwhb"] Oct 06 10:14:39 crc kubenswrapper[4824]: I1006 10:14:39.939398 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-57d769cc4f-cnwhb" podUID="97cd45b1-470f-483b-b1ed-6dfd5fed1ca8" containerName="dnsmasq-dns" containerID="cri-o://c55a2252f37111bdab065db4f6503c909cd8a7fb4769f9893846fd731cd879e7" gracePeriod=10 Oct 06 10:14:40 crc kubenswrapper[4824]: I1006 10:14:40.522666 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-cnwhb" Oct 06 10:14:40 crc kubenswrapper[4824]: I1006 10:14:40.643896 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/97cd45b1-470f-483b-b1ed-6dfd5fed1ca8-dns-svc\") pod \"97cd45b1-470f-483b-b1ed-6dfd5fed1ca8\" (UID: \"97cd45b1-470f-483b-b1ed-6dfd5fed1ca8\") " Oct 06 10:14:40 crc kubenswrapper[4824]: I1006 10:14:40.643972 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pbw72\" (UniqueName: \"kubernetes.io/projected/97cd45b1-470f-483b-b1ed-6dfd5fed1ca8-kube-api-access-pbw72\") pod \"97cd45b1-470f-483b-b1ed-6dfd5fed1ca8\" (UID: \"97cd45b1-470f-483b-b1ed-6dfd5fed1ca8\") " Oct 06 10:14:40 crc kubenswrapper[4824]: I1006 10:14:40.644145 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/97cd45b1-470f-483b-b1ed-6dfd5fed1ca8-config\") pod \"97cd45b1-470f-483b-b1ed-6dfd5fed1ca8\" (UID: \"97cd45b1-470f-483b-b1ed-6dfd5fed1ca8\") " Oct 06 10:14:40 crc kubenswrapper[4824]: I1006 10:14:40.665232 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/97cd45b1-470f-483b-b1ed-6dfd5fed1ca8-kube-api-access-pbw72" (OuterVolumeSpecName: "kube-api-access-pbw72") pod "97cd45b1-470f-483b-b1ed-6dfd5fed1ca8" (UID: "97cd45b1-470f-483b-b1ed-6dfd5fed1ca8"). InnerVolumeSpecName "kube-api-access-pbw72". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:14:40 crc kubenswrapper[4824]: I1006 10:14:40.710118 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/97cd45b1-470f-483b-b1ed-6dfd5fed1ca8-config" (OuterVolumeSpecName: "config") pod "97cd45b1-470f-483b-b1ed-6dfd5fed1ca8" (UID: "97cd45b1-470f-483b-b1ed-6dfd5fed1ca8"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:14:40 crc kubenswrapper[4824]: I1006 10:14:40.710456 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/97cd45b1-470f-483b-b1ed-6dfd5fed1ca8-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "97cd45b1-470f-483b-b1ed-6dfd5fed1ca8" (UID: "97cd45b1-470f-483b-b1ed-6dfd5fed1ca8"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:14:40 crc kubenswrapper[4824]: I1006 10:14:40.717550 4824 generic.go:334] "Generic (PLEG): container finished" podID="97cd45b1-470f-483b-b1ed-6dfd5fed1ca8" containerID="c55a2252f37111bdab065db4f6503c909cd8a7fb4769f9893846fd731cd879e7" exitCode=0 Oct 06 10:14:40 crc kubenswrapper[4824]: I1006 10:14:40.717638 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-cnwhb" event={"ID":"97cd45b1-470f-483b-b1ed-6dfd5fed1ca8","Type":"ContainerDied","Data":"c55a2252f37111bdab065db4f6503c909cd8a7fb4769f9893846fd731cd879e7"} Oct 06 10:14:40 crc kubenswrapper[4824]: I1006 10:14:40.717670 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-cnwhb" event={"ID":"97cd45b1-470f-483b-b1ed-6dfd5fed1ca8","Type":"ContainerDied","Data":"2c2b14f57a26c7784a998f41c494e1840f6fe44fb9ad5607e6ebc16cfcdf70fb"} Oct 06 10:14:40 crc kubenswrapper[4824]: I1006 10:14:40.717696 4824 scope.go:117] "RemoveContainer" containerID="c55a2252f37111bdab065db4f6503c909cd8a7fb4769f9893846fd731cd879e7" Oct 06 10:14:40 crc kubenswrapper[4824]: I1006 10:14:40.717893 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-cnwhb" Oct 06 10:14:40 crc kubenswrapper[4824]: I1006 10:14:40.722205 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-z6wvs" event={"ID":"7f85d924-3fee-4cec-a9df-5a0e41567719","Type":"ContainerStarted","Data":"8aa27ff0a4048f550ddefbd81c880e705996166ea12fb2ea76ee4b0330a8bb31"} Oct 06 10:14:40 crc kubenswrapper[4824]: I1006 10:14:40.742997 4824 scope.go:117] "RemoveContainer" containerID="949861b1f09b2a44df67727602bedc146340d3bf143c1435c6da5c97af7e69bf" Oct 06 10:14:40 crc kubenswrapper[4824]: I1006 10:14:40.745892 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-z6wvs" podStartSLOduration=2.2020483730000002 podStartE2EDuration="6.745878337s" podCreationTimestamp="2025-10-06 10:14:34 +0000 UTC" firstStartedPulling="2025-10-06 10:14:35.494421444 +0000 UTC m=+984.858844305" lastFinishedPulling="2025-10-06 10:14:40.038251408 +0000 UTC m=+989.402674269" observedRunningTime="2025-10-06 10:14:40.741611711 +0000 UTC m=+990.106034572" watchObservedRunningTime="2025-10-06 10:14:40.745878337 +0000 UTC m=+990.110301198" Oct 06 10:14:40 crc kubenswrapper[4824]: I1006 10:14:40.746206 4824 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/97cd45b1-470f-483b-b1ed-6dfd5fed1ca8-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 06 10:14:40 crc kubenswrapper[4824]: I1006 10:14:40.746235 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pbw72\" (UniqueName: \"kubernetes.io/projected/97cd45b1-470f-483b-b1ed-6dfd5fed1ca8-kube-api-access-pbw72\") on node \"crc\" DevicePath \"\"" Oct 06 10:14:40 crc kubenswrapper[4824]: I1006 10:14:40.746250 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/97cd45b1-470f-483b-b1ed-6dfd5fed1ca8-config\") on node \"crc\" DevicePath \"\"" Oct 06 10:14:40 crc kubenswrapper[4824]: I1006 10:14:40.763615 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-cnwhb"] Oct 06 10:14:40 crc kubenswrapper[4824]: I1006 10:14:40.771352 4824 scope.go:117] "RemoveContainer" containerID="c55a2252f37111bdab065db4f6503c909cd8a7fb4769f9893846fd731cd879e7" Oct 06 10:14:40 crc kubenswrapper[4824]: E1006 10:14:40.772306 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c55a2252f37111bdab065db4f6503c909cd8a7fb4769f9893846fd731cd879e7\": container with ID starting with c55a2252f37111bdab065db4f6503c909cd8a7fb4769f9893846fd731cd879e7 not found: ID does not exist" containerID="c55a2252f37111bdab065db4f6503c909cd8a7fb4769f9893846fd731cd879e7" Oct 06 10:14:40 crc kubenswrapper[4824]: I1006 10:14:40.772363 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c55a2252f37111bdab065db4f6503c909cd8a7fb4769f9893846fd731cd879e7"} err="failed to get container status \"c55a2252f37111bdab065db4f6503c909cd8a7fb4769f9893846fd731cd879e7\": rpc error: code = NotFound desc = could not find container \"c55a2252f37111bdab065db4f6503c909cd8a7fb4769f9893846fd731cd879e7\": container with ID starting with c55a2252f37111bdab065db4f6503c909cd8a7fb4769f9893846fd731cd879e7 not found: ID does not exist" Oct 06 10:14:40 crc kubenswrapper[4824]: I1006 10:14:40.772396 4824 scope.go:117] "RemoveContainer" containerID="949861b1f09b2a44df67727602bedc146340d3bf143c1435c6da5c97af7e69bf" Oct 06 10:14:40 crc kubenswrapper[4824]: E1006 10:14:40.772903 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"949861b1f09b2a44df67727602bedc146340d3bf143c1435c6da5c97af7e69bf\": container with ID starting with 949861b1f09b2a44df67727602bedc146340d3bf143c1435c6da5c97af7e69bf not found: ID does not exist" containerID="949861b1f09b2a44df67727602bedc146340d3bf143c1435c6da5c97af7e69bf" Oct 06 10:14:40 crc kubenswrapper[4824]: I1006 10:14:40.772926 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"949861b1f09b2a44df67727602bedc146340d3bf143c1435c6da5c97af7e69bf"} err="failed to get container status \"949861b1f09b2a44df67727602bedc146340d3bf143c1435c6da5c97af7e69bf\": rpc error: code = NotFound desc = could not find container \"949861b1f09b2a44df67727602bedc146340d3bf143c1435c6da5c97af7e69bf\": container with ID starting with 949861b1f09b2a44df67727602bedc146340d3bf143c1435c6da5c97af7e69bf not found: ID does not exist" Oct 06 10:14:40 crc kubenswrapper[4824]: I1006 10:14:40.774612 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-cnwhb"] Oct 06 10:14:41 crc kubenswrapper[4824]: I1006 10:14:41.293142 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="97cd45b1-470f-483b-b1ed-6dfd5fed1ca8" path="/var/lib/kubelet/pods/97cd45b1-470f-483b-b1ed-6dfd5fed1ca8/volumes" Oct 06 10:14:42 crc kubenswrapper[4824]: I1006 10:14:42.785530 4824 generic.go:334] "Generic (PLEG): container finished" podID="c8dbad1e-02be-43c8-8f80-ab771ee81742" containerID="9e782d23005bd2ab15abfb2afc9b231130f78b0f01da4901f27b58d3b1eb931d" exitCode=0 Oct 06 10:14:42 crc kubenswrapper[4824]: I1006 10:14:42.785697 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"c8dbad1e-02be-43c8-8f80-ab771ee81742","Type":"ContainerDied","Data":"9e782d23005bd2ab15abfb2afc9b231130f78b0f01da4901f27b58d3b1eb931d"} Oct 06 10:14:43 crc kubenswrapper[4824]: I1006 10:14:43.797922 4824 generic.go:334] "Generic (PLEG): container finished" podID="f7d047da-89e3-4cd4-bf02-e6a3e15c585f" containerID="2512559a4666f95711ee0e9d31717a5aa74ce89aab85c9eec580e792833cd962" exitCode=0 Oct 06 10:14:43 crc kubenswrapper[4824]: I1006 10:14:43.798103 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"f7d047da-89e3-4cd4-bf02-e6a3e15c585f","Type":"ContainerDied","Data":"2512559a4666f95711ee0e9d31717a5aa74ce89aab85c9eec580e792833cd962"} Oct 06 10:14:43 crc kubenswrapper[4824]: I1006 10:14:43.803418 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"c8dbad1e-02be-43c8-8f80-ab771ee81742","Type":"ContainerStarted","Data":"776dc359b5340bb21b168dc69fe78fcefb80aade02e11f0b963bbb9e8b58fb92"} Oct 06 10:14:43 crc kubenswrapper[4824]: I1006 10:14:43.804432 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Oct 06 10:14:43 crc kubenswrapper[4824]: I1006 10:14:43.881150 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=40.298624583 podStartE2EDuration="50.881127309s" podCreationTimestamp="2025-10-06 10:13:53 +0000 UTC" firstStartedPulling="2025-10-06 10:13:59.01030265 +0000 UTC m=+948.374725511" lastFinishedPulling="2025-10-06 10:14:09.592805356 +0000 UTC m=+958.957228237" observedRunningTime="2025-10-06 10:14:43.873786497 +0000 UTC m=+993.238209358" watchObservedRunningTime="2025-10-06 10:14:43.881127309 +0000 UTC m=+993.245550180" Oct 06 10:14:43 crc kubenswrapper[4824]: I1006 10:14:43.915741 4824 patch_prober.go:28] interesting pod/machine-config-daemon-khgzw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 10:14:43 crc kubenswrapper[4824]: I1006 10:14:43.915805 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 10:14:44 crc kubenswrapper[4824]: I1006 10:14:44.814601 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"f7d047da-89e3-4cd4-bf02-e6a3e15c585f","Type":"ContainerStarted","Data":"f8d792090dd667a7c36525cb2a4beb43244cc576a1e4eafe83b7803d2d5cafb8"} Oct 06 10:14:44 crc kubenswrapper[4824]: I1006 10:14:44.815333 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Oct 06 10:14:44 crc kubenswrapper[4824]: I1006 10:14:44.844349 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=42.33892124 podStartE2EDuration="52.844323787s" podCreationTimestamp="2025-10-06 10:13:52 +0000 UTC" firstStartedPulling="2025-10-06 10:13:59.01031599 +0000 UTC m=+948.374738861" lastFinishedPulling="2025-10-06 10:14:09.515718547 +0000 UTC m=+958.880141408" observedRunningTime="2025-10-06 10:14:44.839853277 +0000 UTC m=+994.204276138" watchObservedRunningTime="2025-10-06 10:14:44.844323787 +0000 UTC m=+994.208746648" Oct 06 10:14:46 crc kubenswrapper[4824]: I1006 10:14:46.486516 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/6cdeba93-f1c9-487d-b776-5c2f1bb353d9-etc-swift\") pod \"swift-storage-0\" (UID: \"6cdeba93-f1c9-487d-b776-5c2f1bb353d9\") " pod="openstack/swift-storage-0" Oct 06 10:14:46 crc kubenswrapper[4824]: E1006 10:14:46.486862 4824 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 06 10:14:46 crc kubenswrapper[4824]: E1006 10:14:46.487123 4824 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 06 10:14:46 crc kubenswrapper[4824]: E1006 10:14:46.487239 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/6cdeba93-f1c9-487d-b776-5c2f1bb353d9-etc-swift podName:6cdeba93-f1c9-487d-b776-5c2f1bb353d9 nodeName:}" failed. No retries permitted until 2025-10-06 10:15:02.487205535 +0000 UTC m=+1011.851628426 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/6cdeba93-f1c9-487d-b776-5c2f1bb353d9-etc-swift") pod "swift-storage-0" (UID: "6cdeba93-f1c9-487d-b776-5c2f1bb353d9") : configmap "swift-ring-files" not found Oct 06 10:14:46 crc kubenswrapper[4824]: I1006 10:14:46.685190 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Oct 06 10:14:46 crc kubenswrapper[4824]: I1006 10:14:46.834961 4824 generic.go:334] "Generic (PLEG): container finished" podID="7f85d924-3fee-4cec-a9df-5a0e41567719" containerID="8aa27ff0a4048f550ddefbd81c880e705996166ea12fb2ea76ee4b0330a8bb31" exitCode=0 Oct 06 10:14:46 crc kubenswrapper[4824]: I1006 10:14:46.835017 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-z6wvs" event={"ID":"7f85d924-3fee-4cec-a9df-5a0e41567719","Type":"ContainerDied","Data":"8aa27ff0a4048f550ddefbd81c880e705996166ea12fb2ea76ee4b0330a8bb31"} Oct 06 10:14:46 crc kubenswrapper[4824]: E1006 10:14:46.976950 4824 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7f85d924_3fee_4cec_a9df_5a0e41567719.slice/crio-conmon-8aa27ff0a4048f550ddefbd81c880e705996166ea12fb2ea76ee4b0330a8bb31.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7f85d924_3fee_4cec_a9df_5a0e41567719.slice/crio-8aa27ff0a4048f550ddefbd81c880e705996166ea12fb2ea76ee4b0330a8bb31.scope\": RecentStats: unable to find data in memory cache]" Oct 06 10:14:47 crc kubenswrapper[4824]: I1006 10:14:47.143286 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Oct 06 10:14:47 crc kubenswrapper[4824]: I1006 10:14:47.648790 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-f4pft"] Oct 06 10:14:47 crc kubenswrapper[4824]: E1006 10:14:47.649656 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97cd45b1-470f-483b-b1ed-6dfd5fed1ca8" containerName="init" Oct 06 10:14:47 crc kubenswrapper[4824]: I1006 10:14:47.649674 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="97cd45b1-470f-483b-b1ed-6dfd5fed1ca8" containerName="init" Oct 06 10:14:47 crc kubenswrapper[4824]: E1006 10:14:47.649705 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97cd45b1-470f-483b-b1ed-6dfd5fed1ca8" containerName="dnsmasq-dns" Oct 06 10:14:47 crc kubenswrapper[4824]: I1006 10:14:47.649715 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="97cd45b1-470f-483b-b1ed-6dfd5fed1ca8" containerName="dnsmasq-dns" Oct 06 10:14:47 crc kubenswrapper[4824]: I1006 10:14:47.649893 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="97cd45b1-470f-483b-b1ed-6dfd5fed1ca8" containerName="dnsmasq-dns" Oct 06 10:14:47 crc kubenswrapper[4824]: I1006 10:14:47.650489 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-f4pft" Oct 06 10:14:47 crc kubenswrapper[4824]: I1006 10:14:47.676957 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-f4pft"] Oct 06 10:14:47 crc kubenswrapper[4824]: I1006 10:14:47.712276 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vrfsk\" (UniqueName: \"kubernetes.io/projected/a55f5293-0c03-479d-a0c7-bb4869445794-kube-api-access-vrfsk\") pod \"placement-db-create-f4pft\" (UID: \"a55f5293-0c03-479d-a0c7-bb4869445794\") " pod="openstack/placement-db-create-f4pft" Oct 06 10:14:47 crc kubenswrapper[4824]: I1006 10:14:47.814591 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vrfsk\" (UniqueName: \"kubernetes.io/projected/a55f5293-0c03-479d-a0c7-bb4869445794-kube-api-access-vrfsk\") pod \"placement-db-create-f4pft\" (UID: \"a55f5293-0c03-479d-a0c7-bb4869445794\") " pod="openstack/placement-db-create-f4pft" Oct 06 10:14:47 crc kubenswrapper[4824]: I1006 10:14:47.843781 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vrfsk\" (UniqueName: \"kubernetes.io/projected/a55f5293-0c03-479d-a0c7-bb4869445794-kube-api-access-vrfsk\") pod \"placement-db-create-f4pft\" (UID: \"a55f5293-0c03-479d-a0c7-bb4869445794\") " pod="openstack/placement-db-create-f4pft" Oct 06 10:14:47 crc kubenswrapper[4824]: I1006 10:14:47.971184 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-f4pft" Oct 06 10:14:48 crc kubenswrapper[4824]: I1006 10:14:48.237747 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-z6wvs" Oct 06 10:14:48 crc kubenswrapper[4824]: I1006 10:14:48.321436 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/7f85d924-3fee-4cec-a9df-5a0e41567719-swiftconf\") pod \"7f85d924-3fee-4cec-a9df-5a0e41567719\" (UID: \"7f85d924-3fee-4cec-a9df-5a0e41567719\") " Oct 06 10:14:48 crc kubenswrapper[4824]: I1006 10:14:48.321494 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/7f85d924-3fee-4cec-a9df-5a0e41567719-etc-swift\") pod \"7f85d924-3fee-4cec-a9df-5a0e41567719\" (UID: \"7f85d924-3fee-4cec-a9df-5a0e41567719\") " Oct 06 10:14:48 crc kubenswrapper[4824]: I1006 10:14:48.321546 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bxk9p\" (UniqueName: \"kubernetes.io/projected/7f85d924-3fee-4cec-a9df-5a0e41567719-kube-api-access-bxk9p\") pod \"7f85d924-3fee-4cec-a9df-5a0e41567719\" (UID: \"7f85d924-3fee-4cec-a9df-5a0e41567719\") " Oct 06 10:14:48 crc kubenswrapper[4824]: I1006 10:14:48.322501 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/7f85d924-3fee-4cec-a9df-5a0e41567719-dispersionconf\") pod \"7f85d924-3fee-4cec-a9df-5a0e41567719\" (UID: \"7f85d924-3fee-4cec-a9df-5a0e41567719\") " Oct 06 10:14:48 crc kubenswrapper[4824]: I1006 10:14:48.322566 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7f85d924-3fee-4cec-a9df-5a0e41567719-scripts\") pod \"7f85d924-3fee-4cec-a9df-5a0e41567719\" (UID: \"7f85d924-3fee-4cec-a9df-5a0e41567719\") " Oct 06 10:14:48 crc kubenswrapper[4824]: I1006 10:14:48.322595 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/7f85d924-3fee-4cec-a9df-5a0e41567719-ring-data-devices\") pod \"7f85d924-3fee-4cec-a9df-5a0e41567719\" (UID: \"7f85d924-3fee-4cec-a9df-5a0e41567719\") " Oct 06 10:14:48 crc kubenswrapper[4824]: I1006 10:14:48.322621 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f85d924-3fee-4cec-a9df-5a0e41567719-combined-ca-bundle\") pod \"7f85d924-3fee-4cec-a9df-5a0e41567719\" (UID: \"7f85d924-3fee-4cec-a9df-5a0e41567719\") " Oct 06 10:14:48 crc kubenswrapper[4824]: I1006 10:14:48.322768 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7f85d924-3fee-4cec-a9df-5a0e41567719-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "7f85d924-3fee-4cec-a9df-5a0e41567719" (UID: "7f85d924-3fee-4cec-a9df-5a0e41567719"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:14:48 crc kubenswrapper[4824]: I1006 10:14:48.323187 4824 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/7f85d924-3fee-4cec-a9df-5a0e41567719-etc-swift\") on node \"crc\" DevicePath \"\"" Oct 06 10:14:48 crc kubenswrapper[4824]: I1006 10:14:48.323840 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7f85d924-3fee-4cec-a9df-5a0e41567719-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "7f85d924-3fee-4cec-a9df-5a0e41567719" (UID: "7f85d924-3fee-4cec-a9df-5a0e41567719"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:14:48 crc kubenswrapper[4824]: I1006 10:14:48.329293 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7f85d924-3fee-4cec-a9df-5a0e41567719-kube-api-access-bxk9p" (OuterVolumeSpecName: "kube-api-access-bxk9p") pod "7f85d924-3fee-4cec-a9df-5a0e41567719" (UID: "7f85d924-3fee-4cec-a9df-5a0e41567719"). InnerVolumeSpecName "kube-api-access-bxk9p". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:14:48 crc kubenswrapper[4824]: I1006 10:14:48.331752 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f85d924-3fee-4cec-a9df-5a0e41567719-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "7f85d924-3fee-4cec-a9df-5a0e41567719" (UID: "7f85d924-3fee-4cec-a9df-5a0e41567719"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:14:48 crc kubenswrapper[4824]: I1006 10:14:48.344733 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7f85d924-3fee-4cec-a9df-5a0e41567719-scripts" (OuterVolumeSpecName: "scripts") pod "7f85d924-3fee-4cec-a9df-5a0e41567719" (UID: "7f85d924-3fee-4cec-a9df-5a0e41567719"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:14:48 crc kubenswrapper[4824]: I1006 10:14:48.351432 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f85d924-3fee-4cec-a9df-5a0e41567719-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "7f85d924-3fee-4cec-a9df-5a0e41567719" (UID: "7f85d924-3fee-4cec-a9df-5a0e41567719"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:14:48 crc kubenswrapper[4824]: I1006 10:14:48.352311 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f85d924-3fee-4cec-a9df-5a0e41567719-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7f85d924-3fee-4cec-a9df-5a0e41567719" (UID: "7f85d924-3fee-4cec-a9df-5a0e41567719"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:14:48 crc kubenswrapper[4824]: I1006 10:14:48.424936 4824 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/7f85d924-3fee-4cec-a9df-5a0e41567719-swiftconf\") on node \"crc\" DevicePath \"\"" Oct 06 10:14:48 crc kubenswrapper[4824]: I1006 10:14:48.425221 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bxk9p\" (UniqueName: \"kubernetes.io/projected/7f85d924-3fee-4cec-a9df-5a0e41567719-kube-api-access-bxk9p\") on node \"crc\" DevicePath \"\"" Oct 06 10:14:48 crc kubenswrapper[4824]: I1006 10:14:48.425283 4824 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/7f85d924-3fee-4cec-a9df-5a0e41567719-dispersionconf\") on node \"crc\" DevicePath \"\"" Oct 06 10:14:48 crc kubenswrapper[4824]: I1006 10:14:48.425379 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7f85d924-3fee-4cec-a9df-5a0e41567719-scripts\") on node \"crc\" DevicePath \"\"" Oct 06 10:14:48 crc kubenswrapper[4824]: I1006 10:14:48.425441 4824 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/7f85d924-3fee-4cec-a9df-5a0e41567719-ring-data-devices\") on node \"crc\" DevicePath \"\"" Oct 06 10:14:48 crc kubenswrapper[4824]: I1006 10:14:48.425499 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f85d924-3fee-4cec-a9df-5a0e41567719-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 10:14:48 crc kubenswrapper[4824]: W1006 10:14:48.471620 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda55f5293_0c03_479d_a0c7_bb4869445794.slice/crio-b034a3d93f02061e07c26f6d9b5d0ad2e43a58c1c4d116fc8d0d5762bf160e27 WatchSource:0}: Error finding container b034a3d93f02061e07c26f6d9b5d0ad2e43a58c1c4d116fc8d0d5762bf160e27: Status 404 returned error can't find the container with id b034a3d93f02061e07c26f6d9b5d0ad2e43a58c1c4d116fc8d0d5762bf160e27 Oct 06 10:14:48 crc kubenswrapper[4824]: I1006 10:14:48.475027 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-f4pft"] Oct 06 10:14:48 crc kubenswrapper[4824]: I1006 10:14:48.855330 4824 generic.go:334] "Generic (PLEG): container finished" podID="a55f5293-0c03-479d-a0c7-bb4869445794" containerID="59c36f0cc6dade51bfb70bc93fe29975a6816585123e2e2b7180f65307c8bd9a" exitCode=0 Oct 06 10:14:48 crc kubenswrapper[4824]: I1006 10:14:48.855808 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-f4pft" event={"ID":"a55f5293-0c03-479d-a0c7-bb4869445794","Type":"ContainerDied","Data":"59c36f0cc6dade51bfb70bc93fe29975a6816585123e2e2b7180f65307c8bd9a"} Oct 06 10:14:48 crc kubenswrapper[4824]: I1006 10:14:48.855841 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-f4pft" event={"ID":"a55f5293-0c03-479d-a0c7-bb4869445794","Type":"ContainerStarted","Data":"b034a3d93f02061e07c26f6d9b5d0ad2e43a58c1c4d116fc8d0d5762bf160e27"} Oct 06 10:14:48 crc kubenswrapper[4824]: I1006 10:14:48.858211 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-z6wvs" event={"ID":"7f85d924-3fee-4cec-a9df-5a0e41567719","Type":"ContainerDied","Data":"1c8cf83603bedead0c112f5d7270c206dee2f12caca278e315d20ff3e2eb53c0"} Oct 06 10:14:48 crc kubenswrapper[4824]: I1006 10:14:48.858253 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1c8cf83603bedead0c112f5d7270c206dee2f12caca278e315d20ff3e2eb53c0" Oct 06 10:14:48 crc kubenswrapper[4824]: I1006 10:14:48.858277 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-z6wvs" Oct 06 10:14:50 crc kubenswrapper[4824]: I1006 10:14:50.240664 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-f4pft" Oct 06 10:14:50 crc kubenswrapper[4824]: I1006 10:14:50.362541 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vrfsk\" (UniqueName: \"kubernetes.io/projected/a55f5293-0c03-479d-a0c7-bb4869445794-kube-api-access-vrfsk\") pod \"a55f5293-0c03-479d-a0c7-bb4869445794\" (UID: \"a55f5293-0c03-479d-a0c7-bb4869445794\") " Oct 06 10:14:50 crc kubenswrapper[4824]: I1006 10:14:50.371197 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a55f5293-0c03-479d-a0c7-bb4869445794-kube-api-access-vrfsk" (OuterVolumeSpecName: "kube-api-access-vrfsk") pod "a55f5293-0c03-479d-a0c7-bb4869445794" (UID: "a55f5293-0c03-479d-a0c7-bb4869445794"). InnerVolumeSpecName "kube-api-access-vrfsk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:14:50 crc kubenswrapper[4824]: I1006 10:14:50.464960 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vrfsk\" (UniqueName: \"kubernetes.io/projected/a55f5293-0c03-479d-a0c7-bb4869445794-kube-api-access-vrfsk\") on node \"crc\" DevicePath \"\"" Oct 06 10:14:50 crc kubenswrapper[4824]: I1006 10:14:50.879674 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-f4pft" event={"ID":"a55f5293-0c03-479d-a0c7-bb4869445794","Type":"ContainerDied","Data":"b034a3d93f02061e07c26f6d9b5d0ad2e43a58c1c4d116fc8d0d5762bf160e27"} Oct 06 10:14:50 crc kubenswrapper[4824]: I1006 10:14:50.879723 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-f4pft" Oct 06 10:14:50 crc kubenswrapper[4824]: I1006 10:14:50.879738 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b034a3d93f02061e07c26f6d9b5d0ad2e43a58c1c4d116fc8d0d5762bf160e27" Oct 06 10:14:52 crc kubenswrapper[4824]: I1006 10:14:52.850111 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-5rvk5"] Oct 06 10:14:52 crc kubenswrapper[4824]: E1006 10:14:52.851043 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f85d924-3fee-4cec-a9df-5a0e41567719" containerName="swift-ring-rebalance" Oct 06 10:14:52 crc kubenswrapper[4824]: I1006 10:14:52.851065 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f85d924-3fee-4cec-a9df-5a0e41567719" containerName="swift-ring-rebalance" Oct 06 10:14:52 crc kubenswrapper[4824]: E1006 10:14:52.851091 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a55f5293-0c03-479d-a0c7-bb4869445794" containerName="mariadb-database-create" Oct 06 10:14:52 crc kubenswrapper[4824]: I1006 10:14:52.851106 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="a55f5293-0c03-479d-a0c7-bb4869445794" containerName="mariadb-database-create" Oct 06 10:14:52 crc kubenswrapper[4824]: I1006 10:14:52.851427 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="7f85d924-3fee-4cec-a9df-5a0e41567719" containerName="swift-ring-rebalance" Oct 06 10:14:52 crc kubenswrapper[4824]: I1006 10:14:52.851452 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="a55f5293-0c03-479d-a0c7-bb4869445794" containerName="mariadb-database-create" Oct 06 10:14:52 crc kubenswrapper[4824]: I1006 10:14:52.852537 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-5rvk5" Oct 06 10:14:52 crc kubenswrapper[4824]: I1006 10:14:52.861142 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-5rvk5"] Oct 06 10:14:52 crc kubenswrapper[4824]: I1006 10:14:52.917140 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-48qsh" podUID="03d76a8a-df87-4a48-8fb9-e6a502b37ae8" containerName="ovn-controller" probeResult="failure" output=< Oct 06 10:14:52 crc kubenswrapper[4824]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Oct 06 10:14:52 crc kubenswrapper[4824]: > Oct 06 10:14:52 crc kubenswrapper[4824]: I1006 10:14:52.921129 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q96xk\" (UniqueName: \"kubernetes.io/projected/20c54a88-ce86-425a-b12c-cebd09083d14-kube-api-access-q96xk\") pod \"glance-db-create-5rvk5\" (UID: \"20c54a88-ce86-425a-b12c-cebd09083d14\") " pod="openstack/glance-db-create-5rvk5" Oct 06 10:14:52 crc kubenswrapper[4824]: I1006 10:14:52.936435 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-v22g5" Oct 06 10:14:52 crc kubenswrapper[4824]: I1006 10:14:52.939623 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-v22g5" Oct 06 10:14:53 crc kubenswrapper[4824]: I1006 10:14:53.022512 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q96xk\" (UniqueName: \"kubernetes.io/projected/20c54a88-ce86-425a-b12c-cebd09083d14-kube-api-access-q96xk\") pod \"glance-db-create-5rvk5\" (UID: \"20c54a88-ce86-425a-b12c-cebd09083d14\") " pod="openstack/glance-db-create-5rvk5" Oct 06 10:14:53 crc kubenswrapper[4824]: I1006 10:14:53.045058 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q96xk\" (UniqueName: \"kubernetes.io/projected/20c54a88-ce86-425a-b12c-cebd09083d14-kube-api-access-q96xk\") pod \"glance-db-create-5rvk5\" (UID: \"20c54a88-ce86-425a-b12c-cebd09083d14\") " pod="openstack/glance-db-create-5rvk5" Oct 06 10:14:53 crc kubenswrapper[4824]: I1006 10:14:53.176300 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-48qsh-config-dntgb"] Oct 06 10:14:53 crc kubenswrapper[4824]: I1006 10:14:53.177654 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-5rvk5" Oct 06 10:14:53 crc kubenswrapper[4824]: I1006 10:14:53.179414 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-48qsh-config-dntgb" Oct 06 10:14:53 crc kubenswrapper[4824]: I1006 10:14:53.183468 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Oct 06 10:14:53 crc kubenswrapper[4824]: I1006 10:14:53.193191 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-48qsh-config-dntgb"] Oct 06 10:14:53 crc kubenswrapper[4824]: I1006 10:14:53.225700 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/41bb1f2f-34c5-43a5-b416-fbb5052a9dd8-var-log-ovn\") pod \"ovn-controller-48qsh-config-dntgb\" (UID: \"41bb1f2f-34c5-43a5-b416-fbb5052a9dd8\") " pod="openstack/ovn-controller-48qsh-config-dntgb" Oct 06 10:14:53 crc kubenswrapper[4824]: I1006 10:14:53.225860 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/41bb1f2f-34c5-43a5-b416-fbb5052a9dd8-var-run-ovn\") pod \"ovn-controller-48qsh-config-dntgb\" (UID: \"41bb1f2f-34c5-43a5-b416-fbb5052a9dd8\") " pod="openstack/ovn-controller-48qsh-config-dntgb" Oct 06 10:14:53 crc kubenswrapper[4824]: I1006 10:14:53.225901 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/41bb1f2f-34c5-43a5-b416-fbb5052a9dd8-additional-scripts\") pod \"ovn-controller-48qsh-config-dntgb\" (UID: \"41bb1f2f-34c5-43a5-b416-fbb5052a9dd8\") " pod="openstack/ovn-controller-48qsh-config-dntgb" Oct 06 10:14:53 crc kubenswrapper[4824]: I1006 10:14:53.225927 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/41bb1f2f-34c5-43a5-b416-fbb5052a9dd8-var-run\") pod \"ovn-controller-48qsh-config-dntgb\" (UID: \"41bb1f2f-34c5-43a5-b416-fbb5052a9dd8\") " pod="openstack/ovn-controller-48qsh-config-dntgb" Oct 06 10:14:53 crc kubenswrapper[4824]: I1006 10:14:53.225991 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hxjrf\" (UniqueName: \"kubernetes.io/projected/41bb1f2f-34c5-43a5-b416-fbb5052a9dd8-kube-api-access-hxjrf\") pod \"ovn-controller-48qsh-config-dntgb\" (UID: \"41bb1f2f-34c5-43a5-b416-fbb5052a9dd8\") " pod="openstack/ovn-controller-48qsh-config-dntgb" Oct 06 10:14:53 crc kubenswrapper[4824]: I1006 10:14:53.226014 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/41bb1f2f-34c5-43a5-b416-fbb5052a9dd8-scripts\") pod \"ovn-controller-48qsh-config-dntgb\" (UID: \"41bb1f2f-34c5-43a5-b416-fbb5052a9dd8\") " pod="openstack/ovn-controller-48qsh-config-dntgb" Oct 06 10:14:53 crc kubenswrapper[4824]: I1006 10:14:53.327942 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/41bb1f2f-34c5-43a5-b416-fbb5052a9dd8-var-log-ovn\") pod \"ovn-controller-48qsh-config-dntgb\" (UID: \"41bb1f2f-34c5-43a5-b416-fbb5052a9dd8\") " pod="openstack/ovn-controller-48qsh-config-dntgb" Oct 06 10:14:53 crc kubenswrapper[4824]: I1006 10:14:53.328567 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/41bb1f2f-34c5-43a5-b416-fbb5052a9dd8-var-log-ovn\") pod \"ovn-controller-48qsh-config-dntgb\" (UID: \"41bb1f2f-34c5-43a5-b416-fbb5052a9dd8\") " pod="openstack/ovn-controller-48qsh-config-dntgb" Oct 06 10:14:53 crc kubenswrapper[4824]: I1006 10:14:53.328695 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/41bb1f2f-34c5-43a5-b416-fbb5052a9dd8-var-run-ovn\") pod \"ovn-controller-48qsh-config-dntgb\" (UID: \"41bb1f2f-34c5-43a5-b416-fbb5052a9dd8\") " pod="openstack/ovn-controller-48qsh-config-dntgb" Oct 06 10:14:53 crc kubenswrapper[4824]: I1006 10:14:53.328754 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/41bb1f2f-34c5-43a5-b416-fbb5052a9dd8-additional-scripts\") pod \"ovn-controller-48qsh-config-dntgb\" (UID: \"41bb1f2f-34c5-43a5-b416-fbb5052a9dd8\") " pod="openstack/ovn-controller-48qsh-config-dntgb" Oct 06 10:14:53 crc kubenswrapper[4824]: I1006 10:14:53.328830 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/41bb1f2f-34c5-43a5-b416-fbb5052a9dd8-var-run-ovn\") pod \"ovn-controller-48qsh-config-dntgb\" (UID: \"41bb1f2f-34c5-43a5-b416-fbb5052a9dd8\") " pod="openstack/ovn-controller-48qsh-config-dntgb" Oct 06 10:14:53 crc kubenswrapper[4824]: I1006 10:14:53.329457 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/41bb1f2f-34c5-43a5-b416-fbb5052a9dd8-additional-scripts\") pod \"ovn-controller-48qsh-config-dntgb\" (UID: \"41bb1f2f-34c5-43a5-b416-fbb5052a9dd8\") " pod="openstack/ovn-controller-48qsh-config-dntgb" Oct 06 10:14:53 crc kubenswrapper[4824]: I1006 10:14:53.328779 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/41bb1f2f-34c5-43a5-b416-fbb5052a9dd8-var-run\") pod \"ovn-controller-48qsh-config-dntgb\" (UID: \"41bb1f2f-34c5-43a5-b416-fbb5052a9dd8\") " pod="openstack/ovn-controller-48qsh-config-dntgb" Oct 06 10:14:53 crc kubenswrapper[4824]: I1006 10:14:53.329574 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hxjrf\" (UniqueName: \"kubernetes.io/projected/41bb1f2f-34c5-43a5-b416-fbb5052a9dd8-kube-api-access-hxjrf\") pod \"ovn-controller-48qsh-config-dntgb\" (UID: \"41bb1f2f-34c5-43a5-b416-fbb5052a9dd8\") " pod="openstack/ovn-controller-48qsh-config-dntgb" Oct 06 10:14:53 crc kubenswrapper[4824]: I1006 10:14:53.329600 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/41bb1f2f-34c5-43a5-b416-fbb5052a9dd8-scripts\") pod \"ovn-controller-48qsh-config-dntgb\" (UID: \"41bb1f2f-34c5-43a5-b416-fbb5052a9dd8\") " pod="openstack/ovn-controller-48qsh-config-dntgb" Oct 06 10:14:53 crc kubenswrapper[4824]: I1006 10:14:53.331228 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/41bb1f2f-34c5-43a5-b416-fbb5052a9dd8-scripts\") pod \"ovn-controller-48qsh-config-dntgb\" (UID: \"41bb1f2f-34c5-43a5-b416-fbb5052a9dd8\") " pod="openstack/ovn-controller-48qsh-config-dntgb" Oct 06 10:14:53 crc kubenswrapper[4824]: I1006 10:14:53.331279 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/41bb1f2f-34c5-43a5-b416-fbb5052a9dd8-var-run\") pod \"ovn-controller-48qsh-config-dntgb\" (UID: \"41bb1f2f-34c5-43a5-b416-fbb5052a9dd8\") " pod="openstack/ovn-controller-48qsh-config-dntgb" Oct 06 10:14:53 crc kubenswrapper[4824]: I1006 10:14:53.381850 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hxjrf\" (UniqueName: \"kubernetes.io/projected/41bb1f2f-34c5-43a5-b416-fbb5052a9dd8-kube-api-access-hxjrf\") pod \"ovn-controller-48qsh-config-dntgb\" (UID: \"41bb1f2f-34c5-43a5-b416-fbb5052a9dd8\") " pod="openstack/ovn-controller-48qsh-config-dntgb" Oct 06 10:14:53 crc kubenswrapper[4824]: I1006 10:14:53.498342 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-48qsh-config-dntgb" Oct 06 10:14:53 crc kubenswrapper[4824]: I1006 10:14:53.755433 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-5rvk5"] Oct 06 10:14:53 crc kubenswrapper[4824]: I1006 10:14:53.913783 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-5rvk5" event={"ID":"20c54a88-ce86-425a-b12c-cebd09083d14","Type":"ContainerStarted","Data":"22d7e53f31dc36b44599a26a7625f105a50224c0dca6c2503d6186e260a27117"} Oct 06 10:14:53 crc kubenswrapper[4824]: I1006 10:14:53.979268 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-48qsh-config-dntgb"] Oct 06 10:14:53 crc kubenswrapper[4824]: W1006 10:14:53.980300 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod41bb1f2f_34c5_43a5_b416_fbb5052a9dd8.slice/crio-b561b0dbf690d37a94cb0ad01bc228f59e0c7ff399805379684198972e067d49 WatchSource:0}: Error finding container b561b0dbf690d37a94cb0ad01bc228f59e0c7ff399805379684198972e067d49: Status 404 returned error can't find the container with id b561b0dbf690d37a94cb0ad01bc228f59e0c7ff399805379684198972e067d49 Oct 06 10:14:54 crc kubenswrapper[4824]: I1006 10:14:54.215260 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Oct 06 10:14:54 crc kubenswrapper[4824]: I1006 10:14:54.476642 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-t57w8"] Oct 06 10:14:54 crc kubenswrapper[4824]: I1006 10:14:54.478518 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-t57w8" Oct 06 10:14:54 crc kubenswrapper[4824]: I1006 10:14:54.493416 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-t57w8"] Oct 06 10:14:54 crc kubenswrapper[4824]: I1006 10:14:54.508192 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Oct 06 10:14:54 crc kubenswrapper[4824]: I1006 10:14:54.562810 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2x7cj\" (UniqueName: \"kubernetes.io/projected/c9d76f42-5e79-4057-8823-f80071669be7-kube-api-access-2x7cj\") pod \"cinder-db-create-t57w8\" (UID: \"c9d76f42-5e79-4057-8823-f80071669be7\") " pod="openstack/cinder-db-create-t57w8" Oct 06 10:14:54 crc kubenswrapper[4824]: I1006 10:14:54.576047 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-xzlzd"] Oct 06 10:14:54 crc kubenswrapper[4824]: I1006 10:14:54.579725 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-xzlzd" Oct 06 10:14:54 crc kubenswrapper[4824]: I1006 10:14:54.598145 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-xzlzd"] Oct 06 10:14:54 crc kubenswrapper[4824]: I1006 10:14:54.664737 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tf9zh\" (UniqueName: \"kubernetes.io/projected/ddcbc15f-464c-44de-9eb3-d1d35d48bacb-kube-api-access-tf9zh\") pod \"barbican-db-create-xzlzd\" (UID: \"ddcbc15f-464c-44de-9eb3-d1d35d48bacb\") " pod="openstack/barbican-db-create-xzlzd" Oct 06 10:14:54 crc kubenswrapper[4824]: I1006 10:14:54.665250 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2x7cj\" (UniqueName: \"kubernetes.io/projected/c9d76f42-5e79-4057-8823-f80071669be7-kube-api-access-2x7cj\") pod \"cinder-db-create-t57w8\" (UID: \"c9d76f42-5e79-4057-8823-f80071669be7\") " pod="openstack/cinder-db-create-t57w8" Oct 06 10:14:54 crc kubenswrapper[4824]: I1006 10:14:54.690047 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2x7cj\" (UniqueName: \"kubernetes.io/projected/c9d76f42-5e79-4057-8823-f80071669be7-kube-api-access-2x7cj\") pod \"cinder-db-create-t57w8\" (UID: \"c9d76f42-5e79-4057-8823-f80071669be7\") " pod="openstack/cinder-db-create-t57w8" Oct 06 10:14:54 crc kubenswrapper[4824]: I1006 10:14:54.767371 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tf9zh\" (UniqueName: \"kubernetes.io/projected/ddcbc15f-464c-44de-9eb3-d1d35d48bacb-kube-api-access-tf9zh\") pod \"barbican-db-create-xzlzd\" (UID: \"ddcbc15f-464c-44de-9eb3-d1d35d48bacb\") " pod="openstack/barbican-db-create-xzlzd" Oct 06 10:14:54 crc kubenswrapper[4824]: I1006 10:14:54.785156 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tf9zh\" (UniqueName: \"kubernetes.io/projected/ddcbc15f-464c-44de-9eb3-d1d35d48bacb-kube-api-access-tf9zh\") pod \"barbican-db-create-xzlzd\" (UID: \"ddcbc15f-464c-44de-9eb3-d1d35d48bacb\") " pod="openstack/barbican-db-create-xzlzd" Oct 06 10:14:54 crc kubenswrapper[4824]: I1006 10:14:54.803320 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-t57w8" Oct 06 10:14:54 crc kubenswrapper[4824]: I1006 10:14:54.868758 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-nmf65"] Oct 06 10:14:54 crc kubenswrapper[4824]: I1006 10:14:54.875580 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-nmf65" Oct 06 10:14:54 crc kubenswrapper[4824]: I1006 10:14:54.895825 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-nmf65"] Oct 06 10:14:54 crc kubenswrapper[4824]: I1006 10:14:54.906285 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-xzlzd" Oct 06 10:14:54 crc kubenswrapper[4824]: I1006 10:14:54.966284 4824 generic.go:334] "Generic (PLEG): container finished" podID="20c54a88-ce86-425a-b12c-cebd09083d14" containerID="1b9e78c64608e0c16a1937508d6125f8c093ac04a3f603d41ea1089ecde42934" exitCode=0 Oct 06 10:14:54 crc kubenswrapper[4824]: I1006 10:14:54.966787 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-5rvk5" event={"ID":"20c54a88-ce86-425a-b12c-cebd09083d14","Type":"ContainerDied","Data":"1b9e78c64608e0c16a1937508d6125f8c093ac04a3f603d41ea1089ecde42934"} Oct 06 10:14:54 crc kubenswrapper[4824]: I1006 10:14:54.968786 4824 generic.go:334] "Generic (PLEG): container finished" podID="41bb1f2f-34c5-43a5-b416-fbb5052a9dd8" containerID="c57b55dd4e9a6b22fc4938cdb0ab8c79b1aec852bc7fe714759a448e5e41a598" exitCode=0 Oct 06 10:14:54 crc kubenswrapper[4824]: I1006 10:14:54.968837 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-48qsh-config-dntgb" event={"ID":"41bb1f2f-34c5-43a5-b416-fbb5052a9dd8","Type":"ContainerDied","Data":"c57b55dd4e9a6b22fc4938cdb0ab8c79b1aec852bc7fe714759a448e5e41a598"} Oct 06 10:14:54 crc kubenswrapper[4824]: I1006 10:14:54.968887 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-48qsh-config-dntgb" event={"ID":"41bb1f2f-34c5-43a5-b416-fbb5052a9dd8","Type":"ContainerStarted","Data":"b561b0dbf690d37a94cb0ad01bc228f59e0c7ff399805379684198972e067d49"} Oct 06 10:14:54 crc kubenswrapper[4824]: I1006 10:14:54.974301 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6mghc\" (UniqueName: \"kubernetes.io/projected/ceec1f0f-b38f-459c-99fe-2bf0660b8777-kube-api-access-6mghc\") pod \"neutron-db-create-nmf65\" (UID: \"ceec1f0f-b38f-459c-99fe-2bf0660b8777\") " pod="openstack/neutron-db-create-nmf65" Oct 06 10:14:55 crc kubenswrapper[4824]: I1006 10:14:55.076152 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6mghc\" (UniqueName: \"kubernetes.io/projected/ceec1f0f-b38f-459c-99fe-2bf0660b8777-kube-api-access-6mghc\") pod \"neutron-db-create-nmf65\" (UID: \"ceec1f0f-b38f-459c-99fe-2bf0660b8777\") " pod="openstack/neutron-db-create-nmf65" Oct 06 10:14:55 crc kubenswrapper[4824]: I1006 10:14:55.095484 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6mghc\" (UniqueName: \"kubernetes.io/projected/ceec1f0f-b38f-459c-99fe-2bf0660b8777-kube-api-access-6mghc\") pod \"neutron-db-create-nmf65\" (UID: \"ceec1f0f-b38f-459c-99fe-2bf0660b8777\") " pod="openstack/neutron-db-create-nmf65" Oct 06 10:14:55 crc kubenswrapper[4824]: I1006 10:14:55.260965 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-nmf65" Oct 06 10:14:55 crc kubenswrapper[4824]: I1006 10:14:55.298494 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-t57w8"] Oct 06 10:14:55 crc kubenswrapper[4824]: I1006 10:14:55.415965 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-xzlzd"] Oct 06 10:14:55 crc kubenswrapper[4824]: W1006 10:14:55.421744 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podddcbc15f_464c_44de_9eb3_d1d35d48bacb.slice/crio-52fe3f08d4600e3eafa41262f832f375642e21f58cf0a6bb203ea70432df1207 WatchSource:0}: Error finding container 52fe3f08d4600e3eafa41262f832f375642e21f58cf0a6bb203ea70432df1207: Status 404 returned error can't find the container with id 52fe3f08d4600e3eafa41262f832f375642e21f58cf0a6bb203ea70432df1207 Oct 06 10:14:55 crc kubenswrapper[4824]: I1006 10:14:55.531575 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-nmf65"] Oct 06 10:14:55 crc kubenswrapper[4824]: W1006 10:14:55.535778 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podceec1f0f_b38f_459c_99fe_2bf0660b8777.slice/crio-153f4712fda5eb7507351961c9d660ea9cb028dcf963065728027441ad807b80 WatchSource:0}: Error finding container 153f4712fda5eb7507351961c9d660ea9cb028dcf963065728027441ad807b80: Status 404 returned error can't find the container with id 153f4712fda5eb7507351961c9d660ea9cb028dcf963065728027441ad807b80 Oct 06 10:14:55 crc kubenswrapper[4824]: I1006 10:14:55.982059 4824 generic.go:334] "Generic (PLEG): container finished" podID="ddcbc15f-464c-44de-9eb3-d1d35d48bacb" containerID="c20ec4c680dba96cfce21e0d3a4afec1174331e4d57303bda1d4da5b2de1ef46" exitCode=0 Oct 06 10:14:55 crc kubenswrapper[4824]: I1006 10:14:55.982156 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-xzlzd" event={"ID":"ddcbc15f-464c-44de-9eb3-d1d35d48bacb","Type":"ContainerDied","Data":"c20ec4c680dba96cfce21e0d3a4afec1174331e4d57303bda1d4da5b2de1ef46"} Oct 06 10:14:55 crc kubenswrapper[4824]: I1006 10:14:55.982198 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-xzlzd" event={"ID":"ddcbc15f-464c-44de-9eb3-d1d35d48bacb","Type":"ContainerStarted","Data":"52fe3f08d4600e3eafa41262f832f375642e21f58cf0a6bb203ea70432df1207"} Oct 06 10:14:55 crc kubenswrapper[4824]: I1006 10:14:55.985089 4824 generic.go:334] "Generic (PLEG): container finished" podID="c9d76f42-5e79-4057-8823-f80071669be7" containerID="e9aa345b92d318f37d64b759ec7893557cd190fb224b6cfbba3902cf664d657e" exitCode=0 Oct 06 10:14:55 crc kubenswrapper[4824]: I1006 10:14:55.985176 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-t57w8" event={"ID":"c9d76f42-5e79-4057-8823-f80071669be7","Type":"ContainerDied","Data":"e9aa345b92d318f37d64b759ec7893557cd190fb224b6cfbba3902cf664d657e"} Oct 06 10:14:55 crc kubenswrapper[4824]: I1006 10:14:55.985209 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-t57w8" event={"ID":"c9d76f42-5e79-4057-8823-f80071669be7","Type":"ContainerStarted","Data":"9b0fc1c2f2e01aa831e435e14b2352322d1bfbf416cb9c874d0c3b49d348f321"} Oct 06 10:14:55 crc kubenswrapper[4824]: I1006 10:14:55.987808 4824 generic.go:334] "Generic (PLEG): container finished" podID="ceec1f0f-b38f-459c-99fe-2bf0660b8777" containerID="aab8d3bb855e53d592920c6636a15e3d7292c18be7bd1a3f5740ad628def69bc" exitCode=0 Oct 06 10:14:55 crc kubenswrapper[4824]: I1006 10:14:55.988124 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-nmf65" event={"ID":"ceec1f0f-b38f-459c-99fe-2bf0660b8777","Type":"ContainerDied","Data":"aab8d3bb855e53d592920c6636a15e3d7292c18be7bd1a3f5740ad628def69bc"} Oct 06 10:14:55 crc kubenswrapper[4824]: I1006 10:14:55.988169 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-nmf65" event={"ID":"ceec1f0f-b38f-459c-99fe-2bf0660b8777","Type":"ContainerStarted","Data":"153f4712fda5eb7507351961c9d660ea9cb028dcf963065728027441ad807b80"} Oct 06 10:14:56 crc kubenswrapper[4824]: I1006 10:14:56.444855 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-48qsh-config-dntgb" Oct 06 10:14:56 crc kubenswrapper[4824]: I1006 10:14:56.451641 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-5rvk5" Oct 06 10:14:56 crc kubenswrapper[4824]: I1006 10:14:56.504717 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/41bb1f2f-34c5-43a5-b416-fbb5052a9dd8-var-log-ovn\") pod \"41bb1f2f-34c5-43a5-b416-fbb5052a9dd8\" (UID: \"41bb1f2f-34c5-43a5-b416-fbb5052a9dd8\") " Oct 06 10:14:56 crc kubenswrapper[4824]: I1006 10:14:56.504777 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hxjrf\" (UniqueName: \"kubernetes.io/projected/41bb1f2f-34c5-43a5-b416-fbb5052a9dd8-kube-api-access-hxjrf\") pod \"41bb1f2f-34c5-43a5-b416-fbb5052a9dd8\" (UID: \"41bb1f2f-34c5-43a5-b416-fbb5052a9dd8\") " Oct 06 10:14:56 crc kubenswrapper[4824]: I1006 10:14:56.505061 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/41bb1f2f-34c5-43a5-b416-fbb5052a9dd8-var-run\") pod \"41bb1f2f-34c5-43a5-b416-fbb5052a9dd8\" (UID: \"41bb1f2f-34c5-43a5-b416-fbb5052a9dd8\") " Oct 06 10:14:56 crc kubenswrapper[4824]: I1006 10:14:56.505094 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/41bb1f2f-34c5-43a5-b416-fbb5052a9dd8-additional-scripts\") pod \"41bb1f2f-34c5-43a5-b416-fbb5052a9dd8\" (UID: \"41bb1f2f-34c5-43a5-b416-fbb5052a9dd8\") " Oct 06 10:14:56 crc kubenswrapper[4824]: I1006 10:14:56.505097 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/41bb1f2f-34c5-43a5-b416-fbb5052a9dd8-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "41bb1f2f-34c5-43a5-b416-fbb5052a9dd8" (UID: "41bb1f2f-34c5-43a5-b416-fbb5052a9dd8"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 06 10:14:56 crc kubenswrapper[4824]: I1006 10:14:56.505146 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/41bb1f2f-34c5-43a5-b416-fbb5052a9dd8-scripts\") pod \"41bb1f2f-34c5-43a5-b416-fbb5052a9dd8\" (UID: \"41bb1f2f-34c5-43a5-b416-fbb5052a9dd8\") " Oct 06 10:14:56 crc kubenswrapper[4824]: I1006 10:14:56.505162 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/41bb1f2f-34c5-43a5-b416-fbb5052a9dd8-var-run" (OuterVolumeSpecName: "var-run") pod "41bb1f2f-34c5-43a5-b416-fbb5052a9dd8" (UID: "41bb1f2f-34c5-43a5-b416-fbb5052a9dd8"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 06 10:14:56 crc kubenswrapper[4824]: I1006 10:14:56.505242 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/41bb1f2f-34c5-43a5-b416-fbb5052a9dd8-var-run-ovn\") pod \"41bb1f2f-34c5-43a5-b416-fbb5052a9dd8\" (UID: \"41bb1f2f-34c5-43a5-b416-fbb5052a9dd8\") " Oct 06 10:14:56 crc kubenswrapper[4824]: I1006 10:14:56.505305 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q96xk\" (UniqueName: \"kubernetes.io/projected/20c54a88-ce86-425a-b12c-cebd09083d14-kube-api-access-q96xk\") pod \"20c54a88-ce86-425a-b12c-cebd09083d14\" (UID: \"20c54a88-ce86-425a-b12c-cebd09083d14\") " Oct 06 10:14:56 crc kubenswrapper[4824]: I1006 10:14:56.505373 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/41bb1f2f-34c5-43a5-b416-fbb5052a9dd8-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "41bb1f2f-34c5-43a5-b416-fbb5052a9dd8" (UID: "41bb1f2f-34c5-43a5-b416-fbb5052a9dd8"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 06 10:14:56 crc kubenswrapper[4824]: I1006 10:14:56.506069 4824 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/41bb1f2f-34c5-43a5-b416-fbb5052a9dd8-var-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 06 10:14:56 crc kubenswrapper[4824]: I1006 10:14:56.506097 4824 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/41bb1f2f-34c5-43a5-b416-fbb5052a9dd8-var-log-ovn\") on node \"crc\" DevicePath \"\"" Oct 06 10:14:56 crc kubenswrapper[4824]: I1006 10:14:56.506106 4824 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/41bb1f2f-34c5-43a5-b416-fbb5052a9dd8-var-run\") on node \"crc\" DevicePath \"\"" Oct 06 10:14:56 crc kubenswrapper[4824]: I1006 10:14:56.506188 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/41bb1f2f-34c5-43a5-b416-fbb5052a9dd8-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "41bb1f2f-34c5-43a5-b416-fbb5052a9dd8" (UID: "41bb1f2f-34c5-43a5-b416-fbb5052a9dd8"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:14:56 crc kubenswrapper[4824]: I1006 10:14:56.506528 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/41bb1f2f-34c5-43a5-b416-fbb5052a9dd8-scripts" (OuterVolumeSpecName: "scripts") pod "41bb1f2f-34c5-43a5-b416-fbb5052a9dd8" (UID: "41bb1f2f-34c5-43a5-b416-fbb5052a9dd8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:14:56 crc kubenswrapper[4824]: I1006 10:14:56.515947 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/41bb1f2f-34c5-43a5-b416-fbb5052a9dd8-kube-api-access-hxjrf" (OuterVolumeSpecName: "kube-api-access-hxjrf") pod "41bb1f2f-34c5-43a5-b416-fbb5052a9dd8" (UID: "41bb1f2f-34c5-43a5-b416-fbb5052a9dd8"). InnerVolumeSpecName "kube-api-access-hxjrf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:14:56 crc kubenswrapper[4824]: I1006 10:14:56.516854 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20c54a88-ce86-425a-b12c-cebd09083d14-kube-api-access-q96xk" (OuterVolumeSpecName: "kube-api-access-q96xk") pod "20c54a88-ce86-425a-b12c-cebd09083d14" (UID: "20c54a88-ce86-425a-b12c-cebd09083d14"). InnerVolumeSpecName "kube-api-access-q96xk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:14:56 crc kubenswrapper[4824]: I1006 10:14:56.607334 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hxjrf\" (UniqueName: \"kubernetes.io/projected/41bb1f2f-34c5-43a5-b416-fbb5052a9dd8-kube-api-access-hxjrf\") on node \"crc\" DevicePath \"\"" Oct 06 10:14:56 crc kubenswrapper[4824]: I1006 10:14:56.607382 4824 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/41bb1f2f-34c5-43a5-b416-fbb5052a9dd8-additional-scripts\") on node \"crc\" DevicePath \"\"" Oct 06 10:14:56 crc kubenswrapper[4824]: I1006 10:14:56.607394 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/41bb1f2f-34c5-43a5-b416-fbb5052a9dd8-scripts\") on node \"crc\" DevicePath \"\"" Oct 06 10:14:56 crc kubenswrapper[4824]: I1006 10:14:56.607405 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q96xk\" (UniqueName: \"kubernetes.io/projected/20c54a88-ce86-425a-b12c-cebd09083d14-kube-api-access-q96xk\") on node \"crc\" DevicePath \"\"" Oct 06 10:14:57 crc kubenswrapper[4824]: I1006 10:14:57.000237 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-5rvk5" Oct 06 10:14:57 crc kubenswrapper[4824]: I1006 10:14:57.000436 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-5rvk5" event={"ID":"20c54a88-ce86-425a-b12c-cebd09083d14","Type":"ContainerDied","Data":"22d7e53f31dc36b44599a26a7625f105a50224c0dca6c2503d6186e260a27117"} Oct 06 10:14:57 crc kubenswrapper[4824]: I1006 10:14:57.001944 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="22d7e53f31dc36b44599a26a7625f105a50224c0dca6c2503d6186e260a27117" Oct 06 10:14:57 crc kubenswrapper[4824]: I1006 10:14:57.003058 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-48qsh-config-dntgb" Oct 06 10:14:57 crc kubenswrapper[4824]: I1006 10:14:57.002957 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-48qsh-config-dntgb" event={"ID":"41bb1f2f-34c5-43a5-b416-fbb5052a9dd8","Type":"ContainerDied","Data":"b561b0dbf690d37a94cb0ad01bc228f59e0c7ff399805379684198972e067d49"} Oct 06 10:14:57 crc kubenswrapper[4824]: I1006 10:14:57.003352 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b561b0dbf690d37a94cb0ad01bc228f59e0c7ff399805379684198972e067d49" Oct 06 10:14:57 crc kubenswrapper[4824]: I1006 10:14:57.187223 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-bmgl4"] Oct 06 10:14:57 crc kubenswrapper[4824]: E1006 10:14:57.187609 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20c54a88-ce86-425a-b12c-cebd09083d14" containerName="mariadb-database-create" Oct 06 10:14:57 crc kubenswrapper[4824]: I1006 10:14:57.187629 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="20c54a88-ce86-425a-b12c-cebd09083d14" containerName="mariadb-database-create" Oct 06 10:14:57 crc kubenswrapper[4824]: E1006 10:14:57.187642 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41bb1f2f-34c5-43a5-b416-fbb5052a9dd8" containerName="ovn-config" Oct 06 10:14:57 crc kubenswrapper[4824]: I1006 10:14:57.187649 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="41bb1f2f-34c5-43a5-b416-fbb5052a9dd8" containerName="ovn-config" Oct 06 10:14:57 crc kubenswrapper[4824]: I1006 10:14:57.187840 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="41bb1f2f-34c5-43a5-b416-fbb5052a9dd8" containerName="ovn-config" Oct 06 10:14:57 crc kubenswrapper[4824]: I1006 10:14:57.187873 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="20c54a88-ce86-425a-b12c-cebd09083d14" containerName="mariadb-database-create" Oct 06 10:14:57 crc kubenswrapper[4824]: I1006 10:14:57.188636 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-bmgl4" Oct 06 10:14:57 crc kubenswrapper[4824]: E1006 10:14:57.252382 4824 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod20c54a88_ce86_425a_b12c_cebd09083d14.slice/crio-22d7e53f31dc36b44599a26a7625f105a50224c0dca6c2503d6186e260a27117\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod41bb1f2f_34c5_43a5_b416_fbb5052a9dd8.slice\": RecentStats: unable to find data in memory cache]" Oct 06 10:14:57 crc kubenswrapper[4824]: I1006 10:14:57.276298 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-bmgl4"] Oct 06 10:14:57 crc kubenswrapper[4824]: I1006 10:14:57.320726 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mjpks\" (UniqueName: \"kubernetes.io/projected/256cff80-df3a-481d-a994-05b49d8996c1-kube-api-access-mjpks\") pod \"keystone-db-create-bmgl4\" (UID: \"256cff80-df3a-481d-a994-05b49d8996c1\") " pod="openstack/keystone-db-create-bmgl4" Oct 06 10:14:57 crc kubenswrapper[4824]: I1006 10:14:57.389633 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-xzlzd" Oct 06 10:14:57 crc kubenswrapper[4824]: I1006 10:14:57.424375 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tf9zh\" (UniqueName: \"kubernetes.io/projected/ddcbc15f-464c-44de-9eb3-d1d35d48bacb-kube-api-access-tf9zh\") pod \"ddcbc15f-464c-44de-9eb3-d1d35d48bacb\" (UID: \"ddcbc15f-464c-44de-9eb3-d1d35d48bacb\") " Oct 06 10:14:57 crc kubenswrapper[4824]: I1006 10:14:57.425377 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mjpks\" (UniqueName: \"kubernetes.io/projected/256cff80-df3a-481d-a994-05b49d8996c1-kube-api-access-mjpks\") pod \"keystone-db-create-bmgl4\" (UID: \"256cff80-df3a-481d-a994-05b49d8996c1\") " pod="openstack/keystone-db-create-bmgl4" Oct 06 10:14:57 crc kubenswrapper[4824]: I1006 10:14:57.435248 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ddcbc15f-464c-44de-9eb3-d1d35d48bacb-kube-api-access-tf9zh" (OuterVolumeSpecName: "kube-api-access-tf9zh") pod "ddcbc15f-464c-44de-9eb3-d1d35d48bacb" (UID: "ddcbc15f-464c-44de-9eb3-d1d35d48bacb"). InnerVolumeSpecName "kube-api-access-tf9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:14:57 crc kubenswrapper[4824]: I1006 10:14:57.447755 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mjpks\" (UniqueName: \"kubernetes.io/projected/256cff80-df3a-481d-a994-05b49d8996c1-kube-api-access-mjpks\") pod \"keystone-db-create-bmgl4\" (UID: \"256cff80-df3a-481d-a994-05b49d8996c1\") " pod="openstack/keystone-db-create-bmgl4" Oct 06 10:14:57 crc kubenswrapper[4824]: I1006 10:14:57.542121 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-bmgl4" Oct 06 10:14:57 crc kubenswrapper[4824]: I1006 10:14:57.547344 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tf9zh\" (UniqueName: \"kubernetes.io/projected/ddcbc15f-464c-44de-9eb3-d1d35d48bacb-kube-api-access-tf9zh\") on node \"crc\" DevicePath \"\"" Oct 06 10:14:57 crc kubenswrapper[4824]: I1006 10:14:57.607383 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-48qsh-config-dntgb"] Oct 06 10:14:57 crc kubenswrapper[4824]: I1006 10:14:57.624055 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-48qsh-config-dntgb"] Oct 06 10:14:57 crc kubenswrapper[4824]: I1006 10:14:57.645755 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-nmf65" Oct 06 10:14:57 crc kubenswrapper[4824]: I1006 10:14:57.672085 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-4a6e-account-create-cx6mn"] Oct 06 10:14:57 crc kubenswrapper[4824]: E1006 10:14:57.672609 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ddcbc15f-464c-44de-9eb3-d1d35d48bacb" containerName="mariadb-database-create" Oct 06 10:14:57 crc kubenswrapper[4824]: I1006 10:14:57.672625 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="ddcbc15f-464c-44de-9eb3-d1d35d48bacb" containerName="mariadb-database-create" Oct 06 10:14:57 crc kubenswrapper[4824]: E1006 10:14:57.672637 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ceec1f0f-b38f-459c-99fe-2bf0660b8777" containerName="mariadb-database-create" Oct 06 10:14:57 crc kubenswrapper[4824]: I1006 10:14:57.672645 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="ceec1f0f-b38f-459c-99fe-2bf0660b8777" containerName="mariadb-database-create" Oct 06 10:14:57 crc kubenswrapper[4824]: I1006 10:14:57.672851 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="ceec1f0f-b38f-459c-99fe-2bf0660b8777" containerName="mariadb-database-create" Oct 06 10:14:57 crc kubenswrapper[4824]: I1006 10:14:57.672881 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="ddcbc15f-464c-44de-9eb3-d1d35d48bacb" containerName="mariadb-database-create" Oct 06 10:14:57 crc kubenswrapper[4824]: I1006 10:14:57.673058 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-t57w8" Oct 06 10:14:57 crc kubenswrapper[4824]: I1006 10:14:57.673534 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-4a6e-account-create-cx6mn" Oct 06 10:14:57 crc kubenswrapper[4824]: I1006 10:14:57.679506 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Oct 06 10:14:57 crc kubenswrapper[4824]: I1006 10:14:57.696966 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-4a6e-account-create-cx6mn"] Oct 06 10:14:57 crc kubenswrapper[4824]: I1006 10:14:57.752458 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2x7cj\" (UniqueName: \"kubernetes.io/projected/c9d76f42-5e79-4057-8823-f80071669be7-kube-api-access-2x7cj\") pod \"c9d76f42-5e79-4057-8823-f80071669be7\" (UID: \"c9d76f42-5e79-4057-8823-f80071669be7\") " Oct 06 10:14:57 crc kubenswrapper[4824]: I1006 10:14:57.752646 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6mghc\" (UniqueName: \"kubernetes.io/projected/ceec1f0f-b38f-459c-99fe-2bf0660b8777-kube-api-access-6mghc\") pod \"ceec1f0f-b38f-459c-99fe-2bf0660b8777\" (UID: \"ceec1f0f-b38f-459c-99fe-2bf0660b8777\") " Oct 06 10:14:57 crc kubenswrapper[4824]: I1006 10:14:57.753008 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xcxhd\" (UniqueName: \"kubernetes.io/projected/7b66b5d9-1c7c-4f7d-aa8d-cba5b9f6c2f2-kube-api-access-xcxhd\") pod \"placement-4a6e-account-create-cx6mn\" (UID: \"7b66b5d9-1c7c-4f7d-aa8d-cba5b9f6c2f2\") " pod="openstack/placement-4a6e-account-create-cx6mn" Oct 06 10:14:57 crc kubenswrapper[4824]: I1006 10:14:57.756041 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c9d76f42-5e79-4057-8823-f80071669be7-kube-api-access-2x7cj" (OuterVolumeSpecName: "kube-api-access-2x7cj") pod "c9d76f42-5e79-4057-8823-f80071669be7" (UID: "c9d76f42-5e79-4057-8823-f80071669be7"). InnerVolumeSpecName "kube-api-access-2x7cj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:14:57 crc kubenswrapper[4824]: I1006 10:14:57.756338 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ceec1f0f-b38f-459c-99fe-2bf0660b8777-kube-api-access-6mghc" (OuterVolumeSpecName: "kube-api-access-6mghc") pod "ceec1f0f-b38f-459c-99fe-2bf0660b8777" (UID: "ceec1f0f-b38f-459c-99fe-2bf0660b8777"). InnerVolumeSpecName "kube-api-access-6mghc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:14:57 crc kubenswrapper[4824]: I1006 10:14:57.854941 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xcxhd\" (UniqueName: \"kubernetes.io/projected/7b66b5d9-1c7c-4f7d-aa8d-cba5b9f6c2f2-kube-api-access-xcxhd\") pod \"placement-4a6e-account-create-cx6mn\" (UID: \"7b66b5d9-1c7c-4f7d-aa8d-cba5b9f6c2f2\") " pod="openstack/placement-4a6e-account-create-cx6mn" Oct 06 10:14:57 crc kubenswrapper[4824]: I1006 10:14:57.855135 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6mghc\" (UniqueName: \"kubernetes.io/projected/ceec1f0f-b38f-459c-99fe-2bf0660b8777-kube-api-access-6mghc\") on node \"crc\" DevicePath \"\"" Oct 06 10:14:57 crc kubenswrapper[4824]: I1006 10:14:57.855147 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2x7cj\" (UniqueName: \"kubernetes.io/projected/c9d76f42-5e79-4057-8823-f80071669be7-kube-api-access-2x7cj\") on node \"crc\" DevicePath \"\"" Oct 06 10:14:57 crc kubenswrapper[4824]: I1006 10:14:57.875392 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xcxhd\" (UniqueName: \"kubernetes.io/projected/7b66b5d9-1c7c-4f7d-aa8d-cba5b9f6c2f2-kube-api-access-xcxhd\") pod \"placement-4a6e-account-create-cx6mn\" (UID: \"7b66b5d9-1c7c-4f7d-aa8d-cba5b9f6c2f2\") " pod="openstack/placement-4a6e-account-create-cx6mn" Oct 06 10:14:57 crc kubenswrapper[4824]: I1006 10:14:57.920578 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-48qsh" Oct 06 10:14:58 crc kubenswrapper[4824]: I1006 10:14:58.004499 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-4a6e-account-create-cx6mn" Oct 06 10:14:58 crc kubenswrapper[4824]: I1006 10:14:58.011463 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-nmf65" event={"ID":"ceec1f0f-b38f-459c-99fe-2bf0660b8777","Type":"ContainerDied","Data":"153f4712fda5eb7507351961c9d660ea9cb028dcf963065728027441ad807b80"} Oct 06 10:14:58 crc kubenswrapper[4824]: I1006 10:14:58.011507 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="153f4712fda5eb7507351961c9d660ea9cb028dcf963065728027441ad807b80" Oct 06 10:14:58 crc kubenswrapper[4824]: I1006 10:14:58.011567 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-nmf65" Oct 06 10:14:58 crc kubenswrapper[4824]: I1006 10:14:58.014271 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-xzlzd" Oct 06 10:14:58 crc kubenswrapper[4824]: I1006 10:14:58.014302 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-xzlzd" event={"ID":"ddcbc15f-464c-44de-9eb3-d1d35d48bacb","Type":"ContainerDied","Data":"52fe3f08d4600e3eafa41262f832f375642e21f58cf0a6bb203ea70432df1207"} Oct 06 10:14:58 crc kubenswrapper[4824]: I1006 10:14:58.014355 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="52fe3f08d4600e3eafa41262f832f375642e21f58cf0a6bb203ea70432df1207" Oct 06 10:14:58 crc kubenswrapper[4824]: I1006 10:14:58.016574 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-t57w8" event={"ID":"c9d76f42-5e79-4057-8823-f80071669be7","Type":"ContainerDied","Data":"9b0fc1c2f2e01aa831e435e14b2352322d1bfbf416cb9c874d0c3b49d348f321"} Oct 06 10:14:58 crc kubenswrapper[4824]: I1006 10:14:58.016608 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9b0fc1c2f2e01aa831e435e14b2352322d1bfbf416cb9c874d0c3b49d348f321" Oct 06 10:14:58 crc kubenswrapper[4824]: I1006 10:14:58.016676 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-t57w8" Oct 06 10:14:58 crc kubenswrapper[4824]: I1006 10:14:58.067487 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-bmgl4"] Oct 06 10:14:58 crc kubenswrapper[4824]: W1006 10:14:58.074649 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod256cff80_df3a_481d_a994_05b49d8996c1.slice/crio-46252b5138b393087d3abcf32792637ae974eef72c3c6b64d301ba96aa02b688 WatchSource:0}: Error finding container 46252b5138b393087d3abcf32792637ae974eef72c3c6b64d301ba96aa02b688: Status 404 returned error can't find the container with id 46252b5138b393087d3abcf32792637ae974eef72c3c6b64d301ba96aa02b688 Oct 06 10:14:58 crc kubenswrapper[4824]: I1006 10:14:58.488559 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-4a6e-account-create-cx6mn"] Oct 06 10:14:58 crc kubenswrapper[4824]: W1006 10:14:58.491376 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7b66b5d9_1c7c_4f7d_aa8d_cba5b9f6c2f2.slice/crio-5d86b6fa5548dd4ddca95b0f89da70b8ab569597462fa85f4affd1b0461f2fb7 WatchSource:0}: Error finding container 5d86b6fa5548dd4ddca95b0f89da70b8ab569597462fa85f4affd1b0461f2fb7: Status 404 returned error can't find the container with id 5d86b6fa5548dd4ddca95b0f89da70b8ab569597462fa85f4affd1b0461f2fb7 Oct 06 10:14:59 crc kubenswrapper[4824]: I1006 10:14:59.032132 4824 generic.go:334] "Generic (PLEG): container finished" podID="256cff80-df3a-481d-a994-05b49d8996c1" containerID="03a50ad9606a91f866ae2e6342d1b63486e17701e34c1c5ab2d58988ccb2b1a1" exitCode=0 Oct 06 10:14:59 crc kubenswrapper[4824]: I1006 10:14:59.032202 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-bmgl4" event={"ID":"256cff80-df3a-481d-a994-05b49d8996c1","Type":"ContainerDied","Data":"03a50ad9606a91f866ae2e6342d1b63486e17701e34c1c5ab2d58988ccb2b1a1"} Oct 06 10:14:59 crc kubenswrapper[4824]: I1006 10:14:59.032659 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-bmgl4" event={"ID":"256cff80-df3a-481d-a994-05b49d8996c1","Type":"ContainerStarted","Data":"46252b5138b393087d3abcf32792637ae974eef72c3c6b64d301ba96aa02b688"} Oct 06 10:14:59 crc kubenswrapper[4824]: I1006 10:14:59.035266 4824 generic.go:334] "Generic (PLEG): container finished" podID="7b66b5d9-1c7c-4f7d-aa8d-cba5b9f6c2f2" containerID="87c81576adcaa23fb9324494c703d1c3f84bf4dfc82a548e6d873feb13063e88" exitCode=0 Oct 06 10:14:59 crc kubenswrapper[4824]: I1006 10:14:59.035330 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-4a6e-account-create-cx6mn" event={"ID":"7b66b5d9-1c7c-4f7d-aa8d-cba5b9f6c2f2","Type":"ContainerDied","Data":"87c81576adcaa23fb9324494c703d1c3f84bf4dfc82a548e6d873feb13063e88"} Oct 06 10:14:59 crc kubenswrapper[4824]: I1006 10:14:59.035397 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-4a6e-account-create-cx6mn" event={"ID":"7b66b5d9-1c7c-4f7d-aa8d-cba5b9f6c2f2","Type":"ContainerStarted","Data":"5d86b6fa5548dd4ddca95b0f89da70b8ab569597462fa85f4affd1b0461f2fb7"} Oct 06 10:14:59 crc kubenswrapper[4824]: I1006 10:14:59.292305 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="41bb1f2f-34c5-43a5-b416-fbb5052a9dd8" path="/var/lib/kubelet/pods/41bb1f2f-34c5-43a5-b416-fbb5052a9dd8/volumes" Oct 06 10:15:00 crc kubenswrapper[4824]: I1006 10:15:00.152237 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29329095-vgrt8"] Oct 06 10:15:00 crc kubenswrapper[4824]: E1006 10:15:00.152600 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9d76f42-5e79-4057-8823-f80071669be7" containerName="mariadb-database-create" Oct 06 10:15:00 crc kubenswrapper[4824]: I1006 10:15:00.152612 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9d76f42-5e79-4057-8823-f80071669be7" containerName="mariadb-database-create" Oct 06 10:15:00 crc kubenswrapper[4824]: I1006 10:15:00.152791 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="c9d76f42-5e79-4057-8823-f80071669be7" containerName="mariadb-database-create" Oct 06 10:15:00 crc kubenswrapper[4824]: I1006 10:15:00.153442 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29329095-vgrt8" Oct 06 10:15:00 crc kubenswrapper[4824]: I1006 10:15:00.157159 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 06 10:15:00 crc kubenswrapper[4824]: I1006 10:15:00.157410 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 06 10:15:00 crc kubenswrapper[4824]: I1006 10:15:00.210054 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29329095-vgrt8"] Oct 06 10:15:00 crc kubenswrapper[4824]: I1006 10:15:00.309733 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/29e4895c-1440-496e-a3f8-06635f103456-config-volume\") pod \"collect-profiles-29329095-vgrt8\" (UID: \"29e4895c-1440-496e-a3f8-06635f103456\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329095-vgrt8" Oct 06 10:15:00 crc kubenswrapper[4824]: I1006 10:15:00.310294 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/29e4895c-1440-496e-a3f8-06635f103456-secret-volume\") pod \"collect-profiles-29329095-vgrt8\" (UID: \"29e4895c-1440-496e-a3f8-06635f103456\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329095-vgrt8" Oct 06 10:15:00 crc kubenswrapper[4824]: I1006 10:15:00.310407 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-98mbl\" (UniqueName: \"kubernetes.io/projected/29e4895c-1440-496e-a3f8-06635f103456-kube-api-access-98mbl\") pod \"collect-profiles-29329095-vgrt8\" (UID: \"29e4895c-1440-496e-a3f8-06635f103456\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329095-vgrt8" Oct 06 10:15:00 crc kubenswrapper[4824]: I1006 10:15:00.411383 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-98mbl\" (UniqueName: \"kubernetes.io/projected/29e4895c-1440-496e-a3f8-06635f103456-kube-api-access-98mbl\") pod \"collect-profiles-29329095-vgrt8\" (UID: \"29e4895c-1440-496e-a3f8-06635f103456\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329095-vgrt8" Oct 06 10:15:00 crc kubenswrapper[4824]: I1006 10:15:00.411463 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/29e4895c-1440-496e-a3f8-06635f103456-config-volume\") pod \"collect-profiles-29329095-vgrt8\" (UID: \"29e4895c-1440-496e-a3f8-06635f103456\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329095-vgrt8" Oct 06 10:15:00 crc kubenswrapper[4824]: I1006 10:15:00.411501 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/29e4895c-1440-496e-a3f8-06635f103456-secret-volume\") pod \"collect-profiles-29329095-vgrt8\" (UID: \"29e4895c-1440-496e-a3f8-06635f103456\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329095-vgrt8" Oct 06 10:15:00 crc kubenswrapper[4824]: I1006 10:15:00.412812 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/29e4895c-1440-496e-a3f8-06635f103456-config-volume\") pod \"collect-profiles-29329095-vgrt8\" (UID: \"29e4895c-1440-496e-a3f8-06635f103456\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329095-vgrt8" Oct 06 10:15:00 crc kubenswrapper[4824]: I1006 10:15:00.427431 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/29e4895c-1440-496e-a3f8-06635f103456-secret-volume\") pod \"collect-profiles-29329095-vgrt8\" (UID: \"29e4895c-1440-496e-a3f8-06635f103456\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329095-vgrt8" Oct 06 10:15:00 crc kubenswrapper[4824]: I1006 10:15:00.439693 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-98mbl\" (UniqueName: \"kubernetes.io/projected/29e4895c-1440-496e-a3f8-06635f103456-kube-api-access-98mbl\") pod \"collect-profiles-29329095-vgrt8\" (UID: \"29e4895c-1440-496e-a3f8-06635f103456\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329095-vgrt8" Oct 06 10:15:00 crc kubenswrapper[4824]: I1006 10:15:00.479275 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29329095-vgrt8" Oct 06 10:15:00 crc kubenswrapper[4824]: I1006 10:15:00.584210 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-bmgl4" Oct 06 10:15:00 crc kubenswrapper[4824]: I1006 10:15:00.630652 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-4a6e-account-create-cx6mn" Oct 06 10:15:00 crc kubenswrapper[4824]: I1006 10:15:00.728404 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcxhd\" (UniqueName: \"kubernetes.io/projected/7b66b5d9-1c7c-4f7d-aa8d-cba5b9f6c2f2-kube-api-access-xcxhd\") pod \"7b66b5d9-1c7c-4f7d-aa8d-cba5b9f6c2f2\" (UID: \"7b66b5d9-1c7c-4f7d-aa8d-cba5b9f6c2f2\") " Oct 06 10:15:00 crc kubenswrapper[4824]: I1006 10:15:00.728887 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mjpks\" (UniqueName: \"kubernetes.io/projected/256cff80-df3a-481d-a994-05b49d8996c1-kube-api-access-mjpks\") pod \"256cff80-df3a-481d-a994-05b49d8996c1\" (UID: \"256cff80-df3a-481d-a994-05b49d8996c1\") " Oct 06 10:15:00 crc kubenswrapper[4824]: I1006 10:15:00.734555 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7b66b5d9-1c7c-4f7d-aa8d-cba5b9f6c2f2-kube-api-access-xcxhd" (OuterVolumeSpecName: "kube-api-access-xcxhd") pod "7b66b5d9-1c7c-4f7d-aa8d-cba5b9f6c2f2" (UID: "7b66b5d9-1c7c-4f7d-aa8d-cba5b9f6c2f2"). InnerVolumeSpecName "kube-api-access-xcxhd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:15:00 crc kubenswrapper[4824]: I1006 10:15:00.735837 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/256cff80-df3a-481d-a994-05b49d8996c1-kube-api-access-mjpks" (OuterVolumeSpecName: "kube-api-access-mjpks") pod "256cff80-df3a-481d-a994-05b49d8996c1" (UID: "256cff80-df3a-481d-a994-05b49d8996c1"). InnerVolumeSpecName "kube-api-access-mjpks". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:15:00 crc kubenswrapper[4824]: I1006 10:15:00.831071 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcxhd\" (UniqueName: \"kubernetes.io/projected/7b66b5d9-1c7c-4f7d-aa8d-cba5b9f6c2f2-kube-api-access-xcxhd\") on node \"crc\" DevicePath \"\"" Oct 06 10:15:00 crc kubenswrapper[4824]: I1006 10:15:00.831122 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mjpks\" (UniqueName: \"kubernetes.io/projected/256cff80-df3a-481d-a994-05b49d8996c1-kube-api-access-mjpks\") on node \"crc\" DevicePath \"\"" Oct 06 10:15:01 crc kubenswrapper[4824]: I1006 10:15:01.006667 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29329095-vgrt8"] Oct 06 10:15:01 crc kubenswrapper[4824]: W1006 10:15:01.009791 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod29e4895c_1440_496e_a3f8_06635f103456.slice/crio-e62d6f756caeee22fb192b0e1deac1c9bc9441bcd3b5fe46437456b9a143f262 WatchSource:0}: Error finding container e62d6f756caeee22fb192b0e1deac1c9bc9441bcd3b5fe46437456b9a143f262: Status 404 returned error can't find the container with id e62d6f756caeee22fb192b0e1deac1c9bc9441bcd3b5fe46437456b9a143f262 Oct 06 10:15:01 crc kubenswrapper[4824]: I1006 10:15:01.079588 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-bmgl4" Oct 06 10:15:01 crc kubenswrapper[4824]: I1006 10:15:01.079621 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-bmgl4" event={"ID":"256cff80-df3a-481d-a994-05b49d8996c1","Type":"ContainerDied","Data":"46252b5138b393087d3abcf32792637ae974eef72c3c6b64d301ba96aa02b688"} Oct 06 10:15:01 crc kubenswrapper[4824]: I1006 10:15:01.079678 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="46252b5138b393087d3abcf32792637ae974eef72c3c6b64d301ba96aa02b688" Oct 06 10:15:01 crc kubenswrapper[4824]: I1006 10:15:01.081884 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-4a6e-account-create-cx6mn" event={"ID":"7b66b5d9-1c7c-4f7d-aa8d-cba5b9f6c2f2","Type":"ContainerDied","Data":"5d86b6fa5548dd4ddca95b0f89da70b8ab569597462fa85f4affd1b0461f2fb7"} Oct 06 10:15:01 crc kubenswrapper[4824]: I1006 10:15:01.081941 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5d86b6fa5548dd4ddca95b0f89da70b8ab569597462fa85f4affd1b0461f2fb7" Oct 06 10:15:01 crc kubenswrapper[4824]: I1006 10:15:01.081963 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-4a6e-account-create-cx6mn" Oct 06 10:15:01 crc kubenswrapper[4824]: I1006 10:15:01.083641 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29329095-vgrt8" event={"ID":"29e4895c-1440-496e-a3f8-06635f103456","Type":"ContainerStarted","Data":"e62d6f756caeee22fb192b0e1deac1c9bc9441bcd3b5fe46437456b9a143f262"} Oct 06 10:15:02 crc kubenswrapper[4824]: I1006 10:15:02.097515 4824 generic.go:334] "Generic (PLEG): container finished" podID="29e4895c-1440-496e-a3f8-06635f103456" containerID="40f3924e148377f827f93497c59e0add28dc20bb1fc85ce545164fb23eb032a4" exitCode=0 Oct 06 10:15:02 crc kubenswrapper[4824]: I1006 10:15:02.097588 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29329095-vgrt8" event={"ID":"29e4895c-1440-496e-a3f8-06635f103456","Type":"ContainerDied","Data":"40f3924e148377f827f93497c59e0add28dc20bb1fc85ce545164fb23eb032a4"} Oct 06 10:15:02 crc kubenswrapper[4824]: I1006 10:15:02.565032 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/6cdeba93-f1c9-487d-b776-5c2f1bb353d9-etc-swift\") pod \"swift-storage-0\" (UID: \"6cdeba93-f1c9-487d-b776-5c2f1bb353d9\") " pod="openstack/swift-storage-0" Oct 06 10:15:02 crc kubenswrapper[4824]: I1006 10:15:02.577082 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/6cdeba93-f1c9-487d-b776-5c2f1bb353d9-etc-swift\") pod \"swift-storage-0\" (UID: \"6cdeba93-f1c9-487d-b776-5c2f1bb353d9\") " pod="openstack/swift-storage-0" Oct 06 10:15:02 crc kubenswrapper[4824]: I1006 10:15:02.727302 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Oct 06 10:15:02 crc kubenswrapper[4824]: I1006 10:15:02.984636 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-faf1-account-create-85qr7"] Oct 06 10:15:02 crc kubenswrapper[4824]: E1006 10:15:02.985436 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b66b5d9-1c7c-4f7d-aa8d-cba5b9f6c2f2" containerName="mariadb-account-create" Oct 06 10:15:02 crc kubenswrapper[4824]: I1006 10:15:02.985453 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b66b5d9-1c7c-4f7d-aa8d-cba5b9f6c2f2" containerName="mariadb-account-create" Oct 06 10:15:02 crc kubenswrapper[4824]: E1006 10:15:02.985485 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="256cff80-df3a-481d-a994-05b49d8996c1" containerName="mariadb-database-create" Oct 06 10:15:02 crc kubenswrapper[4824]: I1006 10:15:02.985493 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="256cff80-df3a-481d-a994-05b49d8996c1" containerName="mariadb-database-create" Oct 06 10:15:02 crc kubenswrapper[4824]: I1006 10:15:02.985664 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="7b66b5d9-1c7c-4f7d-aa8d-cba5b9f6c2f2" containerName="mariadb-account-create" Oct 06 10:15:02 crc kubenswrapper[4824]: I1006 10:15:02.985677 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="256cff80-df3a-481d-a994-05b49d8996c1" containerName="mariadb-database-create" Oct 06 10:15:02 crc kubenswrapper[4824]: I1006 10:15:02.986768 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-faf1-account-create-85qr7" Oct 06 10:15:02 crc kubenswrapper[4824]: I1006 10:15:02.996510 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-faf1-account-create-85qr7"] Oct 06 10:15:02 crc kubenswrapper[4824]: I1006 10:15:02.996852 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Oct 06 10:15:03 crc kubenswrapper[4824]: I1006 10:15:03.074523 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-snqpv\" (UniqueName: \"kubernetes.io/projected/c9571f84-f5be-4f46-b939-2ad4d5a42daf-kube-api-access-snqpv\") pod \"glance-faf1-account-create-85qr7\" (UID: \"c9571f84-f5be-4f46-b939-2ad4d5a42daf\") " pod="openstack/glance-faf1-account-create-85qr7" Oct 06 10:15:03 crc kubenswrapper[4824]: I1006 10:15:03.176543 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-snqpv\" (UniqueName: \"kubernetes.io/projected/c9571f84-f5be-4f46-b939-2ad4d5a42daf-kube-api-access-snqpv\") pod \"glance-faf1-account-create-85qr7\" (UID: \"c9571f84-f5be-4f46-b939-2ad4d5a42daf\") " pod="openstack/glance-faf1-account-create-85qr7" Oct 06 10:15:03 crc kubenswrapper[4824]: I1006 10:15:03.199273 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-snqpv\" (UniqueName: \"kubernetes.io/projected/c9571f84-f5be-4f46-b939-2ad4d5a42daf-kube-api-access-snqpv\") pod \"glance-faf1-account-create-85qr7\" (UID: \"c9571f84-f5be-4f46-b939-2ad4d5a42daf\") " pod="openstack/glance-faf1-account-create-85qr7" Oct 06 10:15:03 crc kubenswrapper[4824]: I1006 10:15:03.328754 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-faf1-account-create-85qr7" Oct 06 10:15:03 crc kubenswrapper[4824]: I1006 10:15:03.416816 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Oct 06 10:15:03 crc kubenswrapper[4824]: W1006 10:15:03.424815 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6cdeba93_f1c9_487d_b776_5c2f1bb353d9.slice/crio-3d7f3cabb359fa18f12c237d0b07b0566d9abd71ab828e1c145608424b769aa0 WatchSource:0}: Error finding container 3d7f3cabb359fa18f12c237d0b07b0566d9abd71ab828e1c145608424b769aa0: Status 404 returned error can't find the container with id 3d7f3cabb359fa18f12c237d0b07b0566d9abd71ab828e1c145608424b769aa0 Oct 06 10:15:03 crc kubenswrapper[4824]: I1006 10:15:03.432458 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29329095-vgrt8" Oct 06 10:15:03 crc kubenswrapper[4824]: I1006 10:15:03.584586 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/29e4895c-1440-496e-a3f8-06635f103456-config-volume\") pod \"29e4895c-1440-496e-a3f8-06635f103456\" (UID: \"29e4895c-1440-496e-a3f8-06635f103456\") " Oct 06 10:15:03 crc kubenswrapper[4824]: I1006 10:15:03.585183 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/29e4895c-1440-496e-a3f8-06635f103456-secret-volume\") pod \"29e4895c-1440-496e-a3f8-06635f103456\" (UID: \"29e4895c-1440-496e-a3f8-06635f103456\") " Oct 06 10:15:03 crc kubenswrapper[4824]: I1006 10:15:03.585481 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-98mbl\" (UniqueName: \"kubernetes.io/projected/29e4895c-1440-496e-a3f8-06635f103456-kube-api-access-98mbl\") pod \"29e4895c-1440-496e-a3f8-06635f103456\" (UID: \"29e4895c-1440-496e-a3f8-06635f103456\") " Oct 06 10:15:03 crc kubenswrapper[4824]: I1006 10:15:03.585597 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/29e4895c-1440-496e-a3f8-06635f103456-config-volume" (OuterVolumeSpecName: "config-volume") pod "29e4895c-1440-496e-a3f8-06635f103456" (UID: "29e4895c-1440-496e-a3f8-06635f103456"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:15:03 crc kubenswrapper[4824]: I1006 10:15:03.586555 4824 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/29e4895c-1440-496e-a3f8-06635f103456-config-volume\") on node \"crc\" DevicePath \"\"" Oct 06 10:15:03 crc kubenswrapper[4824]: I1006 10:15:03.590764 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/29e4895c-1440-496e-a3f8-06635f103456-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "29e4895c-1440-496e-a3f8-06635f103456" (UID: "29e4895c-1440-496e-a3f8-06635f103456"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:15:03 crc kubenswrapper[4824]: I1006 10:15:03.591151 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/29e4895c-1440-496e-a3f8-06635f103456-kube-api-access-98mbl" (OuterVolumeSpecName: "kube-api-access-98mbl") pod "29e4895c-1440-496e-a3f8-06635f103456" (UID: "29e4895c-1440-496e-a3f8-06635f103456"). InnerVolumeSpecName "kube-api-access-98mbl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:15:03 crc kubenswrapper[4824]: I1006 10:15:03.688743 4824 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/29e4895c-1440-496e-a3f8-06635f103456-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 06 10:15:03 crc kubenswrapper[4824]: I1006 10:15:03.688847 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-98mbl\" (UniqueName: \"kubernetes.io/projected/29e4895c-1440-496e-a3f8-06635f103456-kube-api-access-98mbl\") on node \"crc\" DevicePath \"\"" Oct 06 10:15:03 crc kubenswrapper[4824]: I1006 10:15:03.798505 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-faf1-account-create-85qr7"] Oct 06 10:15:04 crc kubenswrapper[4824]: I1006 10:15:04.120638 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29329095-vgrt8" event={"ID":"29e4895c-1440-496e-a3f8-06635f103456","Type":"ContainerDied","Data":"e62d6f756caeee22fb192b0e1deac1c9bc9441bcd3b5fe46437456b9a143f262"} Oct 06 10:15:04 crc kubenswrapper[4824]: I1006 10:15:04.120934 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e62d6f756caeee22fb192b0e1deac1c9bc9441bcd3b5fe46437456b9a143f262" Oct 06 10:15:04 crc kubenswrapper[4824]: I1006 10:15:04.121075 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29329095-vgrt8" Oct 06 10:15:04 crc kubenswrapper[4824]: I1006 10:15:04.123549 4824 generic.go:334] "Generic (PLEG): container finished" podID="c9571f84-f5be-4f46-b939-2ad4d5a42daf" containerID="10dd696c355b056274df6482c88b5dbbf7520a4852f8252ed2269ed130337ac5" exitCode=0 Oct 06 10:15:04 crc kubenswrapper[4824]: I1006 10:15:04.123636 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-faf1-account-create-85qr7" event={"ID":"c9571f84-f5be-4f46-b939-2ad4d5a42daf","Type":"ContainerDied","Data":"10dd696c355b056274df6482c88b5dbbf7520a4852f8252ed2269ed130337ac5"} Oct 06 10:15:04 crc kubenswrapper[4824]: I1006 10:15:04.123697 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-faf1-account-create-85qr7" event={"ID":"c9571f84-f5be-4f46-b939-2ad4d5a42daf","Type":"ContainerStarted","Data":"b7f0a2d2c9024b8535bda218c60ececb81c4277fd05294962b11728783e8458d"} Oct 06 10:15:04 crc kubenswrapper[4824]: I1006 10:15:04.125257 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"6cdeba93-f1c9-487d-b776-5c2f1bb353d9","Type":"ContainerStarted","Data":"3d7f3cabb359fa18f12c237d0b07b0566d9abd71ab828e1c145608424b769aa0"} Oct 06 10:15:04 crc kubenswrapper[4824]: I1006 10:15:04.575685 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-bbe4-account-create-ghj2p"] Oct 06 10:15:04 crc kubenswrapper[4824]: E1006 10:15:04.576281 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="29e4895c-1440-496e-a3f8-06635f103456" containerName="collect-profiles" Oct 06 10:15:04 crc kubenswrapper[4824]: I1006 10:15:04.576301 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="29e4895c-1440-496e-a3f8-06635f103456" containerName="collect-profiles" Oct 06 10:15:04 crc kubenswrapper[4824]: I1006 10:15:04.576532 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="29e4895c-1440-496e-a3f8-06635f103456" containerName="collect-profiles" Oct 06 10:15:04 crc kubenswrapper[4824]: I1006 10:15:04.577949 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-bbe4-account-create-ghj2p" Oct 06 10:15:04 crc kubenswrapper[4824]: I1006 10:15:04.585657 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Oct 06 10:15:04 crc kubenswrapper[4824]: I1006 10:15:04.595387 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-bbe4-account-create-ghj2p"] Oct 06 10:15:04 crc kubenswrapper[4824]: I1006 10:15:04.706407 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-d68d-account-create-47bqf"] Oct 06 10:15:04 crc kubenswrapper[4824]: I1006 10:15:04.707794 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-d68d-account-create-47bqf" Oct 06 10:15:04 crc kubenswrapper[4824]: I1006 10:15:04.708589 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qhrlh\" (UniqueName: \"kubernetes.io/projected/4c9b54e8-db5a-4837-ab99-30f509b9d2fb-kube-api-access-qhrlh\") pod \"cinder-bbe4-account-create-ghj2p\" (UID: \"4c9b54e8-db5a-4837-ab99-30f509b9d2fb\") " pod="openstack/cinder-bbe4-account-create-ghj2p" Oct 06 10:15:04 crc kubenswrapper[4824]: I1006 10:15:04.717707 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Oct 06 10:15:04 crc kubenswrapper[4824]: I1006 10:15:04.721687 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-d68d-account-create-47bqf"] Oct 06 10:15:04 crc kubenswrapper[4824]: I1006 10:15:04.817258 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qhrlh\" (UniqueName: \"kubernetes.io/projected/4c9b54e8-db5a-4837-ab99-30f509b9d2fb-kube-api-access-qhrlh\") pod \"cinder-bbe4-account-create-ghj2p\" (UID: \"4c9b54e8-db5a-4837-ab99-30f509b9d2fb\") " pod="openstack/cinder-bbe4-account-create-ghj2p" Oct 06 10:15:04 crc kubenswrapper[4824]: I1006 10:15:04.817355 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nf67d\" (UniqueName: \"kubernetes.io/projected/258a753c-080b-44aa-9636-bfb6f7195b23-kube-api-access-nf67d\") pod \"barbican-d68d-account-create-47bqf\" (UID: \"258a753c-080b-44aa-9636-bfb6f7195b23\") " pod="openstack/barbican-d68d-account-create-47bqf" Oct 06 10:15:04 crc kubenswrapper[4824]: I1006 10:15:04.839864 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qhrlh\" (UniqueName: \"kubernetes.io/projected/4c9b54e8-db5a-4837-ab99-30f509b9d2fb-kube-api-access-qhrlh\") pod \"cinder-bbe4-account-create-ghj2p\" (UID: \"4c9b54e8-db5a-4837-ab99-30f509b9d2fb\") " pod="openstack/cinder-bbe4-account-create-ghj2p" Oct 06 10:15:04 crc kubenswrapper[4824]: I1006 10:15:04.904884 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-bbe4-account-create-ghj2p" Oct 06 10:15:04 crc kubenswrapper[4824]: I1006 10:15:04.920027 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nf67d\" (UniqueName: \"kubernetes.io/projected/258a753c-080b-44aa-9636-bfb6f7195b23-kube-api-access-nf67d\") pod \"barbican-d68d-account-create-47bqf\" (UID: \"258a753c-080b-44aa-9636-bfb6f7195b23\") " pod="openstack/barbican-d68d-account-create-47bqf" Oct 06 10:15:04 crc kubenswrapper[4824]: I1006 10:15:04.948872 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nf67d\" (UniqueName: \"kubernetes.io/projected/258a753c-080b-44aa-9636-bfb6f7195b23-kube-api-access-nf67d\") pod \"barbican-d68d-account-create-47bqf\" (UID: \"258a753c-080b-44aa-9636-bfb6f7195b23\") " pod="openstack/barbican-d68d-account-create-47bqf" Oct 06 10:15:05 crc kubenswrapper[4824]: I1006 10:15:05.022051 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-8782-account-create-86hgc"] Oct 06 10:15:05 crc kubenswrapper[4824]: I1006 10:15:05.028676 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-8782-account-create-86hgc" Oct 06 10:15:05 crc kubenswrapper[4824]: I1006 10:15:05.034962 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Oct 06 10:15:05 crc kubenswrapper[4824]: I1006 10:15:05.035812 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-d68d-account-create-47bqf" Oct 06 10:15:05 crc kubenswrapper[4824]: I1006 10:15:05.037654 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-8782-account-create-86hgc"] Oct 06 10:15:05 crc kubenswrapper[4824]: I1006 10:15:05.129944 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w6z85\" (UniqueName: \"kubernetes.io/projected/34475172-27f5-47b0-824e-71eff9ed3570-kube-api-access-w6z85\") pod \"neutron-8782-account-create-86hgc\" (UID: \"34475172-27f5-47b0-824e-71eff9ed3570\") " pod="openstack/neutron-8782-account-create-86hgc" Oct 06 10:15:05 crc kubenswrapper[4824]: I1006 10:15:05.139616 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"6cdeba93-f1c9-487d-b776-5c2f1bb353d9","Type":"ContainerStarted","Data":"d33f35dc5d024e548c80d883b05069dd057c8c956d59bf4e1d400419f94e2202"} Oct 06 10:15:05 crc kubenswrapper[4824]: I1006 10:15:05.139692 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"6cdeba93-f1c9-487d-b776-5c2f1bb353d9","Type":"ContainerStarted","Data":"6abb9ffbacf734361f6f7ed31c18771ddce0315f9e1139e336b593e2f43df9b8"} Oct 06 10:15:05 crc kubenswrapper[4824]: I1006 10:15:05.233587 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w6z85\" (UniqueName: \"kubernetes.io/projected/34475172-27f5-47b0-824e-71eff9ed3570-kube-api-access-w6z85\") pod \"neutron-8782-account-create-86hgc\" (UID: \"34475172-27f5-47b0-824e-71eff9ed3570\") " pod="openstack/neutron-8782-account-create-86hgc" Oct 06 10:15:05 crc kubenswrapper[4824]: I1006 10:15:05.244798 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-bbe4-account-create-ghj2p"] Oct 06 10:15:05 crc kubenswrapper[4824]: I1006 10:15:05.258318 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w6z85\" (UniqueName: \"kubernetes.io/projected/34475172-27f5-47b0-824e-71eff9ed3570-kube-api-access-w6z85\") pod \"neutron-8782-account-create-86hgc\" (UID: \"34475172-27f5-47b0-824e-71eff9ed3570\") " pod="openstack/neutron-8782-account-create-86hgc" Oct 06 10:15:05 crc kubenswrapper[4824]: W1006 10:15:05.267483 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4c9b54e8_db5a_4837_ab99_30f509b9d2fb.slice/crio-a60231bbff224a9acfc6b793aa254bdb27f0d97e32d9a7f1c56f0cf332aeefb9 WatchSource:0}: Error finding container a60231bbff224a9acfc6b793aa254bdb27f0d97e32d9a7f1c56f0cf332aeefb9: Status 404 returned error can't find the container with id a60231bbff224a9acfc6b793aa254bdb27f0d97e32d9a7f1c56f0cf332aeefb9 Oct 06 10:15:05 crc kubenswrapper[4824]: I1006 10:15:05.367507 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-8782-account-create-86hgc" Oct 06 10:15:05 crc kubenswrapper[4824]: I1006 10:15:05.461957 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-faf1-account-create-85qr7" Oct 06 10:15:05 crc kubenswrapper[4824]: I1006 10:15:05.543503 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-snqpv\" (UniqueName: \"kubernetes.io/projected/c9571f84-f5be-4f46-b939-2ad4d5a42daf-kube-api-access-snqpv\") pod \"c9571f84-f5be-4f46-b939-2ad4d5a42daf\" (UID: \"c9571f84-f5be-4f46-b939-2ad4d5a42daf\") " Oct 06 10:15:05 crc kubenswrapper[4824]: I1006 10:15:05.549556 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c9571f84-f5be-4f46-b939-2ad4d5a42daf-kube-api-access-snqpv" (OuterVolumeSpecName: "kube-api-access-snqpv") pod "c9571f84-f5be-4f46-b939-2ad4d5a42daf" (UID: "c9571f84-f5be-4f46-b939-2ad4d5a42daf"). InnerVolumeSpecName "kube-api-access-snqpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:15:05 crc kubenswrapper[4824]: I1006 10:15:05.620597 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-d68d-account-create-47bqf"] Oct 06 10:15:05 crc kubenswrapper[4824]: W1006 10:15:05.645079 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod258a753c_080b_44aa_9636_bfb6f7195b23.slice/crio-f1b480c60da4534f384a7cb7b3afd3e9838fdc0f34d76e031f2e512c028e879b WatchSource:0}: Error finding container f1b480c60da4534f384a7cb7b3afd3e9838fdc0f34d76e031f2e512c028e879b: Status 404 returned error can't find the container with id f1b480c60da4534f384a7cb7b3afd3e9838fdc0f34d76e031f2e512c028e879b Oct 06 10:15:05 crc kubenswrapper[4824]: I1006 10:15:05.646329 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-snqpv\" (UniqueName: \"kubernetes.io/projected/c9571f84-f5be-4f46-b939-2ad4d5a42daf-kube-api-access-snqpv\") on node \"crc\" DevicePath \"\"" Oct 06 10:15:05 crc kubenswrapper[4824]: I1006 10:15:05.662390 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-8782-account-create-86hgc"] Oct 06 10:15:06 crc kubenswrapper[4824]: I1006 10:15:06.162384 4824 generic.go:334] "Generic (PLEG): container finished" podID="34475172-27f5-47b0-824e-71eff9ed3570" containerID="a423d8ebda96fde12328f4eb430b9d6d71aa4c43d91b811bbab42c96f2fb8468" exitCode=0 Oct 06 10:15:06 crc kubenswrapper[4824]: I1006 10:15:06.162941 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-8782-account-create-86hgc" event={"ID":"34475172-27f5-47b0-824e-71eff9ed3570","Type":"ContainerDied","Data":"a423d8ebda96fde12328f4eb430b9d6d71aa4c43d91b811bbab42c96f2fb8468"} Oct 06 10:15:06 crc kubenswrapper[4824]: I1006 10:15:06.163012 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-8782-account-create-86hgc" event={"ID":"34475172-27f5-47b0-824e-71eff9ed3570","Type":"ContainerStarted","Data":"123cfabb41a880b073e97c31737c1ffdc14dccd2362d799c488a9705373b1730"} Oct 06 10:15:06 crc kubenswrapper[4824]: I1006 10:15:06.167228 4824 generic.go:334] "Generic (PLEG): container finished" podID="4c9b54e8-db5a-4837-ab99-30f509b9d2fb" containerID="f2bd5fad013050b478969598e1456a7593913eafa49996689b79b535d76a889d" exitCode=0 Oct 06 10:15:06 crc kubenswrapper[4824]: I1006 10:15:06.167333 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-bbe4-account-create-ghj2p" event={"ID":"4c9b54e8-db5a-4837-ab99-30f509b9d2fb","Type":"ContainerDied","Data":"f2bd5fad013050b478969598e1456a7593913eafa49996689b79b535d76a889d"} Oct 06 10:15:06 crc kubenswrapper[4824]: I1006 10:15:06.167384 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-bbe4-account-create-ghj2p" event={"ID":"4c9b54e8-db5a-4837-ab99-30f509b9d2fb","Type":"ContainerStarted","Data":"a60231bbff224a9acfc6b793aa254bdb27f0d97e32d9a7f1c56f0cf332aeefb9"} Oct 06 10:15:06 crc kubenswrapper[4824]: I1006 10:15:06.172311 4824 generic.go:334] "Generic (PLEG): container finished" podID="258a753c-080b-44aa-9636-bfb6f7195b23" containerID="78b87860dc45f07561d049b019a8b61447f0972f95e18d7d73e1fb8a6b4d874f" exitCode=0 Oct 06 10:15:06 crc kubenswrapper[4824]: I1006 10:15:06.172405 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-d68d-account-create-47bqf" event={"ID":"258a753c-080b-44aa-9636-bfb6f7195b23","Type":"ContainerDied","Data":"78b87860dc45f07561d049b019a8b61447f0972f95e18d7d73e1fb8a6b4d874f"} Oct 06 10:15:06 crc kubenswrapper[4824]: I1006 10:15:06.172435 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-d68d-account-create-47bqf" event={"ID":"258a753c-080b-44aa-9636-bfb6f7195b23","Type":"ContainerStarted","Data":"f1b480c60da4534f384a7cb7b3afd3e9838fdc0f34d76e031f2e512c028e879b"} Oct 06 10:15:06 crc kubenswrapper[4824]: I1006 10:15:06.175429 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-faf1-account-create-85qr7" Oct 06 10:15:06 crc kubenswrapper[4824]: I1006 10:15:06.175406 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-faf1-account-create-85qr7" event={"ID":"c9571f84-f5be-4f46-b939-2ad4d5a42daf","Type":"ContainerDied","Data":"b7f0a2d2c9024b8535bda218c60ececb81c4277fd05294962b11728783e8458d"} Oct 06 10:15:06 crc kubenswrapper[4824]: I1006 10:15:06.175710 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b7f0a2d2c9024b8535bda218c60ececb81c4277fd05294962b11728783e8458d" Oct 06 10:15:06 crc kubenswrapper[4824]: I1006 10:15:06.186922 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"6cdeba93-f1c9-487d-b776-5c2f1bb353d9","Type":"ContainerStarted","Data":"7bc487687c01b9d583f02aec9b0e28a4837de5a88c2c5d368d2199609c233f88"} Oct 06 10:15:06 crc kubenswrapper[4824]: I1006 10:15:06.189265 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"6cdeba93-f1c9-487d-b776-5c2f1bb353d9","Type":"ContainerStarted","Data":"7948fcb77a33a68cecd5cf07a7c2480b8e2d326b34a77924372d27d3d1dbc487"} Oct 06 10:15:07 crc kubenswrapper[4824]: I1006 10:15:07.336568 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-b964-account-create-h2khv"] Oct 06 10:15:07 crc kubenswrapper[4824]: E1006 10:15:07.337510 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9571f84-f5be-4f46-b939-2ad4d5a42daf" containerName="mariadb-account-create" Oct 06 10:15:07 crc kubenswrapper[4824]: I1006 10:15:07.337526 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9571f84-f5be-4f46-b939-2ad4d5a42daf" containerName="mariadb-account-create" Oct 06 10:15:07 crc kubenswrapper[4824]: I1006 10:15:07.337747 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="c9571f84-f5be-4f46-b939-2ad4d5a42daf" containerName="mariadb-account-create" Oct 06 10:15:07 crc kubenswrapper[4824]: I1006 10:15:07.341201 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-b964-account-create-h2khv" Oct 06 10:15:07 crc kubenswrapper[4824]: I1006 10:15:07.344353 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Oct 06 10:15:07 crc kubenswrapper[4824]: I1006 10:15:07.345841 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-b964-account-create-h2khv"] Oct 06 10:15:07 crc kubenswrapper[4824]: I1006 10:15:07.485306 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vbz8d\" (UniqueName: \"kubernetes.io/projected/c9fd14a4-e0f8-4905-88ad-77371475d9c2-kube-api-access-vbz8d\") pod \"keystone-b964-account-create-h2khv\" (UID: \"c9fd14a4-e0f8-4905-88ad-77371475d9c2\") " pod="openstack/keystone-b964-account-create-h2khv" Oct 06 10:15:07 crc kubenswrapper[4824]: I1006 10:15:07.586539 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vbz8d\" (UniqueName: \"kubernetes.io/projected/c9fd14a4-e0f8-4905-88ad-77371475d9c2-kube-api-access-vbz8d\") pod \"keystone-b964-account-create-h2khv\" (UID: \"c9fd14a4-e0f8-4905-88ad-77371475d9c2\") " pod="openstack/keystone-b964-account-create-h2khv" Oct 06 10:15:07 crc kubenswrapper[4824]: I1006 10:15:07.611183 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vbz8d\" (UniqueName: \"kubernetes.io/projected/c9fd14a4-e0f8-4905-88ad-77371475d9c2-kube-api-access-vbz8d\") pod \"keystone-b964-account-create-h2khv\" (UID: \"c9fd14a4-e0f8-4905-88ad-77371475d9c2\") " pod="openstack/keystone-b964-account-create-h2khv" Oct 06 10:15:07 crc kubenswrapper[4824]: I1006 10:15:07.671314 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-b964-account-create-h2khv" Oct 06 10:15:07 crc kubenswrapper[4824]: I1006 10:15:07.843262 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-d68d-account-create-47bqf" Oct 06 10:15:07 crc kubenswrapper[4824]: I1006 10:15:07.846530 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-8782-account-create-86hgc" Oct 06 10:15:07 crc kubenswrapper[4824]: I1006 10:15:07.970199 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-bbe4-account-create-ghj2p" Oct 06 10:15:08 crc kubenswrapper[4824]: I1006 10:15:08.038923 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nf67d\" (UniqueName: \"kubernetes.io/projected/258a753c-080b-44aa-9636-bfb6f7195b23-kube-api-access-nf67d\") pod \"258a753c-080b-44aa-9636-bfb6f7195b23\" (UID: \"258a753c-080b-44aa-9636-bfb6f7195b23\") " Oct 06 10:15:08 crc kubenswrapper[4824]: I1006 10:15:08.039254 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w6z85\" (UniqueName: \"kubernetes.io/projected/34475172-27f5-47b0-824e-71eff9ed3570-kube-api-access-w6z85\") pod \"34475172-27f5-47b0-824e-71eff9ed3570\" (UID: \"34475172-27f5-47b0-824e-71eff9ed3570\") " Oct 06 10:15:08 crc kubenswrapper[4824]: I1006 10:15:08.039314 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qhrlh\" (UniqueName: \"kubernetes.io/projected/4c9b54e8-db5a-4837-ab99-30f509b9d2fb-kube-api-access-qhrlh\") pod \"4c9b54e8-db5a-4837-ab99-30f509b9d2fb\" (UID: \"4c9b54e8-db5a-4837-ab99-30f509b9d2fb\") " Oct 06 10:15:08 crc kubenswrapper[4824]: I1006 10:15:08.045787 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4c9b54e8-db5a-4837-ab99-30f509b9d2fb-kube-api-access-qhrlh" (OuterVolumeSpecName: "kube-api-access-qhrlh") pod "4c9b54e8-db5a-4837-ab99-30f509b9d2fb" (UID: "4c9b54e8-db5a-4837-ab99-30f509b9d2fb"). InnerVolumeSpecName "kube-api-access-qhrlh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:15:08 crc kubenswrapper[4824]: I1006 10:15:08.045929 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/34475172-27f5-47b0-824e-71eff9ed3570-kube-api-access-w6z85" (OuterVolumeSpecName: "kube-api-access-w6z85") pod "34475172-27f5-47b0-824e-71eff9ed3570" (UID: "34475172-27f5-47b0-824e-71eff9ed3570"). InnerVolumeSpecName "kube-api-access-w6z85". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:15:08 crc kubenswrapper[4824]: I1006 10:15:08.046082 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/258a753c-080b-44aa-9636-bfb6f7195b23-kube-api-access-nf67d" (OuterVolumeSpecName: "kube-api-access-nf67d") pod "258a753c-080b-44aa-9636-bfb6f7195b23" (UID: "258a753c-080b-44aa-9636-bfb6f7195b23"). InnerVolumeSpecName "kube-api-access-nf67d". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:15:08 crc kubenswrapper[4824]: I1006 10:15:08.137540 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-59pkf"] Oct 06 10:15:08 crc kubenswrapper[4824]: E1006 10:15:08.138656 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34475172-27f5-47b0-824e-71eff9ed3570" containerName="mariadb-account-create" Oct 06 10:15:08 crc kubenswrapper[4824]: I1006 10:15:08.138681 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="34475172-27f5-47b0-824e-71eff9ed3570" containerName="mariadb-account-create" Oct 06 10:15:08 crc kubenswrapper[4824]: E1006 10:15:08.138706 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="258a753c-080b-44aa-9636-bfb6f7195b23" containerName="mariadb-account-create" Oct 06 10:15:08 crc kubenswrapper[4824]: I1006 10:15:08.138715 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="258a753c-080b-44aa-9636-bfb6f7195b23" containerName="mariadb-account-create" Oct 06 10:15:08 crc kubenswrapper[4824]: E1006 10:15:08.138739 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c9b54e8-db5a-4837-ab99-30f509b9d2fb" containerName="mariadb-account-create" Oct 06 10:15:08 crc kubenswrapper[4824]: I1006 10:15:08.138746 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c9b54e8-db5a-4837-ab99-30f509b9d2fb" containerName="mariadb-account-create" Oct 06 10:15:08 crc kubenswrapper[4824]: I1006 10:15:08.139049 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c9b54e8-db5a-4837-ab99-30f509b9d2fb" containerName="mariadb-account-create" Oct 06 10:15:08 crc kubenswrapper[4824]: I1006 10:15:08.139115 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="258a753c-080b-44aa-9636-bfb6f7195b23" containerName="mariadb-account-create" Oct 06 10:15:08 crc kubenswrapper[4824]: I1006 10:15:08.139150 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="34475172-27f5-47b0-824e-71eff9ed3570" containerName="mariadb-account-create" Oct 06 10:15:08 crc kubenswrapper[4824]: I1006 10:15:08.140110 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-59pkf" Oct 06 10:15:08 crc kubenswrapper[4824]: I1006 10:15:08.141426 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nf67d\" (UniqueName: \"kubernetes.io/projected/258a753c-080b-44aa-9636-bfb6f7195b23-kube-api-access-nf67d\") on node \"crc\" DevicePath \"\"" Oct 06 10:15:08 crc kubenswrapper[4824]: I1006 10:15:08.141468 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w6z85\" (UniqueName: \"kubernetes.io/projected/34475172-27f5-47b0-824e-71eff9ed3570-kube-api-access-w6z85\") on node \"crc\" DevicePath \"\"" Oct 06 10:15:08 crc kubenswrapper[4824]: I1006 10:15:08.141478 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qhrlh\" (UniqueName: \"kubernetes.io/projected/4c9b54e8-db5a-4837-ab99-30f509b9d2fb-kube-api-access-qhrlh\") on node \"crc\" DevicePath \"\"" Oct 06 10:15:08 crc kubenswrapper[4824]: I1006 10:15:08.143472 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Oct 06 10:15:08 crc kubenswrapper[4824]: I1006 10:15:08.143488 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-ctsvg" Oct 06 10:15:08 crc kubenswrapper[4824]: I1006 10:15:08.156768 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-59pkf"] Oct 06 10:15:08 crc kubenswrapper[4824]: I1006 10:15:08.209930 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-d68d-account-create-47bqf" event={"ID":"258a753c-080b-44aa-9636-bfb6f7195b23","Type":"ContainerDied","Data":"f1b480c60da4534f384a7cb7b3afd3e9838fdc0f34d76e031f2e512c028e879b"} Oct 06 10:15:08 crc kubenswrapper[4824]: I1006 10:15:08.210004 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f1b480c60da4534f384a7cb7b3afd3e9838fdc0f34d76e031f2e512c028e879b" Oct 06 10:15:08 crc kubenswrapper[4824]: I1006 10:15:08.210092 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-d68d-account-create-47bqf" Oct 06 10:15:08 crc kubenswrapper[4824]: I1006 10:15:08.218956 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"6cdeba93-f1c9-487d-b776-5c2f1bb353d9","Type":"ContainerStarted","Data":"58397fe0f76c14ced236015d827ded9d311b7ea9c1533e06daaa9d815d4a3aa1"} Oct 06 10:15:08 crc kubenswrapper[4824]: I1006 10:15:08.219055 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"6cdeba93-f1c9-487d-b776-5c2f1bb353d9","Type":"ContainerStarted","Data":"25c5d5af25ba6fd3db0ee100e008eebb08005e479ea9860c84e5b1a5b5acb376"} Oct 06 10:15:08 crc kubenswrapper[4824]: I1006 10:15:08.223831 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-8782-account-create-86hgc" event={"ID":"34475172-27f5-47b0-824e-71eff9ed3570","Type":"ContainerDied","Data":"123cfabb41a880b073e97c31737c1ffdc14dccd2362d799c488a9705373b1730"} Oct 06 10:15:08 crc kubenswrapper[4824]: I1006 10:15:08.223906 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="123cfabb41a880b073e97c31737c1ffdc14dccd2362d799c488a9705373b1730" Oct 06 10:15:08 crc kubenswrapper[4824]: I1006 10:15:08.224093 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-8782-account-create-86hgc" Oct 06 10:15:08 crc kubenswrapper[4824]: I1006 10:15:08.226685 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-bbe4-account-create-ghj2p" event={"ID":"4c9b54e8-db5a-4837-ab99-30f509b9d2fb","Type":"ContainerDied","Data":"a60231bbff224a9acfc6b793aa254bdb27f0d97e32d9a7f1c56f0cf332aeefb9"} Oct 06 10:15:08 crc kubenswrapper[4824]: I1006 10:15:08.226735 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a60231bbff224a9acfc6b793aa254bdb27f0d97e32d9a7f1c56f0cf332aeefb9" Oct 06 10:15:08 crc kubenswrapper[4824]: I1006 10:15:08.226807 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-bbe4-account-create-ghj2p" Oct 06 10:15:08 crc kubenswrapper[4824]: I1006 10:15:08.262999 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-b964-account-create-h2khv"] Oct 06 10:15:08 crc kubenswrapper[4824]: W1006 10:15:08.274469 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc9fd14a4_e0f8_4905_88ad_77371475d9c2.slice/crio-6b2435de687be9fbcad6acfbfd723b853343ac76798a8806e0dabba77fcca1e7 WatchSource:0}: Error finding container 6b2435de687be9fbcad6acfbfd723b853343ac76798a8806e0dabba77fcca1e7: Status 404 returned error can't find the container with id 6b2435de687be9fbcad6acfbfd723b853343ac76798a8806e0dabba77fcca1e7 Oct 06 10:15:08 crc kubenswrapper[4824]: I1006 10:15:08.345617 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/1bc5d3bf-5bde-4ce8-9b2a-50498e628b86-db-sync-config-data\") pod \"glance-db-sync-59pkf\" (UID: \"1bc5d3bf-5bde-4ce8-9b2a-50498e628b86\") " pod="openstack/glance-db-sync-59pkf" Oct 06 10:15:08 crc kubenswrapper[4824]: I1006 10:15:08.345716 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mwj5b\" (UniqueName: \"kubernetes.io/projected/1bc5d3bf-5bde-4ce8-9b2a-50498e628b86-kube-api-access-mwj5b\") pod \"glance-db-sync-59pkf\" (UID: \"1bc5d3bf-5bde-4ce8-9b2a-50498e628b86\") " pod="openstack/glance-db-sync-59pkf" Oct 06 10:15:08 crc kubenswrapper[4824]: I1006 10:15:08.345746 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1bc5d3bf-5bde-4ce8-9b2a-50498e628b86-config-data\") pod \"glance-db-sync-59pkf\" (UID: \"1bc5d3bf-5bde-4ce8-9b2a-50498e628b86\") " pod="openstack/glance-db-sync-59pkf" Oct 06 10:15:08 crc kubenswrapper[4824]: I1006 10:15:08.345774 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bc5d3bf-5bde-4ce8-9b2a-50498e628b86-combined-ca-bundle\") pod \"glance-db-sync-59pkf\" (UID: \"1bc5d3bf-5bde-4ce8-9b2a-50498e628b86\") " pod="openstack/glance-db-sync-59pkf" Oct 06 10:15:08 crc kubenswrapper[4824]: I1006 10:15:08.447430 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/1bc5d3bf-5bde-4ce8-9b2a-50498e628b86-db-sync-config-data\") pod \"glance-db-sync-59pkf\" (UID: \"1bc5d3bf-5bde-4ce8-9b2a-50498e628b86\") " pod="openstack/glance-db-sync-59pkf" Oct 06 10:15:08 crc kubenswrapper[4824]: I1006 10:15:08.447543 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mwj5b\" (UniqueName: \"kubernetes.io/projected/1bc5d3bf-5bde-4ce8-9b2a-50498e628b86-kube-api-access-mwj5b\") pod \"glance-db-sync-59pkf\" (UID: \"1bc5d3bf-5bde-4ce8-9b2a-50498e628b86\") " pod="openstack/glance-db-sync-59pkf" Oct 06 10:15:08 crc kubenswrapper[4824]: I1006 10:15:08.447585 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1bc5d3bf-5bde-4ce8-9b2a-50498e628b86-config-data\") pod \"glance-db-sync-59pkf\" (UID: \"1bc5d3bf-5bde-4ce8-9b2a-50498e628b86\") " pod="openstack/glance-db-sync-59pkf" Oct 06 10:15:08 crc kubenswrapper[4824]: I1006 10:15:08.447611 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bc5d3bf-5bde-4ce8-9b2a-50498e628b86-combined-ca-bundle\") pod \"glance-db-sync-59pkf\" (UID: \"1bc5d3bf-5bde-4ce8-9b2a-50498e628b86\") " pod="openstack/glance-db-sync-59pkf" Oct 06 10:15:08 crc kubenswrapper[4824]: I1006 10:15:08.456850 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bc5d3bf-5bde-4ce8-9b2a-50498e628b86-combined-ca-bundle\") pod \"glance-db-sync-59pkf\" (UID: \"1bc5d3bf-5bde-4ce8-9b2a-50498e628b86\") " pod="openstack/glance-db-sync-59pkf" Oct 06 10:15:08 crc kubenswrapper[4824]: I1006 10:15:08.461404 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/1bc5d3bf-5bde-4ce8-9b2a-50498e628b86-db-sync-config-data\") pod \"glance-db-sync-59pkf\" (UID: \"1bc5d3bf-5bde-4ce8-9b2a-50498e628b86\") " pod="openstack/glance-db-sync-59pkf" Oct 06 10:15:08 crc kubenswrapper[4824]: I1006 10:15:08.461913 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1bc5d3bf-5bde-4ce8-9b2a-50498e628b86-config-data\") pod \"glance-db-sync-59pkf\" (UID: \"1bc5d3bf-5bde-4ce8-9b2a-50498e628b86\") " pod="openstack/glance-db-sync-59pkf" Oct 06 10:15:08 crc kubenswrapper[4824]: I1006 10:15:08.482250 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mwj5b\" (UniqueName: \"kubernetes.io/projected/1bc5d3bf-5bde-4ce8-9b2a-50498e628b86-kube-api-access-mwj5b\") pod \"glance-db-sync-59pkf\" (UID: \"1bc5d3bf-5bde-4ce8-9b2a-50498e628b86\") " pod="openstack/glance-db-sync-59pkf" Oct 06 10:15:08 crc kubenswrapper[4824]: I1006 10:15:08.758334 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-59pkf" Oct 06 10:15:09 crc kubenswrapper[4824]: I1006 10:15:09.164271 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-59pkf"] Oct 06 10:15:09 crc kubenswrapper[4824]: W1006 10:15:09.178812 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1bc5d3bf_5bde_4ce8_9b2a_50498e628b86.slice/crio-8e6c6202927867d3cb6b8341a56e868f2c2f7f8d022d17414b123d2e4afa7828 WatchSource:0}: Error finding container 8e6c6202927867d3cb6b8341a56e868f2c2f7f8d022d17414b123d2e4afa7828: Status 404 returned error can't find the container with id 8e6c6202927867d3cb6b8341a56e868f2c2f7f8d022d17414b123d2e4afa7828 Oct 06 10:15:09 crc kubenswrapper[4824]: I1006 10:15:09.243356 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-b964-account-create-h2khv" event={"ID":"c9fd14a4-e0f8-4905-88ad-77371475d9c2","Type":"ContainerStarted","Data":"e05cd16913d32642ed8363d7ffe7f9912b69414d4757d773e592dfb555d63428"} Oct 06 10:15:09 crc kubenswrapper[4824]: I1006 10:15:09.243829 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-b964-account-create-h2khv" event={"ID":"c9fd14a4-e0f8-4905-88ad-77371475d9c2","Type":"ContainerStarted","Data":"6b2435de687be9fbcad6acfbfd723b853343ac76798a8806e0dabba77fcca1e7"} Oct 06 10:15:09 crc kubenswrapper[4824]: I1006 10:15:09.246584 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-59pkf" event={"ID":"1bc5d3bf-5bde-4ce8-9b2a-50498e628b86","Type":"ContainerStarted","Data":"8e6c6202927867d3cb6b8341a56e868f2c2f7f8d022d17414b123d2e4afa7828"} Oct 06 10:15:09 crc kubenswrapper[4824]: I1006 10:15:09.252507 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"6cdeba93-f1c9-487d-b776-5c2f1bb353d9","Type":"ContainerStarted","Data":"607516ea24b5d246a4ddda2ab308815a39fc371d7a90422f6d32f4eeb0f183c9"} Oct 06 10:15:09 crc kubenswrapper[4824]: I1006 10:15:09.266913 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-b964-account-create-h2khv" podStartSLOduration=2.266879923 podStartE2EDuration="2.266879923s" podCreationTimestamp="2025-10-06 10:15:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:15:09.259182223 +0000 UTC m=+1018.623605084" watchObservedRunningTime="2025-10-06 10:15:09.266879923 +0000 UTC m=+1018.631302774" Oct 06 10:15:10 crc kubenswrapper[4824]: I1006 10:15:10.283333 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"6cdeba93-f1c9-487d-b776-5c2f1bb353d9","Type":"ContainerStarted","Data":"bd19f84462fbd0305f2f5ee290eedb9d85bc1e8f6f954b275d6201f60a2e732b"} Oct 06 10:15:10 crc kubenswrapper[4824]: I1006 10:15:10.285345 4824 generic.go:334] "Generic (PLEG): container finished" podID="c9fd14a4-e0f8-4905-88ad-77371475d9c2" containerID="e05cd16913d32642ed8363d7ffe7f9912b69414d4757d773e592dfb555d63428" exitCode=0 Oct 06 10:15:10 crc kubenswrapper[4824]: I1006 10:15:10.285414 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-b964-account-create-h2khv" event={"ID":"c9fd14a4-e0f8-4905-88ad-77371475d9c2","Type":"ContainerDied","Data":"e05cd16913d32642ed8363d7ffe7f9912b69414d4757d773e592dfb555d63428"} Oct 06 10:15:11 crc kubenswrapper[4824]: I1006 10:15:11.317423 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"6cdeba93-f1c9-487d-b776-5c2f1bb353d9","Type":"ContainerStarted","Data":"b48ddaec8d9d6c868497666284e9021c29a3f00d1d7cca576acc8bfd44aca742"} Oct 06 10:15:11 crc kubenswrapper[4824]: I1006 10:15:11.317825 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"6cdeba93-f1c9-487d-b776-5c2f1bb353d9","Type":"ContainerStarted","Data":"74da7f1e81e70f2cd6cabc25fd24e700659cfb0c9c4119eec70a499b9958a5e4"} Oct 06 10:15:11 crc kubenswrapper[4824]: I1006 10:15:11.694175 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-b964-account-create-h2khv" Oct 06 10:15:11 crc kubenswrapper[4824]: I1006 10:15:11.813689 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vbz8d\" (UniqueName: \"kubernetes.io/projected/c9fd14a4-e0f8-4905-88ad-77371475d9c2-kube-api-access-vbz8d\") pod \"c9fd14a4-e0f8-4905-88ad-77371475d9c2\" (UID: \"c9fd14a4-e0f8-4905-88ad-77371475d9c2\") " Oct 06 10:15:11 crc kubenswrapper[4824]: I1006 10:15:11.821024 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c9fd14a4-e0f8-4905-88ad-77371475d9c2-kube-api-access-vbz8d" (OuterVolumeSpecName: "kube-api-access-vbz8d") pod "c9fd14a4-e0f8-4905-88ad-77371475d9c2" (UID: "c9fd14a4-e0f8-4905-88ad-77371475d9c2"). InnerVolumeSpecName "kube-api-access-vbz8d". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:15:11 crc kubenswrapper[4824]: I1006 10:15:11.916953 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vbz8d\" (UniqueName: \"kubernetes.io/projected/c9fd14a4-e0f8-4905-88ad-77371475d9c2-kube-api-access-vbz8d\") on node \"crc\" DevicePath \"\"" Oct 06 10:15:12 crc kubenswrapper[4824]: I1006 10:15:12.339137 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"6cdeba93-f1c9-487d-b776-5c2f1bb353d9","Type":"ContainerStarted","Data":"59267a835ad45373d9d209464a3af9b099f25be6adf88853eabc43f2ad0b4476"} Oct 06 10:15:12 crc kubenswrapper[4824]: I1006 10:15:12.339186 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"6cdeba93-f1c9-487d-b776-5c2f1bb353d9","Type":"ContainerStarted","Data":"fb0472acdcf2dc79f38440e90957ac58984b7491088ede46183f5e3f99a07633"} Oct 06 10:15:12 crc kubenswrapper[4824]: I1006 10:15:12.339196 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"6cdeba93-f1c9-487d-b776-5c2f1bb353d9","Type":"ContainerStarted","Data":"a67c629e44f0a2d7239f95149952d4503667ffec936943ac987980e595434acb"} Oct 06 10:15:12 crc kubenswrapper[4824]: I1006 10:15:12.339208 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"6cdeba93-f1c9-487d-b776-5c2f1bb353d9","Type":"ContainerStarted","Data":"0fc7375291fa072c59f4509bc63e084f5134ebeb425df982465538557b6b99dd"} Oct 06 10:15:12 crc kubenswrapper[4824]: I1006 10:15:12.339218 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"6cdeba93-f1c9-487d-b776-5c2f1bb353d9","Type":"ContainerStarted","Data":"ee7857cb6d95aea13f0ac3d38bf189d97018fd23d8805994f2e065f32fed24ab"} Oct 06 10:15:12 crc kubenswrapper[4824]: I1006 10:15:12.340591 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-b964-account-create-h2khv" event={"ID":"c9fd14a4-e0f8-4905-88ad-77371475d9c2","Type":"ContainerDied","Data":"6b2435de687be9fbcad6acfbfd723b853343ac76798a8806e0dabba77fcca1e7"} Oct 06 10:15:12 crc kubenswrapper[4824]: I1006 10:15:12.340618 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6b2435de687be9fbcad6acfbfd723b853343ac76798a8806e0dabba77fcca1e7" Oct 06 10:15:12 crc kubenswrapper[4824]: I1006 10:15:12.340662 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-b964-account-create-h2khv" Oct 06 10:15:12 crc kubenswrapper[4824]: I1006 10:15:12.388742 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=35.992190967 podStartE2EDuration="43.388717024s" podCreationTimestamp="2025-10-06 10:14:29 +0000 UTC" firstStartedPulling="2025-10-06 10:15:03.428804825 +0000 UTC m=+1012.793227686" lastFinishedPulling="2025-10-06 10:15:10.825330882 +0000 UTC m=+1020.189753743" observedRunningTime="2025-10-06 10:15:12.372740629 +0000 UTC m=+1021.737163490" watchObservedRunningTime="2025-10-06 10:15:12.388717024 +0000 UTC m=+1021.753139895" Oct 06 10:15:12 crc kubenswrapper[4824]: I1006 10:15:12.730818 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-wtjf5"] Oct 06 10:15:12 crc kubenswrapper[4824]: E1006 10:15:12.731184 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9fd14a4-e0f8-4905-88ad-77371475d9c2" containerName="mariadb-account-create" Oct 06 10:15:12 crc kubenswrapper[4824]: I1006 10:15:12.731199 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9fd14a4-e0f8-4905-88ad-77371475d9c2" containerName="mariadb-account-create" Oct 06 10:15:12 crc kubenswrapper[4824]: I1006 10:15:12.731389 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="c9fd14a4-e0f8-4905-88ad-77371475d9c2" containerName="mariadb-account-create" Oct 06 10:15:12 crc kubenswrapper[4824]: I1006 10:15:12.732249 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77585f5f8c-wtjf5" Oct 06 10:15:12 crc kubenswrapper[4824]: I1006 10:15:12.738924 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Oct 06 10:15:12 crc kubenswrapper[4824]: I1006 10:15:12.749137 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-wtjf5"] Oct 06 10:15:12 crc kubenswrapper[4824]: I1006 10:15:12.836201 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/51ecd7bf-8b94-4c8c-af6e-a7e74e7f7c1c-dns-swift-storage-0\") pod \"dnsmasq-dns-77585f5f8c-wtjf5\" (UID: \"51ecd7bf-8b94-4c8c-af6e-a7e74e7f7c1c\") " pod="openstack/dnsmasq-dns-77585f5f8c-wtjf5" Oct 06 10:15:12 crc kubenswrapper[4824]: I1006 10:15:12.836271 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/51ecd7bf-8b94-4c8c-af6e-a7e74e7f7c1c-dns-svc\") pod \"dnsmasq-dns-77585f5f8c-wtjf5\" (UID: \"51ecd7bf-8b94-4c8c-af6e-a7e74e7f7c1c\") " pod="openstack/dnsmasq-dns-77585f5f8c-wtjf5" Oct 06 10:15:12 crc kubenswrapper[4824]: I1006 10:15:12.836340 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/51ecd7bf-8b94-4c8c-af6e-a7e74e7f7c1c-config\") pod \"dnsmasq-dns-77585f5f8c-wtjf5\" (UID: \"51ecd7bf-8b94-4c8c-af6e-a7e74e7f7c1c\") " pod="openstack/dnsmasq-dns-77585f5f8c-wtjf5" Oct 06 10:15:12 crc kubenswrapper[4824]: I1006 10:15:12.836525 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/51ecd7bf-8b94-4c8c-af6e-a7e74e7f7c1c-ovsdbserver-sb\") pod \"dnsmasq-dns-77585f5f8c-wtjf5\" (UID: \"51ecd7bf-8b94-4c8c-af6e-a7e74e7f7c1c\") " pod="openstack/dnsmasq-dns-77585f5f8c-wtjf5" Oct 06 10:15:12 crc kubenswrapper[4824]: I1006 10:15:12.837032 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dl6mz\" (UniqueName: \"kubernetes.io/projected/51ecd7bf-8b94-4c8c-af6e-a7e74e7f7c1c-kube-api-access-dl6mz\") pod \"dnsmasq-dns-77585f5f8c-wtjf5\" (UID: \"51ecd7bf-8b94-4c8c-af6e-a7e74e7f7c1c\") " pod="openstack/dnsmasq-dns-77585f5f8c-wtjf5" Oct 06 10:15:12 crc kubenswrapper[4824]: I1006 10:15:12.837228 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/51ecd7bf-8b94-4c8c-af6e-a7e74e7f7c1c-ovsdbserver-nb\") pod \"dnsmasq-dns-77585f5f8c-wtjf5\" (UID: \"51ecd7bf-8b94-4c8c-af6e-a7e74e7f7c1c\") " pod="openstack/dnsmasq-dns-77585f5f8c-wtjf5" Oct 06 10:15:12 crc kubenswrapper[4824]: I1006 10:15:12.920863 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-s7fhb"] Oct 06 10:15:12 crc kubenswrapper[4824]: I1006 10:15:12.928649 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-s7fhb" Oct 06 10:15:12 crc kubenswrapper[4824]: I1006 10:15:12.931840 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 06 10:15:12 crc kubenswrapper[4824]: I1006 10:15:12.932504 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 06 10:15:12 crc kubenswrapper[4824]: I1006 10:15:12.933113 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-lxhxz" Oct 06 10:15:12 crc kubenswrapper[4824]: I1006 10:15:12.933216 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 06 10:15:12 crc kubenswrapper[4824]: I1006 10:15:12.938182 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-s7fhb"] Oct 06 10:15:12 crc kubenswrapper[4824]: I1006 10:15:12.941153 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dl6mz\" (UniqueName: \"kubernetes.io/projected/51ecd7bf-8b94-4c8c-af6e-a7e74e7f7c1c-kube-api-access-dl6mz\") pod \"dnsmasq-dns-77585f5f8c-wtjf5\" (UID: \"51ecd7bf-8b94-4c8c-af6e-a7e74e7f7c1c\") " pod="openstack/dnsmasq-dns-77585f5f8c-wtjf5" Oct 06 10:15:12 crc kubenswrapper[4824]: I1006 10:15:12.941289 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/51ecd7bf-8b94-4c8c-af6e-a7e74e7f7c1c-ovsdbserver-nb\") pod \"dnsmasq-dns-77585f5f8c-wtjf5\" (UID: \"51ecd7bf-8b94-4c8c-af6e-a7e74e7f7c1c\") " pod="openstack/dnsmasq-dns-77585f5f8c-wtjf5" Oct 06 10:15:12 crc kubenswrapper[4824]: I1006 10:15:12.941340 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/51ecd7bf-8b94-4c8c-af6e-a7e74e7f7c1c-dns-swift-storage-0\") pod \"dnsmasq-dns-77585f5f8c-wtjf5\" (UID: \"51ecd7bf-8b94-4c8c-af6e-a7e74e7f7c1c\") " pod="openstack/dnsmasq-dns-77585f5f8c-wtjf5" Oct 06 10:15:12 crc kubenswrapper[4824]: I1006 10:15:12.941362 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/51ecd7bf-8b94-4c8c-af6e-a7e74e7f7c1c-dns-svc\") pod \"dnsmasq-dns-77585f5f8c-wtjf5\" (UID: \"51ecd7bf-8b94-4c8c-af6e-a7e74e7f7c1c\") " pod="openstack/dnsmasq-dns-77585f5f8c-wtjf5" Oct 06 10:15:12 crc kubenswrapper[4824]: I1006 10:15:12.941457 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/51ecd7bf-8b94-4c8c-af6e-a7e74e7f7c1c-config\") pod \"dnsmasq-dns-77585f5f8c-wtjf5\" (UID: \"51ecd7bf-8b94-4c8c-af6e-a7e74e7f7c1c\") " pod="openstack/dnsmasq-dns-77585f5f8c-wtjf5" Oct 06 10:15:12 crc kubenswrapper[4824]: I1006 10:15:12.941499 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/51ecd7bf-8b94-4c8c-af6e-a7e74e7f7c1c-ovsdbserver-sb\") pod \"dnsmasq-dns-77585f5f8c-wtjf5\" (UID: \"51ecd7bf-8b94-4c8c-af6e-a7e74e7f7c1c\") " pod="openstack/dnsmasq-dns-77585f5f8c-wtjf5" Oct 06 10:15:12 crc kubenswrapper[4824]: I1006 10:15:12.942922 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/51ecd7bf-8b94-4c8c-af6e-a7e74e7f7c1c-ovsdbserver-sb\") pod \"dnsmasq-dns-77585f5f8c-wtjf5\" (UID: \"51ecd7bf-8b94-4c8c-af6e-a7e74e7f7c1c\") " pod="openstack/dnsmasq-dns-77585f5f8c-wtjf5" Oct 06 10:15:12 crc kubenswrapper[4824]: I1006 10:15:12.943117 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/51ecd7bf-8b94-4c8c-af6e-a7e74e7f7c1c-config\") pod \"dnsmasq-dns-77585f5f8c-wtjf5\" (UID: \"51ecd7bf-8b94-4c8c-af6e-a7e74e7f7c1c\") " pod="openstack/dnsmasq-dns-77585f5f8c-wtjf5" Oct 06 10:15:12 crc kubenswrapper[4824]: I1006 10:15:12.943301 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/51ecd7bf-8b94-4c8c-af6e-a7e74e7f7c1c-dns-svc\") pod \"dnsmasq-dns-77585f5f8c-wtjf5\" (UID: \"51ecd7bf-8b94-4c8c-af6e-a7e74e7f7c1c\") " pod="openstack/dnsmasq-dns-77585f5f8c-wtjf5" Oct 06 10:15:12 crc kubenswrapper[4824]: I1006 10:15:12.943925 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/51ecd7bf-8b94-4c8c-af6e-a7e74e7f7c1c-dns-swift-storage-0\") pod \"dnsmasq-dns-77585f5f8c-wtjf5\" (UID: \"51ecd7bf-8b94-4c8c-af6e-a7e74e7f7c1c\") " pod="openstack/dnsmasq-dns-77585f5f8c-wtjf5" Oct 06 10:15:12 crc kubenswrapper[4824]: I1006 10:15:12.944015 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/51ecd7bf-8b94-4c8c-af6e-a7e74e7f7c1c-ovsdbserver-nb\") pod \"dnsmasq-dns-77585f5f8c-wtjf5\" (UID: \"51ecd7bf-8b94-4c8c-af6e-a7e74e7f7c1c\") " pod="openstack/dnsmasq-dns-77585f5f8c-wtjf5" Oct 06 10:15:12 crc kubenswrapper[4824]: I1006 10:15:12.971029 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dl6mz\" (UniqueName: \"kubernetes.io/projected/51ecd7bf-8b94-4c8c-af6e-a7e74e7f7c1c-kube-api-access-dl6mz\") pod \"dnsmasq-dns-77585f5f8c-wtjf5\" (UID: \"51ecd7bf-8b94-4c8c-af6e-a7e74e7f7c1c\") " pod="openstack/dnsmasq-dns-77585f5f8c-wtjf5" Oct 06 10:15:13 crc kubenswrapper[4824]: I1006 10:15:13.042825 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9v4c9\" (UniqueName: \"kubernetes.io/projected/e843b5e1-97a8-4a1c-8fa0-f3d8cf155ea0-kube-api-access-9v4c9\") pod \"keystone-db-sync-s7fhb\" (UID: \"e843b5e1-97a8-4a1c-8fa0-f3d8cf155ea0\") " pod="openstack/keystone-db-sync-s7fhb" Oct 06 10:15:13 crc kubenswrapper[4824]: I1006 10:15:13.043006 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e843b5e1-97a8-4a1c-8fa0-f3d8cf155ea0-config-data\") pod \"keystone-db-sync-s7fhb\" (UID: \"e843b5e1-97a8-4a1c-8fa0-f3d8cf155ea0\") " pod="openstack/keystone-db-sync-s7fhb" Oct 06 10:15:13 crc kubenswrapper[4824]: I1006 10:15:13.043085 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e843b5e1-97a8-4a1c-8fa0-f3d8cf155ea0-combined-ca-bundle\") pod \"keystone-db-sync-s7fhb\" (UID: \"e843b5e1-97a8-4a1c-8fa0-f3d8cf155ea0\") " pod="openstack/keystone-db-sync-s7fhb" Oct 06 10:15:13 crc kubenswrapper[4824]: I1006 10:15:13.071823 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77585f5f8c-wtjf5" Oct 06 10:15:13 crc kubenswrapper[4824]: I1006 10:15:13.144692 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9v4c9\" (UniqueName: \"kubernetes.io/projected/e843b5e1-97a8-4a1c-8fa0-f3d8cf155ea0-kube-api-access-9v4c9\") pod \"keystone-db-sync-s7fhb\" (UID: \"e843b5e1-97a8-4a1c-8fa0-f3d8cf155ea0\") " pod="openstack/keystone-db-sync-s7fhb" Oct 06 10:15:13 crc kubenswrapper[4824]: I1006 10:15:13.144788 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e843b5e1-97a8-4a1c-8fa0-f3d8cf155ea0-config-data\") pod \"keystone-db-sync-s7fhb\" (UID: \"e843b5e1-97a8-4a1c-8fa0-f3d8cf155ea0\") " pod="openstack/keystone-db-sync-s7fhb" Oct 06 10:15:13 crc kubenswrapper[4824]: I1006 10:15:13.144817 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e843b5e1-97a8-4a1c-8fa0-f3d8cf155ea0-combined-ca-bundle\") pod \"keystone-db-sync-s7fhb\" (UID: \"e843b5e1-97a8-4a1c-8fa0-f3d8cf155ea0\") " pod="openstack/keystone-db-sync-s7fhb" Oct 06 10:15:13 crc kubenswrapper[4824]: I1006 10:15:13.149955 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e843b5e1-97a8-4a1c-8fa0-f3d8cf155ea0-combined-ca-bundle\") pod \"keystone-db-sync-s7fhb\" (UID: \"e843b5e1-97a8-4a1c-8fa0-f3d8cf155ea0\") " pod="openstack/keystone-db-sync-s7fhb" Oct 06 10:15:13 crc kubenswrapper[4824]: I1006 10:15:13.150113 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e843b5e1-97a8-4a1c-8fa0-f3d8cf155ea0-config-data\") pod \"keystone-db-sync-s7fhb\" (UID: \"e843b5e1-97a8-4a1c-8fa0-f3d8cf155ea0\") " pod="openstack/keystone-db-sync-s7fhb" Oct 06 10:15:13 crc kubenswrapper[4824]: I1006 10:15:13.169652 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9v4c9\" (UniqueName: \"kubernetes.io/projected/e843b5e1-97a8-4a1c-8fa0-f3d8cf155ea0-kube-api-access-9v4c9\") pod \"keystone-db-sync-s7fhb\" (UID: \"e843b5e1-97a8-4a1c-8fa0-f3d8cf155ea0\") " pod="openstack/keystone-db-sync-s7fhb" Oct 06 10:15:13 crc kubenswrapper[4824]: I1006 10:15:13.256473 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-s7fhb" Oct 06 10:15:13 crc kubenswrapper[4824]: I1006 10:15:13.623972 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-wtjf5"] Oct 06 10:15:13 crc kubenswrapper[4824]: I1006 10:15:13.644585 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-s7fhb"] Oct 06 10:15:13 crc kubenswrapper[4824]: W1006 10:15:13.647551 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod51ecd7bf_8b94_4c8c_af6e_a7e74e7f7c1c.slice/crio-11d534cf0352c0c2d7dfea22ab9090584ee138420f40dc3e46ac9ddacd83f120 WatchSource:0}: Error finding container 11d534cf0352c0c2d7dfea22ab9090584ee138420f40dc3e46ac9ddacd83f120: Status 404 returned error can't find the container with id 11d534cf0352c0c2d7dfea22ab9090584ee138420f40dc3e46ac9ddacd83f120 Oct 06 10:15:13 crc kubenswrapper[4824]: W1006 10:15:13.651779 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode843b5e1_97a8_4a1c_8fa0_f3d8cf155ea0.slice/crio-c12f731634f31923dee3721868b11bf96850bb2fe531ee1485ecc73284d5f331 WatchSource:0}: Error finding container c12f731634f31923dee3721868b11bf96850bb2fe531ee1485ecc73284d5f331: Status 404 returned error can't find the container with id c12f731634f31923dee3721868b11bf96850bb2fe531ee1485ecc73284d5f331 Oct 06 10:15:13 crc kubenswrapper[4824]: I1006 10:15:13.915840 4824 patch_prober.go:28] interesting pod/machine-config-daemon-khgzw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 10:15:13 crc kubenswrapper[4824]: I1006 10:15:13.915917 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 10:15:14 crc kubenswrapper[4824]: I1006 10:15:14.390183 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-s7fhb" event={"ID":"e843b5e1-97a8-4a1c-8fa0-f3d8cf155ea0","Type":"ContainerStarted","Data":"c12f731634f31923dee3721868b11bf96850bb2fe531ee1485ecc73284d5f331"} Oct 06 10:15:14 crc kubenswrapper[4824]: I1006 10:15:14.393670 4824 generic.go:334] "Generic (PLEG): container finished" podID="51ecd7bf-8b94-4c8c-af6e-a7e74e7f7c1c" containerID="aff700ad3efac962ccd117169b4ca28a7581f213e1f027952ab79d16a6167ae0" exitCode=0 Oct 06 10:15:14 crc kubenswrapper[4824]: I1006 10:15:14.393729 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77585f5f8c-wtjf5" event={"ID":"51ecd7bf-8b94-4c8c-af6e-a7e74e7f7c1c","Type":"ContainerDied","Data":"aff700ad3efac962ccd117169b4ca28a7581f213e1f027952ab79d16a6167ae0"} Oct 06 10:15:14 crc kubenswrapper[4824]: I1006 10:15:14.393768 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77585f5f8c-wtjf5" event={"ID":"51ecd7bf-8b94-4c8c-af6e-a7e74e7f7c1c","Type":"ContainerStarted","Data":"11d534cf0352c0c2d7dfea22ab9090584ee138420f40dc3e46ac9ddacd83f120"} Oct 06 10:15:15 crc kubenswrapper[4824]: I1006 10:15:15.407330 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77585f5f8c-wtjf5" event={"ID":"51ecd7bf-8b94-4c8c-af6e-a7e74e7f7c1c","Type":"ContainerStarted","Data":"e17f227e1abc0d9b1729053104cdd67ed12db77ffc53e5658e7c5da09ebc3383"} Oct 06 10:15:15 crc kubenswrapper[4824]: I1006 10:15:15.407892 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-77585f5f8c-wtjf5" Oct 06 10:15:15 crc kubenswrapper[4824]: I1006 10:15:15.437139 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-77585f5f8c-wtjf5" podStartSLOduration=3.437117748 podStartE2EDuration="3.437117748s" podCreationTimestamp="2025-10-06 10:15:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:15:15.433244122 +0000 UTC m=+1024.797667013" watchObservedRunningTime="2025-10-06 10:15:15.437117748 +0000 UTC m=+1024.801540599" Oct 06 10:15:23 crc kubenswrapper[4824]: I1006 10:15:23.074173 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-77585f5f8c-wtjf5" Oct 06 10:15:23 crc kubenswrapper[4824]: I1006 10:15:23.132192 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-cwvfc"] Oct 06 10:15:23 crc kubenswrapper[4824]: I1006 10:15:23.132771 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-698758b865-cwvfc" podUID="b4ac2488-c99e-4380-8e82-e89207990c45" containerName="dnsmasq-dns" containerID="cri-o://50ee8359d95369ba4299b5af2a6db387bcddd6aee11c5ce974e82a168eb8a3a0" gracePeriod=10 Oct 06 10:15:24 crc kubenswrapper[4824]: I1006 10:15:24.495834 4824 generic.go:334] "Generic (PLEG): container finished" podID="b4ac2488-c99e-4380-8e82-e89207990c45" containerID="50ee8359d95369ba4299b5af2a6db387bcddd6aee11c5ce974e82a168eb8a3a0" exitCode=0 Oct 06 10:15:24 crc kubenswrapper[4824]: I1006 10:15:24.495949 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-cwvfc" event={"ID":"b4ac2488-c99e-4380-8e82-e89207990c45","Type":"ContainerDied","Data":"50ee8359d95369ba4299b5af2a6db387bcddd6aee11c5ce974e82a168eb8a3a0"} Oct 06 10:15:24 crc kubenswrapper[4824]: I1006 10:15:24.851638 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-698758b865-cwvfc" podUID="b4ac2488-c99e-4380-8e82-e89207990c45" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.115:5353: connect: connection refused" Oct 06 10:15:27 crc kubenswrapper[4824]: E1006 10:15:27.955738 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-keystone:current-podified" Oct 06 10:15:27 crc kubenswrapper[4824]: E1006 10:15:27.956579 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:keystone-db-sync,Image:quay.io/podified-antelope-centos9/openstack-keystone:current-podified,Command:[/bin/bash],Args:[-c keystone-manage db_sync],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/keystone/keystone.conf,SubPath:keystone.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9v4c9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42425,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42425,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-db-sync-s7fhb_openstack(e843b5e1-97a8-4a1c-8fa0-f3d8cf155ea0): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 06 10:15:27 crc kubenswrapper[4824]: E1006 10:15:27.957856 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"keystone-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/keystone-db-sync-s7fhb" podUID="e843b5e1-97a8-4a1c-8fa0-f3d8cf155ea0" Oct 06 10:15:28 crc kubenswrapper[4824]: I1006 10:15:28.311118 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-cwvfc" Oct 06 10:15:28 crc kubenswrapper[4824]: I1006 10:15:28.455035 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b4ac2488-c99e-4380-8e82-e89207990c45-dns-svc\") pod \"b4ac2488-c99e-4380-8e82-e89207990c45\" (UID: \"b4ac2488-c99e-4380-8e82-e89207990c45\") " Oct 06 10:15:28 crc kubenswrapper[4824]: I1006 10:15:28.455126 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b4ac2488-c99e-4380-8e82-e89207990c45-config\") pod \"b4ac2488-c99e-4380-8e82-e89207990c45\" (UID: \"b4ac2488-c99e-4380-8e82-e89207990c45\") " Oct 06 10:15:28 crc kubenswrapper[4824]: I1006 10:15:28.455159 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b4ac2488-c99e-4380-8e82-e89207990c45-ovsdbserver-nb\") pod \"b4ac2488-c99e-4380-8e82-e89207990c45\" (UID: \"b4ac2488-c99e-4380-8e82-e89207990c45\") " Oct 06 10:15:28 crc kubenswrapper[4824]: I1006 10:15:28.455365 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vwmp5\" (UniqueName: \"kubernetes.io/projected/b4ac2488-c99e-4380-8e82-e89207990c45-kube-api-access-vwmp5\") pod \"b4ac2488-c99e-4380-8e82-e89207990c45\" (UID: \"b4ac2488-c99e-4380-8e82-e89207990c45\") " Oct 06 10:15:28 crc kubenswrapper[4824]: I1006 10:15:28.455405 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b4ac2488-c99e-4380-8e82-e89207990c45-ovsdbserver-sb\") pod \"b4ac2488-c99e-4380-8e82-e89207990c45\" (UID: \"b4ac2488-c99e-4380-8e82-e89207990c45\") " Oct 06 10:15:28 crc kubenswrapper[4824]: I1006 10:15:28.460610 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b4ac2488-c99e-4380-8e82-e89207990c45-kube-api-access-vwmp5" (OuterVolumeSpecName: "kube-api-access-vwmp5") pod "b4ac2488-c99e-4380-8e82-e89207990c45" (UID: "b4ac2488-c99e-4380-8e82-e89207990c45"). InnerVolumeSpecName "kube-api-access-vwmp5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:15:28 crc kubenswrapper[4824]: I1006 10:15:28.498783 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b4ac2488-c99e-4380-8e82-e89207990c45-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "b4ac2488-c99e-4380-8e82-e89207990c45" (UID: "b4ac2488-c99e-4380-8e82-e89207990c45"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:15:28 crc kubenswrapper[4824]: I1006 10:15:28.500330 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b4ac2488-c99e-4380-8e82-e89207990c45-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b4ac2488-c99e-4380-8e82-e89207990c45" (UID: "b4ac2488-c99e-4380-8e82-e89207990c45"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:15:28 crc kubenswrapper[4824]: I1006 10:15:28.500348 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b4ac2488-c99e-4380-8e82-e89207990c45-config" (OuterVolumeSpecName: "config") pod "b4ac2488-c99e-4380-8e82-e89207990c45" (UID: "b4ac2488-c99e-4380-8e82-e89207990c45"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:15:28 crc kubenswrapper[4824]: I1006 10:15:28.502644 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b4ac2488-c99e-4380-8e82-e89207990c45-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "b4ac2488-c99e-4380-8e82-e89207990c45" (UID: "b4ac2488-c99e-4380-8e82-e89207990c45"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:15:28 crc kubenswrapper[4824]: I1006 10:15:28.545624 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-cwvfc" event={"ID":"b4ac2488-c99e-4380-8e82-e89207990c45","Type":"ContainerDied","Data":"6c52816fe2b22eba75635e304c08a8a12f3001fef7d495eceb80a745fc7a5a63"} Oct 06 10:15:28 crc kubenswrapper[4824]: I1006 10:15:28.545713 4824 scope.go:117] "RemoveContainer" containerID="50ee8359d95369ba4299b5af2a6db387bcddd6aee11c5ce974e82a168eb8a3a0" Oct 06 10:15:28 crc kubenswrapper[4824]: I1006 10:15:28.545652 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-cwvfc" Oct 06 10:15:28 crc kubenswrapper[4824]: E1006 10:15:28.548343 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"keystone-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-keystone:current-podified\\\"\"" pod="openstack/keystone-db-sync-s7fhb" podUID="e843b5e1-97a8-4a1c-8fa0-f3d8cf155ea0" Oct 06 10:15:28 crc kubenswrapper[4824]: I1006 10:15:28.557760 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vwmp5\" (UniqueName: \"kubernetes.io/projected/b4ac2488-c99e-4380-8e82-e89207990c45-kube-api-access-vwmp5\") on node \"crc\" DevicePath \"\"" Oct 06 10:15:28 crc kubenswrapper[4824]: I1006 10:15:28.558170 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b4ac2488-c99e-4380-8e82-e89207990c45-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 06 10:15:28 crc kubenswrapper[4824]: I1006 10:15:28.558184 4824 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b4ac2488-c99e-4380-8e82-e89207990c45-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 06 10:15:28 crc kubenswrapper[4824]: I1006 10:15:28.558196 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b4ac2488-c99e-4380-8e82-e89207990c45-config\") on node \"crc\" DevicePath \"\"" Oct 06 10:15:28 crc kubenswrapper[4824]: I1006 10:15:28.558205 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b4ac2488-c99e-4380-8e82-e89207990c45-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 06 10:15:28 crc kubenswrapper[4824]: I1006 10:15:28.581342 4824 scope.go:117] "RemoveContainer" containerID="6019be6266c0c1b65010463c73882989c76979fbcb0b846a20dc58af346d84e8" Oct 06 10:15:28 crc kubenswrapper[4824]: I1006 10:15:28.596506 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-cwvfc"] Oct 06 10:15:28 crc kubenswrapper[4824]: I1006 10:15:28.605974 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-698758b865-cwvfc"] Oct 06 10:15:29 crc kubenswrapper[4824]: I1006 10:15:29.303267 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b4ac2488-c99e-4380-8e82-e89207990c45" path="/var/lib/kubelet/pods/b4ac2488-c99e-4380-8e82-e89207990c45/volumes" Oct 06 10:15:29 crc kubenswrapper[4824]: I1006 10:15:29.561827 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-59pkf" event={"ID":"1bc5d3bf-5bde-4ce8-9b2a-50498e628b86","Type":"ContainerStarted","Data":"0bb94fc42b32d2a0dfc8802382f088d25416230c9e728cf1be47d5229e9d1ebe"} Oct 06 10:15:29 crc kubenswrapper[4824]: I1006 10:15:29.585091 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-59pkf" podStartSLOduration=2.70995055 podStartE2EDuration="21.585068675s" podCreationTimestamp="2025-10-06 10:15:08 +0000 UTC" firstStartedPulling="2025-10-06 10:15:09.183659466 +0000 UTC m=+1018.548082337" lastFinishedPulling="2025-10-06 10:15:28.058777601 +0000 UTC m=+1037.423200462" observedRunningTime="2025-10-06 10:15:29.579699512 +0000 UTC m=+1038.944122383" watchObservedRunningTime="2025-10-06 10:15:29.585068675 +0000 UTC m=+1038.949491546" Oct 06 10:15:34 crc kubenswrapper[4824]: I1006 10:15:34.630042 4824 generic.go:334] "Generic (PLEG): container finished" podID="1bc5d3bf-5bde-4ce8-9b2a-50498e628b86" containerID="0bb94fc42b32d2a0dfc8802382f088d25416230c9e728cf1be47d5229e9d1ebe" exitCode=0 Oct 06 10:15:34 crc kubenswrapper[4824]: I1006 10:15:34.630150 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-59pkf" event={"ID":"1bc5d3bf-5bde-4ce8-9b2a-50498e628b86","Type":"ContainerDied","Data":"0bb94fc42b32d2a0dfc8802382f088d25416230c9e728cf1be47d5229e9d1ebe"} Oct 06 10:15:36 crc kubenswrapper[4824]: I1006 10:15:36.032335 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-59pkf" Oct 06 10:15:36 crc kubenswrapper[4824]: I1006 10:15:36.038068 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bc5d3bf-5bde-4ce8-9b2a-50498e628b86-combined-ca-bundle\") pod \"1bc5d3bf-5bde-4ce8-9b2a-50498e628b86\" (UID: \"1bc5d3bf-5bde-4ce8-9b2a-50498e628b86\") " Oct 06 10:15:36 crc kubenswrapper[4824]: I1006 10:15:36.038150 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1bc5d3bf-5bde-4ce8-9b2a-50498e628b86-config-data\") pod \"1bc5d3bf-5bde-4ce8-9b2a-50498e628b86\" (UID: \"1bc5d3bf-5bde-4ce8-9b2a-50498e628b86\") " Oct 06 10:15:36 crc kubenswrapper[4824]: I1006 10:15:36.103184 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bc5d3bf-5bde-4ce8-9b2a-50498e628b86-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1bc5d3bf-5bde-4ce8-9b2a-50498e628b86" (UID: "1bc5d3bf-5bde-4ce8-9b2a-50498e628b86"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:15:36 crc kubenswrapper[4824]: I1006 10:15:36.139890 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mwj5b\" (UniqueName: \"kubernetes.io/projected/1bc5d3bf-5bde-4ce8-9b2a-50498e628b86-kube-api-access-mwj5b\") pod \"1bc5d3bf-5bde-4ce8-9b2a-50498e628b86\" (UID: \"1bc5d3bf-5bde-4ce8-9b2a-50498e628b86\") " Oct 06 10:15:36 crc kubenswrapper[4824]: I1006 10:15:36.140058 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/1bc5d3bf-5bde-4ce8-9b2a-50498e628b86-db-sync-config-data\") pod \"1bc5d3bf-5bde-4ce8-9b2a-50498e628b86\" (UID: \"1bc5d3bf-5bde-4ce8-9b2a-50498e628b86\") " Oct 06 10:15:36 crc kubenswrapper[4824]: I1006 10:15:36.140512 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bc5d3bf-5bde-4ce8-9b2a-50498e628b86-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 10:15:36 crc kubenswrapper[4824]: I1006 10:15:36.144516 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bc5d3bf-5bde-4ce8-9b2a-50498e628b86-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "1bc5d3bf-5bde-4ce8-9b2a-50498e628b86" (UID: "1bc5d3bf-5bde-4ce8-9b2a-50498e628b86"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:15:36 crc kubenswrapper[4824]: I1006 10:15:36.145572 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bc5d3bf-5bde-4ce8-9b2a-50498e628b86-kube-api-access-mwj5b" (OuterVolumeSpecName: "kube-api-access-mwj5b") pod "1bc5d3bf-5bde-4ce8-9b2a-50498e628b86" (UID: "1bc5d3bf-5bde-4ce8-9b2a-50498e628b86"). InnerVolumeSpecName "kube-api-access-mwj5b". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:15:36 crc kubenswrapper[4824]: I1006 10:15:36.146013 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bc5d3bf-5bde-4ce8-9b2a-50498e628b86-config-data" (OuterVolumeSpecName: "config-data") pod "1bc5d3bf-5bde-4ce8-9b2a-50498e628b86" (UID: "1bc5d3bf-5bde-4ce8-9b2a-50498e628b86"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:15:36 crc kubenswrapper[4824]: I1006 10:15:36.241468 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mwj5b\" (UniqueName: \"kubernetes.io/projected/1bc5d3bf-5bde-4ce8-9b2a-50498e628b86-kube-api-access-mwj5b\") on node \"crc\" DevicePath \"\"" Oct 06 10:15:36 crc kubenswrapper[4824]: I1006 10:15:36.241507 4824 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/1bc5d3bf-5bde-4ce8-9b2a-50498e628b86-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 10:15:36 crc kubenswrapper[4824]: I1006 10:15:36.241516 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1bc5d3bf-5bde-4ce8-9b2a-50498e628b86-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 10:15:36 crc kubenswrapper[4824]: I1006 10:15:36.656162 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-59pkf" event={"ID":"1bc5d3bf-5bde-4ce8-9b2a-50498e628b86","Type":"ContainerDied","Data":"8e6c6202927867d3cb6b8341a56e868f2c2f7f8d022d17414b123d2e4afa7828"} Oct 06 10:15:36 crc kubenswrapper[4824]: I1006 10:15:36.656238 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8e6c6202927867d3cb6b8341a56e868f2c2f7f8d022d17414b123d2e4afa7828" Oct 06 10:15:36 crc kubenswrapper[4824]: I1006 10:15:36.656238 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-59pkf" Oct 06 10:15:37 crc kubenswrapper[4824]: I1006 10:15:37.169817 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7ff5475cc9-9l788"] Oct 06 10:15:37 crc kubenswrapper[4824]: E1006 10:15:37.174862 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4ac2488-c99e-4380-8e82-e89207990c45" containerName="dnsmasq-dns" Oct 06 10:15:37 crc kubenswrapper[4824]: I1006 10:15:37.174957 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4ac2488-c99e-4380-8e82-e89207990c45" containerName="dnsmasq-dns" Oct 06 10:15:37 crc kubenswrapper[4824]: E1006 10:15:37.175119 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4ac2488-c99e-4380-8e82-e89207990c45" containerName="init" Oct 06 10:15:37 crc kubenswrapper[4824]: I1006 10:15:37.175198 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4ac2488-c99e-4380-8e82-e89207990c45" containerName="init" Oct 06 10:15:37 crc kubenswrapper[4824]: E1006 10:15:37.175290 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1bc5d3bf-5bde-4ce8-9b2a-50498e628b86" containerName="glance-db-sync" Oct 06 10:15:37 crc kubenswrapper[4824]: I1006 10:15:37.175358 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bc5d3bf-5bde-4ce8-9b2a-50498e628b86" containerName="glance-db-sync" Oct 06 10:15:37 crc kubenswrapper[4824]: I1006 10:15:37.175633 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="b4ac2488-c99e-4380-8e82-e89207990c45" containerName="dnsmasq-dns" Oct 06 10:15:37 crc kubenswrapper[4824]: I1006 10:15:37.175750 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="1bc5d3bf-5bde-4ce8-9b2a-50498e628b86" containerName="glance-db-sync" Oct 06 10:15:37 crc kubenswrapper[4824]: I1006 10:15:37.177078 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7ff5475cc9-9l788" Oct 06 10:15:37 crc kubenswrapper[4824]: I1006 10:15:37.202413 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7ff5475cc9-9l788"] Oct 06 10:15:37 crc kubenswrapper[4824]: I1006 10:15:37.363653 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/67457ff8-d89b-4f2b-8f49-b6323746b4b4-ovsdbserver-nb\") pod \"dnsmasq-dns-7ff5475cc9-9l788\" (UID: \"67457ff8-d89b-4f2b-8f49-b6323746b4b4\") " pod="openstack/dnsmasq-dns-7ff5475cc9-9l788" Oct 06 10:15:37 crc kubenswrapper[4824]: I1006 10:15:37.364131 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h9jx5\" (UniqueName: \"kubernetes.io/projected/67457ff8-d89b-4f2b-8f49-b6323746b4b4-kube-api-access-h9jx5\") pod \"dnsmasq-dns-7ff5475cc9-9l788\" (UID: \"67457ff8-d89b-4f2b-8f49-b6323746b4b4\") " pod="openstack/dnsmasq-dns-7ff5475cc9-9l788" Oct 06 10:15:37 crc kubenswrapper[4824]: I1006 10:15:37.364311 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/67457ff8-d89b-4f2b-8f49-b6323746b4b4-dns-svc\") pod \"dnsmasq-dns-7ff5475cc9-9l788\" (UID: \"67457ff8-d89b-4f2b-8f49-b6323746b4b4\") " pod="openstack/dnsmasq-dns-7ff5475cc9-9l788" Oct 06 10:15:37 crc kubenswrapper[4824]: I1006 10:15:37.364403 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/67457ff8-d89b-4f2b-8f49-b6323746b4b4-config\") pod \"dnsmasq-dns-7ff5475cc9-9l788\" (UID: \"67457ff8-d89b-4f2b-8f49-b6323746b4b4\") " pod="openstack/dnsmasq-dns-7ff5475cc9-9l788" Oct 06 10:15:37 crc kubenswrapper[4824]: I1006 10:15:37.364485 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/67457ff8-d89b-4f2b-8f49-b6323746b4b4-dns-swift-storage-0\") pod \"dnsmasq-dns-7ff5475cc9-9l788\" (UID: \"67457ff8-d89b-4f2b-8f49-b6323746b4b4\") " pod="openstack/dnsmasq-dns-7ff5475cc9-9l788" Oct 06 10:15:37 crc kubenswrapper[4824]: I1006 10:15:37.364591 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/67457ff8-d89b-4f2b-8f49-b6323746b4b4-ovsdbserver-sb\") pod \"dnsmasq-dns-7ff5475cc9-9l788\" (UID: \"67457ff8-d89b-4f2b-8f49-b6323746b4b4\") " pod="openstack/dnsmasq-dns-7ff5475cc9-9l788" Oct 06 10:15:37 crc kubenswrapper[4824]: I1006 10:15:37.466438 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/67457ff8-d89b-4f2b-8f49-b6323746b4b4-ovsdbserver-nb\") pod \"dnsmasq-dns-7ff5475cc9-9l788\" (UID: \"67457ff8-d89b-4f2b-8f49-b6323746b4b4\") " pod="openstack/dnsmasq-dns-7ff5475cc9-9l788" Oct 06 10:15:37 crc kubenswrapper[4824]: I1006 10:15:37.466538 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h9jx5\" (UniqueName: \"kubernetes.io/projected/67457ff8-d89b-4f2b-8f49-b6323746b4b4-kube-api-access-h9jx5\") pod \"dnsmasq-dns-7ff5475cc9-9l788\" (UID: \"67457ff8-d89b-4f2b-8f49-b6323746b4b4\") " pod="openstack/dnsmasq-dns-7ff5475cc9-9l788" Oct 06 10:15:37 crc kubenswrapper[4824]: I1006 10:15:37.466594 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/67457ff8-d89b-4f2b-8f49-b6323746b4b4-dns-svc\") pod \"dnsmasq-dns-7ff5475cc9-9l788\" (UID: \"67457ff8-d89b-4f2b-8f49-b6323746b4b4\") " pod="openstack/dnsmasq-dns-7ff5475cc9-9l788" Oct 06 10:15:37 crc kubenswrapper[4824]: I1006 10:15:37.466623 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/67457ff8-d89b-4f2b-8f49-b6323746b4b4-config\") pod \"dnsmasq-dns-7ff5475cc9-9l788\" (UID: \"67457ff8-d89b-4f2b-8f49-b6323746b4b4\") " pod="openstack/dnsmasq-dns-7ff5475cc9-9l788" Oct 06 10:15:37 crc kubenswrapper[4824]: I1006 10:15:37.466648 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/67457ff8-d89b-4f2b-8f49-b6323746b4b4-dns-swift-storage-0\") pod \"dnsmasq-dns-7ff5475cc9-9l788\" (UID: \"67457ff8-d89b-4f2b-8f49-b6323746b4b4\") " pod="openstack/dnsmasq-dns-7ff5475cc9-9l788" Oct 06 10:15:37 crc kubenswrapper[4824]: I1006 10:15:37.466689 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/67457ff8-d89b-4f2b-8f49-b6323746b4b4-ovsdbserver-sb\") pod \"dnsmasq-dns-7ff5475cc9-9l788\" (UID: \"67457ff8-d89b-4f2b-8f49-b6323746b4b4\") " pod="openstack/dnsmasq-dns-7ff5475cc9-9l788" Oct 06 10:15:37 crc kubenswrapper[4824]: I1006 10:15:37.467625 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/67457ff8-d89b-4f2b-8f49-b6323746b4b4-ovsdbserver-sb\") pod \"dnsmasq-dns-7ff5475cc9-9l788\" (UID: \"67457ff8-d89b-4f2b-8f49-b6323746b4b4\") " pod="openstack/dnsmasq-dns-7ff5475cc9-9l788" Oct 06 10:15:37 crc kubenswrapper[4824]: I1006 10:15:37.467663 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/67457ff8-d89b-4f2b-8f49-b6323746b4b4-ovsdbserver-nb\") pod \"dnsmasq-dns-7ff5475cc9-9l788\" (UID: \"67457ff8-d89b-4f2b-8f49-b6323746b4b4\") " pod="openstack/dnsmasq-dns-7ff5475cc9-9l788" Oct 06 10:15:37 crc kubenswrapper[4824]: I1006 10:15:37.468408 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/67457ff8-d89b-4f2b-8f49-b6323746b4b4-config\") pod \"dnsmasq-dns-7ff5475cc9-9l788\" (UID: \"67457ff8-d89b-4f2b-8f49-b6323746b4b4\") " pod="openstack/dnsmasq-dns-7ff5475cc9-9l788" Oct 06 10:15:37 crc kubenswrapper[4824]: I1006 10:15:37.468548 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/67457ff8-d89b-4f2b-8f49-b6323746b4b4-dns-swift-storage-0\") pod \"dnsmasq-dns-7ff5475cc9-9l788\" (UID: \"67457ff8-d89b-4f2b-8f49-b6323746b4b4\") " pod="openstack/dnsmasq-dns-7ff5475cc9-9l788" Oct 06 10:15:37 crc kubenswrapper[4824]: I1006 10:15:37.469037 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/67457ff8-d89b-4f2b-8f49-b6323746b4b4-dns-svc\") pod \"dnsmasq-dns-7ff5475cc9-9l788\" (UID: \"67457ff8-d89b-4f2b-8f49-b6323746b4b4\") " pod="openstack/dnsmasq-dns-7ff5475cc9-9l788" Oct 06 10:15:37 crc kubenswrapper[4824]: I1006 10:15:37.497315 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h9jx5\" (UniqueName: \"kubernetes.io/projected/67457ff8-d89b-4f2b-8f49-b6323746b4b4-kube-api-access-h9jx5\") pod \"dnsmasq-dns-7ff5475cc9-9l788\" (UID: \"67457ff8-d89b-4f2b-8f49-b6323746b4b4\") " pod="openstack/dnsmasq-dns-7ff5475cc9-9l788" Oct 06 10:15:37 crc kubenswrapper[4824]: I1006 10:15:37.797823 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7ff5475cc9-9l788" Oct 06 10:15:38 crc kubenswrapper[4824]: I1006 10:15:38.221325 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7ff5475cc9-9l788"] Oct 06 10:15:38 crc kubenswrapper[4824]: I1006 10:15:38.674619 4824 generic.go:334] "Generic (PLEG): container finished" podID="67457ff8-d89b-4f2b-8f49-b6323746b4b4" containerID="fcc28e95b9c1cf3a50d7a83b68d38848bb7e2b857eede943b4a727c3f8c0b23c" exitCode=0 Oct 06 10:15:38 crc kubenswrapper[4824]: I1006 10:15:38.675147 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7ff5475cc9-9l788" event={"ID":"67457ff8-d89b-4f2b-8f49-b6323746b4b4","Type":"ContainerDied","Data":"fcc28e95b9c1cf3a50d7a83b68d38848bb7e2b857eede943b4a727c3f8c0b23c"} Oct 06 10:15:38 crc kubenswrapper[4824]: I1006 10:15:38.675178 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7ff5475cc9-9l788" event={"ID":"67457ff8-d89b-4f2b-8f49-b6323746b4b4","Type":"ContainerStarted","Data":"06554e9c3d235541783f28313fcf3695b11706ce9c9bde7757a9ef53d4c595d4"} Oct 06 10:15:39 crc kubenswrapper[4824]: I1006 10:15:39.686296 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7ff5475cc9-9l788" event={"ID":"67457ff8-d89b-4f2b-8f49-b6323746b4b4","Type":"ContainerStarted","Data":"f172c000d0b1e8b0c4db204fddd4668a7710b99c62386eff0a34b706c5d642c1"} Oct 06 10:15:39 crc kubenswrapper[4824]: I1006 10:15:39.686843 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7ff5475cc9-9l788" Oct 06 10:15:39 crc kubenswrapper[4824]: I1006 10:15:39.715564 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7ff5475cc9-9l788" podStartSLOduration=2.715530341 podStartE2EDuration="2.715530341s" podCreationTimestamp="2025-10-06 10:15:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:15:39.709079363 +0000 UTC m=+1049.073502254" watchObservedRunningTime="2025-10-06 10:15:39.715530341 +0000 UTC m=+1049.079953192" Oct 06 10:15:40 crc kubenswrapper[4824]: I1006 10:15:40.696211 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-s7fhb" event={"ID":"e843b5e1-97a8-4a1c-8fa0-f3d8cf155ea0","Type":"ContainerStarted","Data":"f23c07b390abda237c8434ff9ade9da85768a81dd5fc716e96733d119b669bce"} Oct 06 10:15:40 crc kubenswrapper[4824]: I1006 10:15:40.722738 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-s7fhb" podStartSLOduration=2.456652583 podStartE2EDuration="28.722713929s" podCreationTimestamp="2025-10-06 10:15:12 +0000 UTC" firstStartedPulling="2025-10-06 10:15:13.655601563 +0000 UTC m=+1023.020024414" lastFinishedPulling="2025-10-06 10:15:39.921662899 +0000 UTC m=+1049.286085760" observedRunningTime="2025-10-06 10:15:40.717737235 +0000 UTC m=+1050.082160096" watchObservedRunningTime="2025-10-06 10:15:40.722713929 +0000 UTC m=+1050.087136820" Oct 06 10:15:43 crc kubenswrapper[4824]: I1006 10:15:43.735395 4824 generic.go:334] "Generic (PLEG): container finished" podID="e843b5e1-97a8-4a1c-8fa0-f3d8cf155ea0" containerID="f23c07b390abda237c8434ff9ade9da85768a81dd5fc716e96733d119b669bce" exitCode=0 Oct 06 10:15:43 crc kubenswrapper[4824]: I1006 10:15:43.735474 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-s7fhb" event={"ID":"e843b5e1-97a8-4a1c-8fa0-f3d8cf155ea0","Type":"ContainerDied","Data":"f23c07b390abda237c8434ff9ade9da85768a81dd5fc716e96733d119b669bce"} Oct 06 10:15:43 crc kubenswrapper[4824]: I1006 10:15:43.915082 4824 patch_prober.go:28] interesting pod/machine-config-daemon-khgzw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 10:15:43 crc kubenswrapper[4824]: I1006 10:15:43.915173 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 10:15:43 crc kubenswrapper[4824]: I1006 10:15:43.915235 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" Oct 06 10:15:43 crc kubenswrapper[4824]: I1006 10:15:43.916167 4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"797b5c3c4be7e935c4c45dea62e8a371c48208db7d2ac607bd9b1a5fbc29630b"} pod="openshift-machine-config-operator/machine-config-daemon-khgzw" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 06 10:15:43 crc kubenswrapper[4824]: I1006 10:15:43.916292 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" containerName="machine-config-daemon" containerID="cri-o://797b5c3c4be7e935c4c45dea62e8a371c48208db7d2ac607bd9b1a5fbc29630b" gracePeriod=600 Oct 06 10:15:44 crc kubenswrapper[4824]: I1006 10:15:44.755672 4824 generic.go:334] "Generic (PLEG): container finished" podID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" containerID="797b5c3c4be7e935c4c45dea62e8a371c48208db7d2ac607bd9b1a5fbc29630b" exitCode=0 Oct 06 10:15:44 crc kubenswrapper[4824]: I1006 10:15:44.755774 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" event={"ID":"e1140c8c-93f1-4fce-8c06-a69561ff8a6d","Type":"ContainerDied","Data":"797b5c3c4be7e935c4c45dea62e8a371c48208db7d2ac607bd9b1a5fbc29630b"} Oct 06 10:15:44 crc kubenswrapper[4824]: I1006 10:15:44.756207 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" event={"ID":"e1140c8c-93f1-4fce-8c06-a69561ff8a6d","Type":"ContainerStarted","Data":"ad63fa5818247f7e35384bdeb2bf047f095a2d6eed77601da04ba8f26adf2f37"} Oct 06 10:15:44 crc kubenswrapper[4824]: I1006 10:15:44.756245 4824 scope.go:117] "RemoveContainer" containerID="8fa1ede4cec498bdfe66760d5e74fd58534149105068e959659c8b228527cc04" Oct 06 10:15:45 crc kubenswrapper[4824]: I1006 10:15:45.174839 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-s7fhb" Oct 06 10:15:45 crc kubenswrapper[4824]: I1006 10:15:45.339668 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9v4c9\" (UniqueName: \"kubernetes.io/projected/e843b5e1-97a8-4a1c-8fa0-f3d8cf155ea0-kube-api-access-9v4c9\") pod \"e843b5e1-97a8-4a1c-8fa0-f3d8cf155ea0\" (UID: \"e843b5e1-97a8-4a1c-8fa0-f3d8cf155ea0\") " Oct 06 10:15:45 crc kubenswrapper[4824]: I1006 10:15:45.340400 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e843b5e1-97a8-4a1c-8fa0-f3d8cf155ea0-config-data\") pod \"e843b5e1-97a8-4a1c-8fa0-f3d8cf155ea0\" (UID: \"e843b5e1-97a8-4a1c-8fa0-f3d8cf155ea0\") " Oct 06 10:15:45 crc kubenswrapper[4824]: I1006 10:15:45.340751 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e843b5e1-97a8-4a1c-8fa0-f3d8cf155ea0-combined-ca-bundle\") pod \"e843b5e1-97a8-4a1c-8fa0-f3d8cf155ea0\" (UID: \"e843b5e1-97a8-4a1c-8fa0-f3d8cf155ea0\") " Oct 06 10:15:45 crc kubenswrapper[4824]: I1006 10:15:45.361279 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e843b5e1-97a8-4a1c-8fa0-f3d8cf155ea0-kube-api-access-9v4c9" (OuterVolumeSpecName: "kube-api-access-9v4c9") pod "e843b5e1-97a8-4a1c-8fa0-f3d8cf155ea0" (UID: "e843b5e1-97a8-4a1c-8fa0-f3d8cf155ea0"). InnerVolumeSpecName "kube-api-access-9v4c9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:15:45 crc kubenswrapper[4824]: I1006 10:15:45.387707 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e843b5e1-97a8-4a1c-8fa0-f3d8cf155ea0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e843b5e1-97a8-4a1c-8fa0-f3d8cf155ea0" (UID: "e843b5e1-97a8-4a1c-8fa0-f3d8cf155ea0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:15:45 crc kubenswrapper[4824]: I1006 10:15:45.416589 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e843b5e1-97a8-4a1c-8fa0-f3d8cf155ea0-config-data" (OuterVolumeSpecName: "config-data") pod "e843b5e1-97a8-4a1c-8fa0-f3d8cf155ea0" (UID: "e843b5e1-97a8-4a1c-8fa0-f3d8cf155ea0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:15:45 crc kubenswrapper[4824]: I1006 10:15:45.445954 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9v4c9\" (UniqueName: \"kubernetes.io/projected/e843b5e1-97a8-4a1c-8fa0-f3d8cf155ea0-kube-api-access-9v4c9\") on node \"crc\" DevicePath \"\"" Oct 06 10:15:45 crc kubenswrapper[4824]: I1006 10:15:45.446029 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e843b5e1-97a8-4a1c-8fa0-f3d8cf155ea0-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 10:15:45 crc kubenswrapper[4824]: I1006 10:15:45.446281 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e843b5e1-97a8-4a1c-8fa0-f3d8cf155ea0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 10:15:45 crc kubenswrapper[4824]: I1006 10:15:45.772995 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-s7fhb" Oct 06 10:15:45 crc kubenswrapper[4824]: I1006 10:15:45.774073 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-s7fhb" event={"ID":"e843b5e1-97a8-4a1c-8fa0-f3d8cf155ea0","Type":"ContainerDied","Data":"c12f731634f31923dee3721868b11bf96850bb2fe531ee1485ecc73284d5f331"} Oct 06 10:15:45 crc kubenswrapper[4824]: I1006 10:15:45.774869 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c12f731634f31923dee3721868b11bf96850bb2fe531ee1485ecc73284d5f331" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.006226 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-zcq55"] Oct 06 10:15:46 crc kubenswrapper[4824]: E1006 10:15:46.006872 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e843b5e1-97a8-4a1c-8fa0-f3d8cf155ea0" containerName="keystone-db-sync" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.006900 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="e843b5e1-97a8-4a1c-8fa0-f3d8cf155ea0" containerName="keystone-db-sync" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.007234 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="e843b5e1-97a8-4a1c-8fa0-f3d8cf155ea0" containerName="keystone-db-sync" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.008449 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-zcq55" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.011346 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.011634 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-lxhxz" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.011706 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.013034 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.015643 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7ff5475cc9-9l788"] Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.015937 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7ff5475cc9-9l788" podUID="67457ff8-d89b-4f2b-8f49-b6323746b4b4" containerName="dnsmasq-dns" containerID="cri-o://f172c000d0b1e8b0c4db204fddd4668a7710b99c62386eff0a34b706c5d642c1" gracePeriod=10 Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.018324 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7ff5475cc9-9l788" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.028529 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-zcq55"] Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.090963 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5c5cc7c5ff-w7tn4"] Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.093379 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c5cc7c5ff-w7tn4" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.144301 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c5cc7c5ff-w7tn4"] Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.178237 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f735f6b9-a885-4a39-b780-17d8d1b30a88-dns-swift-storage-0\") pod \"dnsmasq-dns-5c5cc7c5ff-w7tn4\" (UID: \"f735f6b9-a885-4a39-b780-17d8d1b30a88\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-w7tn4" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.178349 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c607b17a-75ac-4e7f-8ff3-9d0edb23cc23-config-data\") pod \"keystone-bootstrap-zcq55\" (UID: \"c607b17a-75ac-4e7f-8ff3-9d0edb23cc23\") " pod="openstack/keystone-bootstrap-zcq55" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.178403 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f735f6b9-a885-4a39-b780-17d8d1b30a88-ovsdbserver-nb\") pod \"dnsmasq-dns-5c5cc7c5ff-w7tn4\" (UID: \"f735f6b9-a885-4a39-b780-17d8d1b30a88\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-w7tn4" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.178441 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c607b17a-75ac-4e7f-8ff3-9d0edb23cc23-combined-ca-bundle\") pod \"keystone-bootstrap-zcq55\" (UID: \"c607b17a-75ac-4e7f-8ff3-9d0edb23cc23\") " pod="openstack/keystone-bootstrap-zcq55" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.178581 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c607b17a-75ac-4e7f-8ff3-9d0edb23cc23-credential-keys\") pod \"keystone-bootstrap-zcq55\" (UID: \"c607b17a-75ac-4e7f-8ff3-9d0edb23cc23\") " pod="openstack/keystone-bootstrap-zcq55" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.178621 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b6msl\" (UniqueName: \"kubernetes.io/projected/c607b17a-75ac-4e7f-8ff3-9d0edb23cc23-kube-api-access-b6msl\") pod \"keystone-bootstrap-zcq55\" (UID: \"c607b17a-75ac-4e7f-8ff3-9d0edb23cc23\") " pod="openstack/keystone-bootstrap-zcq55" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.178659 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c607b17a-75ac-4e7f-8ff3-9d0edb23cc23-fernet-keys\") pod \"keystone-bootstrap-zcq55\" (UID: \"c607b17a-75ac-4e7f-8ff3-9d0edb23cc23\") " pod="openstack/keystone-bootstrap-zcq55" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.178722 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c607b17a-75ac-4e7f-8ff3-9d0edb23cc23-scripts\") pod \"keystone-bootstrap-zcq55\" (UID: \"c607b17a-75ac-4e7f-8ff3-9d0edb23cc23\") " pod="openstack/keystone-bootstrap-zcq55" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.178748 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f735f6b9-a885-4a39-b780-17d8d1b30a88-dns-svc\") pod \"dnsmasq-dns-5c5cc7c5ff-w7tn4\" (UID: \"f735f6b9-a885-4a39-b780-17d8d1b30a88\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-w7tn4" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.178837 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f735f6b9-a885-4a39-b780-17d8d1b30a88-config\") pod \"dnsmasq-dns-5c5cc7c5ff-w7tn4\" (UID: \"f735f6b9-a885-4a39-b780-17d8d1b30a88\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-w7tn4" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.178912 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f735f6b9-a885-4a39-b780-17d8d1b30a88-ovsdbserver-sb\") pod \"dnsmasq-dns-5c5cc7c5ff-w7tn4\" (UID: \"f735f6b9-a885-4a39-b780-17d8d1b30a88\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-w7tn4" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.179060 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7k85q\" (UniqueName: \"kubernetes.io/projected/f735f6b9-a885-4a39-b780-17d8d1b30a88-kube-api-access-7k85q\") pod \"dnsmasq-dns-5c5cc7c5ff-w7tn4\" (UID: \"f735f6b9-a885-4a39-b780-17d8d1b30a88\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-w7tn4" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.289705 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f735f6b9-a885-4a39-b780-17d8d1b30a88-ovsdbserver-sb\") pod \"dnsmasq-dns-5c5cc7c5ff-w7tn4\" (UID: \"f735f6b9-a885-4a39-b780-17d8d1b30a88\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-w7tn4" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.289759 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7k85q\" (UniqueName: \"kubernetes.io/projected/f735f6b9-a885-4a39-b780-17d8d1b30a88-kube-api-access-7k85q\") pod \"dnsmasq-dns-5c5cc7c5ff-w7tn4\" (UID: \"f735f6b9-a885-4a39-b780-17d8d1b30a88\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-w7tn4" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.289786 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f735f6b9-a885-4a39-b780-17d8d1b30a88-dns-swift-storage-0\") pod \"dnsmasq-dns-5c5cc7c5ff-w7tn4\" (UID: \"f735f6b9-a885-4a39-b780-17d8d1b30a88\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-w7tn4" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.289816 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c607b17a-75ac-4e7f-8ff3-9d0edb23cc23-config-data\") pod \"keystone-bootstrap-zcq55\" (UID: \"c607b17a-75ac-4e7f-8ff3-9d0edb23cc23\") " pod="openstack/keystone-bootstrap-zcq55" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.289842 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f735f6b9-a885-4a39-b780-17d8d1b30a88-ovsdbserver-nb\") pod \"dnsmasq-dns-5c5cc7c5ff-w7tn4\" (UID: \"f735f6b9-a885-4a39-b780-17d8d1b30a88\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-w7tn4" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.289864 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c607b17a-75ac-4e7f-8ff3-9d0edb23cc23-combined-ca-bundle\") pod \"keystone-bootstrap-zcq55\" (UID: \"c607b17a-75ac-4e7f-8ff3-9d0edb23cc23\") " pod="openstack/keystone-bootstrap-zcq55" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.289899 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c607b17a-75ac-4e7f-8ff3-9d0edb23cc23-credential-keys\") pod \"keystone-bootstrap-zcq55\" (UID: \"c607b17a-75ac-4e7f-8ff3-9d0edb23cc23\") " pod="openstack/keystone-bootstrap-zcq55" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.289918 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b6msl\" (UniqueName: \"kubernetes.io/projected/c607b17a-75ac-4e7f-8ff3-9d0edb23cc23-kube-api-access-b6msl\") pod \"keystone-bootstrap-zcq55\" (UID: \"c607b17a-75ac-4e7f-8ff3-9d0edb23cc23\") " pod="openstack/keystone-bootstrap-zcq55" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.289935 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c607b17a-75ac-4e7f-8ff3-9d0edb23cc23-fernet-keys\") pod \"keystone-bootstrap-zcq55\" (UID: \"c607b17a-75ac-4e7f-8ff3-9d0edb23cc23\") " pod="openstack/keystone-bootstrap-zcq55" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.289962 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c607b17a-75ac-4e7f-8ff3-9d0edb23cc23-scripts\") pod \"keystone-bootstrap-zcq55\" (UID: \"c607b17a-75ac-4e7f-8ff3-9d0edb23cc23\") " pod="openstack/keystone-bootstrap-zcq55" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.289995 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f735f6b9-a885-4a39-b780-17d8d1b30a88-dns-svc\") pod \"dnsmasq-dns-5c5cc7c5ff-w7tn4\" (UID: \"f735f6b9-a885-4a39-b780-17d8d1b30a88\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-w7tn4" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.290035 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f735f6b9-a885-4a39-b780-17d8d1b30a88-config\") pod \"dnsmasq-dns-5c5cc7c5ff-w7tn4\" (UID: \"f735f6b9-a885-4a39-b780-17d8d1b30a88\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-w7tn4" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.290943 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f735f6b9-a885-4a39-b780-17d8d1b30a88-config\") pod \"dnsmasq-dns-5c5cc7c5ff-w7tn4\" (UID: \"f735f6b9-a885-4a39-b780-17d8d1b30a88\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-w7tn4" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.291465 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f735f6b9-a885-4a39-b780-17d8d1b30a88-ovsdbserver-sb\") pod \"dnsmasq-dns-5c5cc7c5ff-w7tn4\" (UID: \"f735f6b9-a885-4a39-b780-17d8d1b30a88\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-w7tn4" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.292251 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f735f6b9-a885-4a39-b780-17d8d1b30a88-dns-swift-storage-0\") pod \"dnsmasq-dns-5c5cc7c5ff-w7tn4\" (UID: \"f735f6b9-a885-4a39-b780-17d8d1b30a88\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-w7tn4" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.303502 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f735f6b9-a885-4a39-b780-17d8d1b30a88-ovsdbserver-nb\") pod \"dnsmasq-dns-5c5cc7c5ff-w7tn4\" (UID: \"f735f6b9-a885-4a39-b780-17d8d1b30a88\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-w7tn4" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.304062 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-ldn9v"] Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.305918 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-ldn9v" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.309848 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c607b17a-75ac-4e7f-8ff3-9d0edb23cc23-combined-ca-bundle\") pod \"keystone-bootstrap-zcq55\" (UID: \"c607b17a-75ac-4e7f-8ff3-9d0edb23cc23\") " pod="openstack/keystone-bootstrap-zcq55" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.320436 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-v75vf" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.320648 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.320793 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.321592 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f735f6b9-a885-4a39-b780-17d8d1b30a88-dns-svc\") pod \"dnsmasq-dns-5c5cc7c5ff-w7tn4\" (UID: \"f735f6b9-a885-4a39-b780-17d8d1b30a88\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-w7tn4" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.324785 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c607b17a-75ac-4e7f-8ff3-9d0edb23cc23-credential-keys\") pod \"keystone-bootstrap-zcq55\" (UID: \"c607b17a-75ac-4e7f-8ff3-9d0edb23cc23\") " pod="openstack/keystone-bootstrap-zcq55" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.330967 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c607b17a-75ac-4e7f-8ff3-9d0edb23cc23-fernet-keys\") pod \"keystone-bootstrap-zcq55\" (UID: \"c607b17a-75ac-4e7f-8ff3-9d0edb23cc23\") " pod="openstack/keystone-bootstrap-zcq55" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.339381 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7k85q\" (UniqueName: \"kubernetes.io/projected/f735f6b9-a885-4a39-b780-17d8d1b30a88-kube-api-access-7k85q\") pod \"dnsmasq-dns-5c5cc7c5ff-w7tn4\" (UID: \"f735f6b9-a885-4a39-b780-17d8d1b30a88\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-w7tn4" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.339614 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c607b17a-75ac-4e7f-8ff3-9d0edb23cc23-config-data\") pod \"keystone-bootstrap-zcq55\" (UID: \"c607b17a-75ac-4e7f-8ff3-9d0edb23cc23\") " pod="openstack/keystone-bootstrap-zcq55" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.339898 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-ldn9v"] Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.341451 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c607b17a-75ac-4e7f-8ff3-9d0edb23cc23-scripts\") pod \"keystone-bootstrap-zcq55\" (UID: \"c607b17a-75ac-4e7f-8ff3-9d0edb23cc23\") " pod="openstack/keystone-bootstrap-zcq55" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.394550 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-7749656757-6jdvp"] Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.403016 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b6msl\" (UniqueName: \"kubernetes.io/projected/c607b17a-75ac-4e7f-8ff3-9d0edb23cc23-kube-api-access-b6msl\") pod \"keystone-bootstrap-zcq55\" (UID: \"c607b17a-75ac-4e7f-8ff3-9d0edb23cc23\") " pod="openstack/keystone-bootstrap-zcq55" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.422292 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7749656757-6jdvp" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.426130 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7749656757-6jdvp"] Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.426350 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.426480 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.437887 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-l699n" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.438521 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.445432 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.447522 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.459311 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.459480 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.464721 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.495432 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2382b80c-75fc-435f-be5f-0f1eb6a5cd17-config-data\") pod \"cinder-db-sync-ldn9v\" (UID: \"2382b80c-75fc-435f-be5f-0f1eb6a5cd17\") " pod="openstack/cinder-db-sync-ldn9v" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.495514 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2382b80c-75fc-435f-be5f-0f1eb6a5cd17-scripts\") pod \"cinder-db-sync-ldn9v\" (UID: \"2382b80c-75fc-435f-be5f-0f1eb6a5cd17\") " pod="openstack/cinder-db-sync-ldn9v" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.495556 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/2382b80c-75fc-435f-be5f-0f1eb6a5cd17-db-sync-config-data\") pod \"cinder-db-sync-ldn9v\" (UID: \"2382b80c-75fc-435f-be5f-0f1eb6a5cd17\") " pod="openstack/cinder-db-sync-ldn9v" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.495632 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2382b80c-75fc-435f-be5f-0f1eb6a5cd17-etc-machine-id\") pod \"cinder-db-sync-ldn9v\" (UID: \"2382b80c-75fc-435f-be5f-0f1eb6a5cd17\") " pod="openstack/cinder-db-sync-ldn9v" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.495673 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2382b80c-75fc-435f-be5f-0f1eb6a5cd17-combined-ca-bundle\") pod \"cinder-db-sync-ldn9v\" (UID: \"2382b80c-75fc-435f-be5f-0f1eb6a5cd17\") " pod="openstack/cinder-db-sync-ldn9v" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.495733 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fqk7s\" (UniqueName: \"kubernetes.io/projected/2382b80c-75fc-435f-be5f-0f1eb6a5cd17-kube-api-access-fqk7s\") pod \"cinder-db-sync-ldn9v\" (UID: \"2382b80c-75fc-435f-be5f-0f1eb6a5cd17\") " pod="openstack/cinder-db-sync-ldn9v" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.515665 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-ljqzg"] Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.517502 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-ljqzg" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.524888 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-zw6pr" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.525167 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.538363 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-ljqzg"] Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.551592 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c5cc7c5ff-w7tn4" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.603892 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-ccd5f54d9-dsnj5"] Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.606267 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/07c13b84-6c2a-4a6c-b23b-df1e14e077b5-horizon-secret-key\") pod \"horizon-7749656757-6jdvp\" (UID: \"07c13b84-6c2a-4a6c-b23b-df1e14e077b5\") " pod="openstack/horizon-7749656757-6jdvp" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.606350 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/13781991-323c-49ae-9c5d-625e2ffd019b-log-httpd\") pod \"ceilometer-0\" (UID: \"13781991-323c-49ae-9c5d-625e2ffd019b\") " pod="openstack/ceilometer-0" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.606433 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2382b80c-75fc-435f-be5f-0f1eb6a5cd17-etc-machine-id\") pod \"cinder-db-sync-ldn9v\" (UID: \"2382b80c-75fc-435f-be5f-0f1eb6a5cd17\") " pod="openstack/cinder-db-sync-ldn9v" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.606461 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/13781991-323c-49ae-9c5d-625e2ffd019b-config-data\") pod \"ceilometer-0\" (UID: \"13781991-323c-49ae-9c5d-625e2ffd019b\") " pod="openstack/ceilometer-0" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.606489 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/13781991-323c-49ae-9c5d-625e2ffd019b-scripts\") pod \"ceilometer-0\" (UID: \"13781991-323c-49ae-9c5d-625e2ffd019b\") " pod="openstack/ceilometer-0" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.606512 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/07c13b84-6c2a-4a6c-b23b-df1e14e077b5-logs\") pod \"horizon-7749656757-6jdvp\" (UID: \"07c13b84-6c2a-4a6c-b23b-df1e14e077b5\") " pod="openstack/horizon-7749656757-6jdvp" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.606531 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/13781991-323c-49ae-9c5d-625e2ffd019b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"13781991-323c-49ae-9c5d-625e2ffd019b\") " pod="openstack/ceilometer-0" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.606564 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2382b80c-75fc-435f-be5f-0f1eb6a5cd17-combined-ca-bundle\") pod \"cinder-db-sync-ldn9v\" (UID: \"2382b80c-75fc-435f-be5f-0f1eb6a5cd17\") " pod="openstack/cinder-db-sync-ldn9v" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.606584 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/07c13b84-6c2a-4a6c-b23b-df1e14e077b5-config-data\") pod \"horizon-7749656757-6jdvp\" (UID: \"07c13b84-6c2a-4a6c-b23b-df1e14e077b5\") " pod="openstack/horizon-7749656757-6jdvp" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.606607 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/07c13b84-6c2a-4a6c-b23b-df1e14e077b5-scripts\") pod \"horizon-7749656757-6jdvp\" (UID: \"07c13b84-6c2a-4a6c-b23b-df1e14e077b5\") " pod="openstack/horizon-7749656757-6jdvp" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.606646 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/13781991-323c-49ae-9c5d-625e2ffd019b-run-httpd\") pod \"ceilometer-0\" (UID: \"13781991-323c-49ae-9c5d-625e2ffd019b\") " pod="openstack/ceilometer-0" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.606674 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vt2nw\" (UniqueName: \"kubernetes.io/projected/07c13b84-6c2a-4a6c-b23b-df1e14e077b5-kube-api-access-vt2nw\") pod \"horizon-7749656757-6jdvp\" (UID: \"07c13b84-6c2a-4a6c-b23b-df1e14e077b5\") " pod="openstack/horizon-7749656757-6jdvp" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.606695 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w6rqx\" (UniqueName: \"kubernetes.io/projected/13781991-323c-49ae-9c5d-625e2ffd019b-kube-api-access-w6rqx\") pod \"ceilometer-0\" (UID: \"13781991-323c-49ae-9c5d-625e2ffd019b\") " pod="openstack/ceilometer-0" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.606719 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fqk7s\" (UniqueName: \"kubernetes.io/projected/2382b80c-75fc-435f-be5f-0f1eb6a5cd17-kube-api-access-fqk7s\") pod \"cinder-db-sync-ldn9v\" (UID: \"2382b80c-75fc-435f-be5f-0f1eb6a5cd17\") " pod="openstack/cinder-db-sync-ldn9v" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.606748 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2382b80c-75fc-435f-be5f-0f1eb6a5cd17-config-data\") pod \"cinder-db-sync-ldn9v\" (UID: \"2382b80c-75fc-435f-be5f-0f1eb6a5cd17\") " pod="openstack/cinder-db-sync-ldn9v" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.606779 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2382b80c-75fc-435f-be5f-0f1eb6a5cd17-scripts\") pod \"cinder-db-sync-ldn9v\" (UID: \"2382b80c-75fc-435f-be5f-0f1eb6a5cd17\") " pod="openstack/cinder-db-sync-ldn9v" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.606804 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13781991-323c-49ae-9c5d-625e2ffd019b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"13781991-323c-49ae-9c5d-625e2ffd019b\") " pod="openstack/ceilometer-0" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.606832 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/2382b80c-75fc-435f-be5f-0f1eb6a5cd17-db-sync-config-data\") pod \"cinder-db-sync-ldn9v\" (UID: \"2382b80c-75fc-435f-be5f-0f1eb6a5cd17\") " pod="openstack/cinder-db-sync-ldn9v" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.607035 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2382b80c-75fc-435f-be5f-0f1eb6a5cd17-etc-machine-id\") pod \"cinder-db-sync-ldn9v\" (UID: \"2382b80c-75fc-435f-be5f-0f1eb6a5cd17\") " pod="openstack/cinder-db-sync-ldn9v" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.613743 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-ccd5f54d9-dsnj5" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.624489 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2382b80c-75fc-435f-be5f-0f1eb6a5cd17-scripts\") pod \"cinder-db-sync-ldn9v\" (UID: \"2382b80c-75fc-435f-be5f-0f1eb6a5cd17\") " pod="openstack/cinder-db-sync-ldn9v" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.632751 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2382b80c-75fc-435f-be5f-0f1eb6a5cd17-combined-ca-bundle\") pod \"cinder-db-sync-ldn9v\" (UID: \"2382b80c-75fc-435f-be5f-0f1eb6a5cd17\") " pod="openstack/cinder-db-sync-ldn9v" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.632832 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-ccd5f54d9-dsnj5"] Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.633237 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-zcq55" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.645206 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/2382b80c-75fc-435f-be5f-0f1eb6a5cd17-db-sync-config-data\") pod \"cinder-db-sync-ldn9v\" (UID: \"2382b80c-75fc-435f-be5f-0f1eb6a5cd17\") " pod="openstack/cinder-db-sync-ldn9v" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.646600 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fqk7s\" (UniqueName: \"kubernetes.io/projected/2382b80c-75fc-435f-be5f-0f1eb6a5cd17-kube-api-access-fqk7s\") pod \"cinder-db-sync-ldn9v\" (UID: \"2382b80c-75fc-435f-be5f-0f1eb6a5cd17\") " pod="openstack/cinder-db-sync-ldn9v" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.654123 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-fzvv8"] Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.655239 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-fzvv8" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.657802 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2382b80c-75fc-435f-be5f-0f1eb6a5cd17-config-data\") pod \"cinder-db-sync-ldn9v\" (UID: \"2382b80c-75fc-435f-be5f-0f1eb6a5cd17\") " pod="openstack/cinder-db-sync-ldn9v" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.668554 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-j6t79" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.669413 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.669597 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.693535 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c5cc7c5ff-w7tn4"] Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.710331 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/07c13b84-6c2a-4a6c-b23b-df1e14e077b5-horizon-secret-key\") pod \"horizon-7749656757-6jdvp\" (UID: \"07c13b84-6c2a-4a6c-b23b-df1e14e077b5\") " pod="openstack/horizon-7749656757-6jdvp" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.710403 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/13781991-323c-49ae-9c5d-625e2ffd019b-log-httpd\") pod \"ceilometer-0\" (UID: \"13781991-323c-49ae-9c5d-625e2ffd019b\") " pod="openstack/ceilometer-0" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.710439 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vcs8n\" (UniqueName: \"kubernetes.io/projected/dfc3bcd2-9baa-4a3a-8588-295e692d8e3e-kube-api-access-vcs8n\") pod \"barbican-db-sync-ljqzg\" (UID: \"dfc3bcd2-9baa-4a3a-8588-295e692d8e3e\") " pod="openstack/barbican-db-sync-ljqzg" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.710460 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/13781991-323c-49ae-9c5d-625e2ffd019b-config-data\") pod \"ceilometer-0\" (UID: \"13781991-323c-49ae-9c5d-625e2ffd019b\") " pod="openstack/ceilometer-0" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.710482 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/13781991-323c-49ae-9c5d-625e2ffd019b-scripts\") pod \"ceilometer-0\" (UID: \"13781991-323c-49ae-9c5d-625e2ffd019b\") " pod="openstack/ceilometer-0" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.710511 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/07c13b84-6c2a-4a6c-b23b-df1e14e077b5-logs\") pod \"horizon-7749656757-6jdvp\" (UID: \"07c13b84-6c2a-4a6c-b23b-df1e14e077b5\") " pod="openstack/horizon-7749656757-6jdvp" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.710528 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/13781991-323c-49ae-9c5d-625e2ffd019b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"13781991-323c-49ae-9c5d-625e2ffd019b\") " pod="openstack/ceilometer-0" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.710560 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/dfc3bcd2-9baa-4a3a-8588-295e692d8e3e-db-sync-config-data\") pod \"barbican-db-sync-ljqzg\" (UID: \"dfc3bcd2-9baa-4a3a-8588-295e692d8e3e\") " pod="openstack/barbican-db-sync-ljqzg" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.710585 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/07c13b84-6c2a-4a6c-b23b-df1e14e077b5-config-data\") pod \"horizon-7749656757-6jdvp\" (UID: \"07c13b84-6c2a-4a6c-b23b-df1e14e077b5\") " pod="openstack/horizon-7749656757-6jdvp" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.710612 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/07c13b84-6c2a-4a6c-b23b-df1e14e077b5-scripts\") pod \"horizon-7749656757-6jdvp\" (UID: \"07c13b84-6c2a-4a6c-b23b-df1e14e077b5\") " pod="openstack/horizon-7749656757-6jdvp" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.710634 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/13781991-323c-49ae-9c5d-625e2ffd019b-run-httpd\") pod \"ceilometer-0\" (UID: \"13781991-323c-49ae-9c5d-625e2ffd019b\") " pod="openstack/ceilometer-0" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.710658 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vt2nw\" (UniqueName: \"kubernetes.io/projected/07c13b84-6c2a-4a6c-b23b-df1e14e077b5-kube-api-access-vt2nw\") pod \"horizon-7749656757-6jdvp\" (UID: \"07c13b84-6c2a-4a6c-b23b-df1e14e077b5\") " pod="openstack/horizon-7749656757-6jdvp" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.710682 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dfc3bcd2-9baa-4a3a-8588-295e692d8e3e-combined-ca-bundle\") pod \"barbican-db-sync-ljqzg\" (UID: \"dfc3bcd2-9baa-4a3a-8588-295e692d8e3e\") " pod="openstack/barbican-db-sync-ljqzg" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.710702 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w6rqx\" (UniqueName: \"kubernetes.io/projected/13781991-323c-49ae-9c5d-625e2ffd019b-kube-api-access-w6rqx\") pod \"ceilometer-0\" (UID: \"13781991-323c-49ae-9c5d-625e2ffd019b\") " pod="openstack/ceilometer-0" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.710746 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13781991-323c-49ae-9c5d-625e2ffd019b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"13781991-323c-49ae-9c5d-625e2ffd019b\") " pod="openstack/ceilometer-0" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.711369 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/13781991-323c-49ae-9c5d-625e2ffd019b-log-httpd\") pod \"ceilometer-0\" (UID: \"13781991-323c-49ae-9c5d-625e2ffd019b\") " pod="openstack/ceilometer-0" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.713003 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/07c13b84-6c2a-4a6c-b23b-df1e14e077b5-config-data\") pod \"horizon-7749656757-6jdvp\" (UID: \"07c13b84-6c2a-4a6c-b23b-df1e14e077b5\") " pod="openstack/horizon-7749656757-6jdvp" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.719768 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/07c13b84-6c2a-4a6c-b23b-df1e14e077b5-scripts\") pod \"horizon-7749656757-6jdvp\" (UID: \"07c13b84-6c2a-4a6c-b23b-df1e14e077b5\") " pod="openstack/horizon-7749656757-6jdvp" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.723573 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/13781991-323c-49ae-9c5d-625e2ffd019b-run-httpd\") pod \"ceilometer-0\" (UID: \"13781991-323c-49ae-9c5d-625e2ffd019b\") " pod="openstack/ceilometer-0" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.726418 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/07c13b84-6c2a-4a6c-b23b-df1e14e077b5-logs\") pod \"horizon-7749656757-6jdvp\" (UID: \"07c13b84-6c2a-4a6c-b23b-df1e14e077b5\") " pod="openstack/horizon-7749656757-6jdvp" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.755806 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/07c13b84-6c2a-4a6c-b23b-df1e14e077b5-horizon-secret-key\") pod \"horizon-7749656757-6jdvp\" (UID: \"07c13b84-6c2a-4a6c-b23b-df1e14e077b5\") " pod="openstack/horizon-7749656757-6jdvp" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.755846 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/13781991-323c-49ae-9c5d-625e2ffd019b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"13781991-323c-49ae-9c5d-625e2ffd019b\") " pod="openstack/ceilometer-0" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.757011 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/13781991-323c-49ae-9c5d-625e2ffd019b-config-data\") pod \"ceilometer-0\" (UID: \"13781991-323c-49ae-9c5d-625e2ffd019b\") " pod="openstack/ceilometer-0" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.757736 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/13781991-323c-49ae-9c5d-625e2ffd019b-scripts\") pod \"ceilometer-0\" (UID: \"13781991-323c-49ae-9c5d-625e2ffd019b\") " pod="openstack/ceilometer-0" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.768999 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13781991-323c-49ae-9c5d-625e2ffd019b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"13781991-323c-49ae-9c5d-625e2ffd019b\") " pod="openstack/ceilometer-0" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.778777 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vt2nw\" (UniqueName: \"kubernetes.io/projected/07c13b84-6c2a-4a6c-b23b-df1e14e077b5-kube-api-access-vt2nw\") pod \"horizon-7749656757-6jdvp\" (UID: \"07c13b84-6c2a-4a6c-b23b-df1e14e077b5\") " pod="openstack/horizon-7749656757-6jdvp" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.780353 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w6rqx\" (UniqueName: \"kubernetes.io/projected/13781991-323c-49ae-9c5d-625e2ffd019b-kube-api-access-w6rqx\") pod \"ceilometer-0\" (UID: \"13781991-323c-49ae-9c5d-625e2ffd019b\") " pod="openstack/ceilometer-0" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.781292 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-ldn9v" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.853711 4824 generic.go:334] "Generic (PLEG): container finished" podID="67457ff8-d89b-4f2b-8f49-b6323746b4b4" containerID="f172c000d0b1e8b0c4db204fddd4668a7710b99c62386eff0a34b706c5d642c1" exitCode=0 Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.853774 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7ff5475cc9-9l788" event={"ID":"67457ff8-d89b-4f2b-8f49-b6323746b4b4","Type":"ContainerDied","Data":"f172c000d0b1e8b0c4db204fddd4668a7710b99c62386eff0a34b706c5d642c1"} Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.897185 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-prrmr\" (UniqueName: \"kubernetes.io/projected/3d28a180-78e5-4977-bccb-9210f5fb0a7f-kube-api-access-prrmr\") pod \"horizon-ccd5f54d9-dsnj5\" (UID: \"3d28a180-78e5-4977-bccb-9210f5fb0a7f\") " pod="openstack/horizon-ccd5f54d9-dsnj5" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.897689 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vcs8n\" (UniqueName: \"kubernetes.io/projected/dfc3bcd2-9baa-4a3a-8588-295e692d8e3e-kube-api-access-vcs8n\") pod \"barbican-db-sync-ljqzg\" (UID: \"dfc3bcd2-9baa-4a3a-8588-295e692d8e3e\") " pod="openstack/barbican-db-sync-ljqzg" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.897739 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zhrg9\" (UniqueName: \"kubernetes.io/projected/554a3f37-c7c8-4d09-9da7-df5319dccecd-kube-api-access-zhrg9\") pod \"placement-db-sync-fzvv8\" (UID: \"554a3f37-c7c8-4d09-9da7-df5319dccecd\") " pod="openstack/placement-db-sync-fzvv8" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.897838 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/dfc3bcd2-9baa-4a3a-8588-295e692d8e3e-db-sync-config-data\") pod \"barbican-db-sync-ljqzg\" (UID: \"dfc3bcd2-9baa-4a3a-8588-295e692d8e3e\") " pod="openstack/barbican-db-sync-ljqzg" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.897902 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3d28a180-78e5-4977-bccb-9210f5fb0a7f-logs\") pod \"horizon-ccd5f54d9-dsnj5\" (UID: \"3d28a180-78e5-4977-bccb-9210f5fb0a7f\") " pod="openstack/horizon-ccd5f54d9-dsnj5" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.897952 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/3d28a180-78e5-4977-bccb-9210f5fb0a7f-horizon-secret-key\") pod \"horizon-ccd5f54d9-dsnj5\" (UID: \"3d28a180-78e5-4977-bccb-9210f5fb0a7f\") " pod="openstack/horizon-ccd5f54d9-dsnj5" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.898003 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/554a3f37-c7c8-4d09-9da7-df5319dccecd-logs\") pod \"placement-db-sync-fzvv8\" (UID: \"554a3f37-c7c8-4d09-9da7-df5319dccecd\") " pod="openstack/placement-db-sync-fzvv8" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.898092 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/554a3f37-c7c8-4d09-9da7-df5319dccecd-config-data\") pod \"placement-db-sync-fzvv8\" (UID: \"554a3f37-c7c8-4d09-9da7-df5319dccecd\") " pod="openstack/placement-db-sync-fzvv8" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.898129 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dfc3bcd2-9baa-4a3a-8588-295e692d8e3e-combined-ca-bundle\") pod \"barbican-db-sync-ljqzg\" (UID: \"dfc3bcd2-9baa-4a3a-8588-295e692d8e3e\") " pod="openstack/barbican-db-sync-ljqzg" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.898162 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/554a3f37-c7c8-4d09-9da7-df5319dccecd-combined-ca-bundle\") pod \"placement-db-sync-fzvv8\" (UID: \"554a3f37-c7c8-4d09-9da7-df5319dccecd\") " pod="openstack/placement-db-sync-fzvv8" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.898231 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/554a3f37-c7c8-4d09-9da7-df5319dccecd-scripts\") pod \"placement-db-sync-fzvv8\" (UID: \"554a3f37-c7c8-4d09-9da7-df5319dccecd\") " pod="openstack/placement-db-sync-fzvv8" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.898311 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3d28a180-78e5-4977-bccb-9210f5fb0a7f-config-data\") pod \"horizon-ccd5f54d9-dsnj5\" (UID: \"3d28a180-78e5-4977-bccb-9210f5fb0a7f\") " pod="openstack/horizon-ccd5f54d9-dsnj5" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.898419 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3d28a180-78e5-4977-bccb-9210f5fb0a7f-scripts\") pod \"horizon-ccd5f54d9-dsnj5\" (UID: \"3d28a180-78e5-4977-bccb-9210f5fb0a7f\") " pod="openstack/horizon-ccd5f54d9-dsnj5" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.898910 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7749656757-6jdvp" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.907486 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.939430 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dfc3bcd2-9baa-4a3a-8588-295e692d8e3e-combined-ca-bundle\") pod \"barbican-db-sync-ljqzg\" (UID: \"dfc3bcd2-9baa-4a3a-8588-295e692d8e3e\") " pod="openstack/barbican-db-sync-ljqzg" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.940782 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/dfc3bcd2-9baa-4a3a-8588-295e692d8e3e-db-sync-config-data\") pod \"barbican-db-sync-ljqzg\" (UID: \"dfc3bcd2-9baa-4a3a-8588-295e692d8e3e\") " pod="openstack/barbican-db-sync-ljqzg" Oct 06 10:15:46 crc kubenswrapper[4824]: I1006 10:15:46.995617 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-fzvv8"] Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.000945 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vcs8n\" (UniqueName: \"kubernetes.io/projected/dfc3bcd2-9baa-4a3a-8588-295e692d8e3e-kube-api-access-vcs8n\") pod \"barbican-db-sync-ljqzg\" (UID: \"dfc3bcd2-9baa-4a3a-8588-295e692d8e3e\") " pod="openstack/barbican-db-sync-ljqzg" Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.011917 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3d28a180-78e5-4977-bccb-9210f5fb0a7f-config-data\") pod \"horizon-ccd5f54d9-dsnj5\" (UID: \"3d28a180-78e5-4977-bccb-9210f5fb0a7f\") " pod="openstack/horizon-ccd5f54d9-dsnj5" Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.015142 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3d28a180-78e5-4977-bccb-9210f5fb0a7f-scripts\") pod \"horizon-ccd5f54d9-dsnj5\" (UID: \"3d28a180-78e5-4977-bccb-9210f5fb0a7f\") " pod="openstack/horizon-ccd5f54d9-dsnj5" Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.018140 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-prrmr\" (UniqueName: \"kubernetes.io/projected/3d28a180-78e5-4977-bccb-9210f5fb0a7f-kube-api-access-prrmr\") pod \"horizon-ccd5f54d9-dsnj5\" (UID: \"3d28a180-78e5-4977-bccb-9210f5fb0a7f\") " pod="openstack/horizon-ccd5f54d9-dsnj5" Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.022223 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zhrg9\" (UniqueName: \"kubernetes.io/projected/554a3f37-c7c8-4d09-9da7-df5319dccecd-kube-api-access-zhrg9\") pod \"placement-db-sync-fzvv8\" (UID: \"554a3f37-c7c8-4d09-9da7-df5319dccecd\") " pod="openstack/placement-db-sync-fzvv8" Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.022395 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3d28a180-78e5-4977-bccb-9210f5fb0a7f-logs\") pod \"horizon-ccd5f54d9-dsnj5\" (UID: \"3d28a180-78e5-4977-bccb-9210f5fb0a7f\") " pod="openstack/horizon-ccd5f54d9-dsnj5" Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.022479 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/3d28a180-78e5-4977-bccb-9210f5fb0a7f-horizon-secret-key\") pod \"horizon-ccd5f54d9-dsnj5\" (UID: \"3d28a180-78e5-4977-bccb-9210f5fb0a7f\") " pod="openstack/horizon-ccd5f54d9-dsnj5" Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.022567 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/554a3f37-c7c8-4d09-9da7-df5319dccecd-logs\") pod \"placement-db-sync-fzvv8\" (UID: \"554a3f37-c7c8-4d09-9da7-df5319dccecd\") " pod="openstack/placement-db-sync-fzvv8" Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.022672 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/554a3f37-c7c8-4d09-9da7-df5319dccecd-config-data\") pod \"placement-db-sync-fzvv8\" (UID: \"554a3f37-c7c8-4d09-9da7-df5319dccecd\") " pod="openstack/placement-db-sync-fzvv8" Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.022744 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/554a3f37-c7c8-4d09-9da7-df5319dccecd-combined-ca-bundle\") pod \"placement-db-sync-fzvv8\" (UID: \"554a3f37-c7c8-4d09-9da7-df5319dccecd\") " pod="openstack/placement-db-sync-fzvv8" Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.022832 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/554a3f37-c7c8-4d09-9da7-df5319dccecd-scripts\") pod \"placement-db-sync-fzvv8\" (UID: \"554a3f37-c7c8-4d09-9da7-df5319dccecd\") " pod="openstack/placement-db-sync-fzvv8" Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.023656 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3d28a180-78e5-4977-bccb-9210f5fb0a7f-logs\") pod \"horizon-ccd5f54d9-dsnj5\" (UID: \"3d28a180-78e5-4977-bccb-9210f5fb0a7f\") " pod="openstack/horizon-ccd5f54d9-dsnj5" Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.024220 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/554a3f37-c7c8-4d09-9da7-df5319dccecd-logs\") pod \"placement-db-sync-fzvv8\" (UID: \"554a3f37-c7c8-4d09-9da7-df5319dccecd\") " pod="openstack/placement-db-sync-fzvv8" Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.017661 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3d28a180-78e5-4977-bccb-9210f5fb0a7f-config-data\") pod \"horizon-ccd5f54d9-dsnj5\" (UID: \"3d28a180-78e5-4977-bccb-9210f5fb0a7f\") " pod="openstack/horizon-ccd5f54d9-dsnj5" Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.018077 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3d28a180-78e5-4977-bccb-9210f5fb0a7f-scripts\") pod \"horizon-ccd5f54d9-dsnj5\" (UID: \"3d28a180-78e5-4977-bccb-9210f5fb0a7f\") " pod="openstack/horizon-ccd5f54d9-dsnj5" Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.038532 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-xsgsg"] Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.039808 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-xsgsg" Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.042684 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/554a3f37-c7c8-4d09-9da7-df5319dccecd-scripts\") pod \"placement-db-sync-fzvv8\" (UID: \"554a3f37-c7c8-4d09-9da7-df5319dccecd\") " pod="openstack/placement-db-sync-fzvv8" Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.043800 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/3d28a180-78e5-4977-bccb-9210f5fb0a7f-horizon-secret-key\") pod \"horizon-ccd5f54d9-dsnj5\" (UID: \"3d28a180-78e5-4977-bccb-9210f5fb0a7f\") " pod="openstack/horizon-ccd5f54d9-dsnj5" Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.047226 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/554a3f37-c7c8-4d09-9da7-df5319dccecd-combined-ca-bundle\") pod \"placement-db-sync-fzvv8\" (UID: \"554a3f37-c7c8-4d09-9da7-df5319dccecd\") " pod="openstack/placement-db-sync-fzvv8" Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.048111 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.049336 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/554a3f37-c7c8-4d09-9da7-df5319dccecd-config-data\") pod \"placement-db-sync-fzvv8\" (UID: \"554a3f37-c7c8-4d09-9da7-df5319dccecd\") " pod="openstack/placement-db-sync-fzvv8" Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.060453 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.069738 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-prrmr\" (UniqueName: \"kubernetes.io/projected/3d28a180-78e5-4977-bccb-9210f5fb0a7f-kube-api-access-prrmr\") pod \"horizon-ccd5f54d9-dsnj5\" (UID: \"3d28a180-78e5-4977-bccb-9210f5fb0a7f\") " pod="openstack/horizon-ccd5f54d9-dsnj5" Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.070214 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-wx2zn" Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.087613 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zhrg9\" (UniqueName: \"kubernetes.io/projected/554a3f37-c7c8-4d09-9da7-df5319dccecd-kube-api-access-zhrg9\") pod \"placement-db-sync-fzvv8\" (UID: \"554a3f37-c7c8-4d09-9da7-df5319dccecd\") " pod="openstack/placement-db-sync-fzvv8" Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.095883 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-fzvv8" Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.127797 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/3d78bdee-7130-48a9-ad24-6c8c6525784c-config\") pod \"neutron-db-sync-xsgsg\" (UID: \"3d78bdee-7130-48a9-ad24-6c8c6525784c\") " pod="openstack/neutron-db-sync-xsgsg" Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.128338 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d78bdee-7130-48a9-ad24-6c8c6525784c-combined-ca-bundle\") pod \"neutron-db-sync-xsgsg\" (UID: \"3d78bdee-7130-48a9-ad24-6c8c6525784c\") " pod="openstack/neutron-db-sync-xsgsg" Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.128449 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8h584\" (UniqueName: \"kubernetes.io/projected/3d78bdee-7130-48a9-ad24-6c8c6525784c-kube-api-access-8h584\") pod \"neutron-db-sync-xsgsg\" (UID: \"3d78bdee-7130-48a9-ad24-6c8c6525784c\") " pod="openstack/neutron-db-sync-xsgsg" Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.161411 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-ljqzg" Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.188292 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-xsgsg"] Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.230363 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/3d78bdee-7130-48a9-ad24-6c8c6525784c-config\") pod \"neutron-db-sync-xsgsg\" (UID: \"3d78bdee-7130-48a9-ad24-6c8c6525784c\") " pod="openstack/neutron-db-sync-xsgsg" Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.230459 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d78bdee-7130-48a9-ad24-6c8c6525784c-combined-ca-bundle\") pod \"neutron-db-sync-xsgsg\" (UID: \"3d78bdee-7130-48a9-ad24-6c8c6525784c\") " pod="openstack/neutron-db-sync-xsgsg" Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.230496 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8h584\" (UniqueName: \"kubernetes.io/projected/3d78bdee-7130-48a9-ad24-6c8c6525784c-kube-api-access-8h584\") pod \"neutron-db-sync-xsgsg\" (UID: \"3d78bdee-7130-48a9-ad24-6c8c6525784c\") " pod="openstack/neutron-db-sync-xsgsg" Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.251713 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d78bdee-7130-48a9-ad24-6c8c6525784c-combined-ca-bundle\") pod \"neutron-db-sync-xsgsg\" (UID: \"3d78bdee-7130-48a9-ad24-6c8c6525784c\") " pod="openstack/neutron-db-sync-xsgsg" Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.272792 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8h584\" (UniqueName: \"kubernetes.io/projected/3d78bdee-7130-48a9-ad24-6c8c6525784c-kube-api-access-8h584\") pod \"neutron-db-sync-xsgsg\" (UID: \"3d78bdee-7130-48a9-ad24-6c8c6525784c\") " pod="openstack/neutron-db-sync-xsgsg" Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.280587 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7ff5475cc9-9l788" Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.281481 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/3d78bdee-7130-48a9-ad24-6c8c6525784c-config\") pod \"neutron-db-sync-xsgsg\" (UID: \"3d78bdee-7130-48a9-ad24-6c8c6525784c\") " pod="openstack/neutron-db-sync-xsgsg" Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.321369 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-ccd5f54d9-dsnj5" Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.331642 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/67457ff8-d89b-4f2b-8f49-b6323746b4b4-ovsdbserver-sb\") pod \"67457ff8-d89b-4f2b-8f49-b6323746b4b4\" (UID: \"67457ff8-d89b-4f2b-8f49-b6323746b4b4\") " Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.331707 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/67457ff8-d89b-4f2b-8f49-b6323746b4b4-config\") pod \"67457ff8-d89b-4f2b-8f49-b6323746b4b4\" (UID: \"67457ff8-d89b-4f2b-8f49-b6323746b4b4\") " Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.331762 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/67457ff8-d89b-4f2b-8f49-b6323746b4b4-ovsdbserver-nb\") pod \"67457ff8-d89b-4f2b-8f49-b6323746b4b4\" (UID: \"67457ff8-d89b-4f2b-8f49-b6323746b4b4\") " Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.332008 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/67457ff8-d89b-4f2b-8f49-b6323746b4b4-dns-swift-storage-0\") pod \"67457ff8-d89b-4f2b-8f49-b6323746b4b4\" (UID: \"67457ff8-d89b-4f2b-8f49-b6323746b4b4\") " Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.332044 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h9jx5\" (UniqueName: \"kubernetes.io/projected/67457ff8-d89b-4f2b-8f49-b6323746b4b4-kube-api-access-h9jx5\") pod \"67457ff8-d89b-4f2b-8f49-b6323746b4b4\" (UID: \"67457ff8-d89b-4f2b-8f49-b6323746b4b4\") " Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.332075 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/67457ff8-d89b-4f2b-8f49-b6323746b4b4-dns-svc\") pod \"67457ff8-d89b-4f2b-8f49-b6323746b4b4\" (UID: \"67457ff8-d89b-4f2b-8f49-b6323746b4b4\") " Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.363966 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/67457ff8-d89b-4f2b-8f49-b6323746b4b4-kube-api-access-h9jx5" (OuterVolumeSpecName: "kube-api-access-h9jx5") pod "67457ff8-d89b-4f2b-8f49-b6323746b4b4" (UID: "67457ff8-d89b-4f2b-8f49-b6323746b4b4"). InnerVolumeSpecName "kube-api-access-h9jx5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.400308 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-xsgsg" Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.434700 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h9jx5\" (UniqueName: \"kubernetes.io/projected/67457ff8-d89b-4f2b-8f49-b6323746b4b4-kube-api-access-h9jx5\") on node \"crc\" DevicePath \"\"" Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.495274 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/67457ff8-d89b-4f2b-8f49-b6323746b4b4-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "67457ff8-d89b-4f2b-8f49-b6323746b4b4" (UID: "67457ff8-d89b-4f2b-8f49-b6323746b4b4"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.507254 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-9r9wx"] Oct 06 10:15:47 crc kubenswrapper[4824]: E1006 10:15:47.507904 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67457ff8-d89b-4f2b-8f49-b6323746b4b4" containerName="init" Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.507928 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="67457ff8-d89b-4f2b-8f49-b6323746b4b4" containerName="init" Oct 06 10:15:47 crc kubenswrapper[4824]: E1006 10:15:47.507952 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67457ff8-d89b-4f2b-8f49-b6323746b4b4" containerName="dnsmasq-dns" Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.507960 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="67457ff8-d89b-4f2b-8f49-b6323746b4b4" containerName="dnsmasq-dns" Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.508178 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="67457ff8-d89b-4f2b-8f49-b6323746b4b4" containerName="dnsmasq-dns" Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.509393 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-9r9wx"] Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.509418 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.511777 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8b5c85b87-9r9wx" Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.516058 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.516158 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.525193 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.525366 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.525469 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-ctsvg" Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.526203 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.531103 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/67457ff8-d89b-4f2b-8f49-b6323746b4b4-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "67457ff8-d89b-4f2b-8f49-b6323746b4b4" (UID: "67457ff8-d89b-4f2b-8f49-b6323746b4b4"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.536547 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/629a3cfa-5d3a-416a-9316-8d64594f34f1-ovsdbserver-sb\") pod \"dnsmasq-dns-8b5c85b87-9r9wx\" (UID: \"629a3cfa-5d3a-416a-9316-8d64594f34f1\") " pod="openstack/dnsmasq-dns-8b5c85b87-9r9wx" Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.536596 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/629a3cfa-5d3a-416a-9316-8d64594f34f1-dns-swift-storage-0\") pod \"dnsmasq-dns-8b5c85b87-9r9wx\" (UID: \"629a3cfa-5d3a-416a-9316-8d64594f34f1\") " pod="openstack/dnsmasq-dns-8b5c85b87-9r9wx" Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.536622 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dk7h7\" (UniqueName: \"kubernetes.io/projected/629a3cfa-5d3a-416a-9316-8d64594f34f1-kube-api-access-dk7h7\") pod \"dnsmasq-dns-8b5c85b87-9r9wx\" (UID: \"629a3cfa-5d3a-416a-9316-8d64594f34f1\") " pod="openstack/dnsmasq-dns-8b5c85b87-9r9wx" Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.536691 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/629a3cfa-5d3a-416a-9316-8d64594f34f1-config\") pod \"dnsmasq-dns-8b5c85b87-9r9wx\" (UID: \"629a3cfa-5d3a-416a-9316-8d64594f34f1\") " pod="openstack/dnsmasq-dns-8b5c85b87-9r9wx" Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.536786 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/629a3cfa-5d3a-416a-9316-8d64594f34f1-dns-svc\") pod \"dnsmasq-dns-8b5c85b87-9r9wx\" (UID: \"629a3cfa-5d3a-416a-9316-8d64594f34f1\") " pod="openstack/dnsmasq-dns-8b5c85b87-9r9wx" Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.536840 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/629a3cfa-5d3a-416a-9316-8d64594f34f1-ovsdbserver-nb\") pod \"dnsmasq-dns-8b5c85b87-9r9wx\" (UID: \"629a3cfa-5d3a-416a-9316-8d64594f34f1\") " pod="openstack/dnsmasq-dns-8b5c85b87-9r9wx" Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.536890 4824 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/67457ff8-d89b-4f2b-8f49-b6323746b4b4-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.536901 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/67457ff8-d89b-4f2b-8f49-b6323746b4b4-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.539482 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.548700 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.549770 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/67457ff8-d89b-4f2b-8f49-b6323746b4b4-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "67457ff8-d89b-4f2b-8f49-b6323746b4b4" (UID: "67457ff8-d89b-4f2b-8f49-b6323746b4b4"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.559775 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/67457ff8-d89b-4f2b-8f49-b6323746b4b4-config" (OuterVolumeSpecName: "config") pod "67457ff8-d89b-4f2b-8f49-b6323746b4b4" (UID: "67457ff8-d89b-4f2b-8f49-b6323746b4b4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.559852 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.560180 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.561872 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/67457ff8-d89b-4f2b-8f49-b6323746b4b4-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "67457ff8-d89b-4f2b-8f49-b6323746b4b4" (UID: "67457ff8-d89b-4f2b-8f49-b6323746b4b4"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.575211 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.638655 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/629a3cfa-5d3a-416a-9316-8d64594f34f1-dns-svc\") pod \"dnsmasq-dns-8b5c85b87-9r9wx\" (UID: \"629a3cfa-5d3a-416a-9316-8d64594f34f1\") " pod="openstack/dnsmasq-dns-8b5c85b87-9r9wx" Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.638923 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e26a57e-6683-48b2-a0ea-3f8ff1027fdc-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"2e26a57e-6683-48b2-a0ea-3f8ff1027fdc\") " pod="openstack/glance-default-external-api-0" Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.638945 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e26a57e-6683-48b2-a0ea-3f8ff1027fdc-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"2e26a57e-6683-48b2-a0ea-3f8ff1027fdc\") " pod="openstack/glance-default-external-api-0" Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.638994 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e26a57e-6683-48b2-a0ea-3f8ff1027fdc-config-data\") pod \"glance-default-external-api-0\" (UID: \"2e26a57e-6683-48b2-a0ea-3f8ff1027fdc\") " pod="openstack/glance-default-external-api-0" Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.639017 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/629a3cfa-5d3a-416a-9316-8d64594f34f1-ovsdbserver-nb\") pod \"dnsmasq-dns-8b5c85b87-9r9wx\" (UID: \"629a3cfa-5d3a-416a-9316-8d64594f34f1\") " pod="openstack/dnsmasq-dns-8b5c85b87-9r9wx" Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.639051 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/629a3cfa-5d3a-416a-9316-8d64594f34f1-ovsdbserver-sb\") pod \"dnsmasq-dns-8b5c85b87-9r9wx\" (UID: \"629a3cfa-5d3a-416a-9316-8d64594f34f1\") " pod="openstack/dnsmasq-dns-8b5c85b87-9r9wx" Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.639086 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/629a3cfa-5d3a-416a-9316-8d64594f34f1-dns-swift-storage-0\") pod \"dnsmasq-dns-8b5c85b87-9r9wx\" (UID: \"629a3cfa-5d3a-416a-9316-8d64594f34f1\") " pod="openstack/dnsmasq-dns-8b5c85b87-9r9wx" Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.639109 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dk7h7\" (UniqueName: \"kubernetes.io/projected/629a3cfa-5d3a-416a-9316-8d64594f34f1-kube-api-access-dk7h7\") pod \"dnsmasq-dns-8b5c85b87-9r9wx\" (UID: \"629a3cfa-5d3a-416a-9316-8d64594f34f1\") " pod="openstack/dnsmasq-dns-8b5c85b87-9r9wx" Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.639133 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2e26a57e-6683-48b2-a0ea-3f8ff1027fdc-scripts\") pod \"glance-default-external-api-0\" (UID: \"2e26a57e-6683-48b2-a0ea-3f8ff1027fdc\") " pod="openstack/glance-default-external-api-0" Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.639166 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"2e26a57e-6683-48b2-a0ea-3f8ff1027fdc\") " pod="openstack/glance-default-external-api-0" Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.639190 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/629a3cfa-5d3a-416a-9316-8d64594f34f1-config\") pod \"dnsmasq-dns-8b5c85b87-9r9wx\" (UID: \"629a3cfa-5d3a-416a-9316-8d64594f34f1\") " pod="openstack/dnsmasq-dns-8b5c85b87-9r9wx" Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.639213 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2e26a57e-6683-48b2-a0ea-3f8ff1027fdc-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"2e26a57e-6683-48b2-a0ea-3f8ff1027fdc\") " pod="openstack/glance-default-external-api-0" Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.639246 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nlv4q\" (UniqueName: \"kubernetes.io/projected/2e26a57e-6683-48b2-a0ea-3f8ff1027fdc-kube-api-access-nlv4q\") pod \"glance-default-external-api-0\" (UID: \"2e26a57e-6683-48b2-a0ea-3f8ff1027fdc\") " pod="openstack/glance-default-external-api-0" Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.639291 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2e26a57e-6683-48b2-a0ea-3f8ff1027fdc-logs\") pod \"glance-default-external-api-0\" (UID: \"2e26a57e-6683-48b2-a0ea-3f8ff1027fdc\") " pod="openstack/glance-default-external-api-0" Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.639357 4824 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/67457ff8-d89b-4f2b-8f49-b6323746b4b4-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.639369 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/67457ff8-d89b-4f2b-8f49-b6323746b4b4-config\") on node \"crc\" DevicePath \"\"" Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.639379 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/67457ff8-d89b-4f2b-8f49-b6323746b4b4-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.640461 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/629a3cfa-5d3a-416a-9316-8d64594f34f1-dns-svc\") pod \"dnsmasq-dns-8b5c85b87-9r9wx\" (UID: \"629a3cfa-5d3a-416a-9316-8d64594f34f1\") " pod="openstack/dnsmasq-dns-8b5c85b87-9r9wx" Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.641547 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/629a3cfa-5d3a-416a-9316-8d64594f34f1-ovsdbserver-nb\") pod \"dnsmasq-dns-8b5c85b87-9r9wx\" (UID: \"629a3cfa-5d3a-416a-9316-8d64594f34f1\") " pod="openstack/dnsmasq-dns-8b5c85b87-9r9wx" Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.642327 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/629a3cfa-5d3a-416a-9316-8d64594f34f1-ovsdbserver-sb\") pod \"dnsmasq-dns-8b5c85b87-9r9wx\" (UID: \"629a3cfa-5d3a-416a-9316-8d64594f34f1\") " pod="openstack/dnsmasq-dns-8b5c85b87-9r9wx" Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.642852 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/629a3cfa-5d3a-416a-9316-8d64594f34f1-dns-swift-storage-0\") pod \"dnsmasq-dns-8b5c85b87-9r9wx\" (UID: \"629a3cfa-5d3a-416a-9316-8d64594f34f1\") " pod="openstack/dnsmasq-dns-8b5c85b87-9r9wx" Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.643844 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/629a3cfa-5d3a-416a-9316-8d64594f34f1-config\") pod \"dnsmasq-dns-8b5c85b87-9r9wx\" (UID: \"629a3cfa-5d3a-416a-9316-8d64594f34f1\") " pod="openstack/dnsmasq-dns-8b5c85b87-9r9wx" Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.669045 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dk7h7\" (UniqueName: \"kubernetes.io/projected/629a3cfa-5d3a-416a-9316-8d64594f34f1-kube-api-access-dk7h7\") pod \"dnsmasq-dns-8b5c85b87-9r9wx\" (UID: \"629a3cfa-5d3a-416a-9316-8d64594f34f1\") " pod="openstack/dnsmasq-dns-8b5c85b87-9r9wx" Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.673314 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c5cc7c5ff-w7tn4"] Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.743557 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/97b2f7d6-3fd8-43ab-85b6-a3d37e218887-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"97b2f7d6-3fd8-43ab-85b6-a3d37e218887\") " pod="openstack/glance-default-internal-api-0" Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.744058 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e26a57e-6683-48b2-a0ea-3f8ff1027fdc-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"2e26a57e-6683-48b2-a0ea-3f8ff1027fdc\") " pod="openstack/glance-default-external-api-0" Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.744079 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e26a57e-6683-48b2-a0ea-3f8ff1027fdc-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"2e26a57e-6683-48b2-a0ea-3f8ff1027fdc\") " pod="openstack/glance-default-external-api-0" Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.744131 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e26a57e-6683-48b2-a0ea-3f8ff1027fdc-config-data\") pod \"glance-default-external-api-0\" (UID: \"2e26a57e-6683-48b2-a0ea-3f8ff1027fdc\") " pod="openstack/glance-default-external-api-0" Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.744236 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dh64c\" (UniqueName: \"kubernetes.io/projected/97b2f7d6-3fd8-43ab-85b6-a3d37e218887-kube-api-access-dh64c\") pod \"glance-default-internal-api-0\" (UID: \"97b2f7d6-3fd8-43ab-85b6-a3d37e218887\") " pod="openstack/glance-default-internal-api-0" Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.744279 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2e26a57e-6683-48b2-a0ea-3f8ff1027fdc-scripts\") pod \"glance-default-external-api-0\" (UID: \"2e26a57e-6683-48b2-a0ea-3f8ff1027fdc\") " pod="openstack/glance-default-external-api-0" Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.744307 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97b2f7d6-3fd8-43ab-85b6-a3d37e218887-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"97b2f7d6-3fd8-43ab-85b6-a3d37e218887\") " pod="openstack/glance-default-internal-api-0" Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.744335 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"2e26a57e-6683-48b2-a0ea-3f8ff1027fdc\") " pod="openstack/glance-default-external-api-0" Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.744354 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"97b2f7d6-3fd8-43ab-85b6-a3d37e218887\") " pod="openstack/glance-default-internal-api-0" Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.744392 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2e26a57e-6683-48b2-a0ea-3f8ff1027fdc-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"2e26a57e-6683-48b2-a0ea-3f8ff1027fdc\") " pod="openstack/glance-default-external-api-0" Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.744409 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/97b2f7d6-3fd8-43ab-85b6-a3d37e218887-config-data\") pod \"glance-default-internal-api-0\" (UID: \"97b2f7d6-3fd8-43ab-85b6-a3d37e218887\") " pod="openstack/glance-default-internal-api-0" Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.744439 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nlv4q\" (UniqueName: \"kubernetes.io/projected/2e26a57e-6683-48b2-a0ea-3f8ff1027fdc-kube-api-access-nlv4q\") pod \"glance-default-external-api-0\" (UID: \"2e26a57e-6683-48b2-a0ea-3f8ff1027fdc\") " pod="openstack/glance-default-external-api-0" Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.744474 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/97b2f7d6-3fd8-43ab-85b6-a3d37e218887-scripts\") pod \"glance-default-internal-api-0\" (UID: \"97b2f7d6-3fd8-43ab-85b6-a3d37e218887\") " pod="openstack/glance-default-internal-api-0" Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.744506 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/97b2f7d6-3fd8-43ab-85b6-a3d37e218887-logs\") pod \"glance-default-internal-api-0\" (UID: \"97b2f7d6-3fd8-43ab-85b6-a3d37e218887\") " pod="openstack/glance-default-internal-api-0" Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.744523 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2e26a57e-6683-48b2-a0ea-3f8ff1027fdc-logs\") pod \"glance-default-external-api-0\" (UID: \"2e26a57e-6683-48b2-a0ea-3f8ff1027fdc\") " pod="openstack/glance-default-external-api-0" Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.744542 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/97b2f7d6-3fd8-43ab-85b6-a3d37e218887-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"97b2f7d6-3fd8-43ab-85b6-a3d37e218887\") " pod="openstack/glance-default-internal-api-0" Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.748017 4824 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"2e26a57e-6683-48b2-a0ea-3f8ff1027fdc\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/glance-default-external-api-0" Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.748450 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2e26a57e-6683-48b2-a0ea-3f8ff1027fdc-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"2e26a57e-6683-48b2-a0ea-3f8ff1027fdc\") " pod="openstack/glance-default-external-api-0" Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.748764 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2e26a57e-6683-48b2-a0ea-3f8ff1027fdc-logs\") pod \"glance-default-external-api-0\" (UID: \"2e26a57e-6683-48b2-a0ea-3f8ff1027fdc\") " pod="openstack/glance-default-external-api-0" Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.753155 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2e26a57e-6683-48b2-a0ea-3f8ff1027fdc-scripts\") pod \"glance-default-external-api-0\" (UID: \"2e26a57e-6683-48b2-a0ea-3f8ff1027fdc\") " pod="openstack/glance-default-external-api-0" Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.765343 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-zcq55"] Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.765818 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e26a57e-6683-48b2-a0ea-3f8ff1027fdc-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"2e26a57e-6683-48b2-a0ea-3f8ff1027fdc\") " pod="openstack/glance-default-external-api-0" Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.770435 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e26a57e-6683-48b2-a0ea-3f8ff1027fdc-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"2e26a57e-6683-48b2-a0ea-3f8ff1027fdc\") " pod="openstack/glance-default-external-api-0" Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.771529 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e26a57e-6683-48b2-a0ea-3f8ff1027fdc-config-data\") pod \"glance-default-external-api-0\" (UID: \"2e26a57e-6683-48b2-a0ea-3f8ff1027fdc\") " pod="openstack/glance-default-external-api-0" Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.781618 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nlv4q\" (UniqueName: \"kubernetes.io/projected/2e26a57e-6683-48b2-a0ea-3f8ff1027fdc-kube-api-access-nlv4q\") pod \"glance-default-external-api-0\" (UID: \"2e26a57e-6683-48b2-a0ea-3f8ff1027fdc\") " pod="openstack/glance-default-external-api-0" Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.789781 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-ldn9v"] Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.797416 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"2e26a57e-6683-48b2-a0ea-3f8ff1027fdc\") " pod="openstack/glance-default-external-api-0" Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.846111 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dh64c\" (UniqueName: \"kubernetes.io/projected/97b2f7d6-3fd8-43ab-85b6-a3d37e218887-kube-api-access-dh64c\") pod \"glance-default-internal-api-0\" (UID: \"97b2f7d6-3fd8-43ab-85b6-a3d37e218887\") " pod="openstack/glance-default-internal-api-0" Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.846624 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97b2f7d6-3fd8-43ab-85b6-a3d37e218887-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"97b2f7d6-3fd8-43ab-85b6-a3d37e218887\") " pod="openstack/glance-default-internal-api-0" Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.846836 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"97b2f7d6-3fd8-43ab-85b6-a3d37e218887\") " pod="openstack/glance-default-internal-api-0" Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.847310 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/97b2f7d6-3fd8-43ab-85b6-a3d37e218887-config-data\") pod \"glance-default-internal-api-0\" (UID: \"97b2f7d6-3fd8-43ab-85b6-a3d37e218887\") " pod="openstack/glance-default-internal-api-0" Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.847369 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/97b2f7d6-3fd8-43ab-85b6-a3d37e218887-scripts\") pod \"glance-default-internal-api-0\" (UID: \"97b2f7d6-3fd8-43ab-85b6-a3d37e218887\") " pod="openstack/glance-default-internal-api-0" Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.847398 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/97b2f7d6-3fd8-43ab-85b6-a3d37e218887-logs\") pod \"glance-default-internal-api-0\" (UID: \"97b2f7d6-3fd8-43ab-85b6-a3d37e218887\") " pod="openstack/glance-default-internal-api-0" Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.847416 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/97b2f7d6-3fd8-43ab-85b6-a3d37e218887-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"97b2f7d6-3fd8-43ab-85b6-a3d37e218887\") " pod="openstack/glance-default-internal-api-0" Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.847474 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/97b2f7d6-3fd8-43ab-85b6-a3d37e218887-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"97b2f7d6-3fd8-43ab-85b6-a3d37e218887\") " pod="openstack/glance-default-internal-api-0" Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.848926 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/97b2f7d6-3fd8-43ab-85b6-a3d37e218887-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"97b2f7d6-3fd8-43ab-85b6-a3d37e218887\") " pod="openstack/glance-default-internal-api-0" Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.851269 4824 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"97b2f7d6-3fd8-43ab-85b6-a3d37e218887\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/glance-default-internal-api-0" Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.851907 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/97b2f7d6-3fd8-43ab-85b6-a3d37e218887-logs\") pod \"glance-default-internal-api-0\" (UID: \"97b2f7d6-3fd8-43ab-85b6-a3d37e218887\") " pod="openstack/glance-default-internal-api-0" Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.856320 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/97b2f7d6-3fd8-43ab-85b6-a3d37e218887-scripts\") pod \"glance-default-internal-api-0\" (UID: \"97b2f7d6-3fd8-43ab-85b6-a3d37e218887\") " pod="openstack/glance-default-internal-api-0" Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.860562 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/97b2f7d6-3fd8-43ab-85b6-a3d37e218887-config-data\") pod \"glance-default-internal-api-0\" (UID: \"97b2f7d6-3fd8-43ab-85b6-a3d37e218887\") " pod="openstack/glance-default-internal-api-0" Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.863125 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8b5c85b87-9r9wx" Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.869490 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97b2f7d6-3fd8-43ab-85b6-a3d37e218887-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"97b2f7d6-3fd8-43ab-85b6-a3d37e218887\") " pod="openstack/glance-default-internal-api-0" Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.871817 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c5cc7c5ff-w7tn4" event={"ID":"f735f6b9-a885-4a39-b780-17d8d1b30a88","Type":"ContainerStarted","Data":"4529389a7035d38d7cda79c71395c4c7d5badfb9ea49bd926b2f342cde21be0b"} Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.871894 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c5cc7c5ff-w7tn4" event={"ID":"f735f6b9-a885-4a39-b780-17d8d1b30a88","Type":"ContainerStarted","Data":"7a4a1791a985935514901bf3702467b7df3a6b688973a5a1c2e2ea00a0d079ee"} Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.872174 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/97b2f7d6-3fd8-43ab-85b6-a3d37e218887-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"97b2f7d6-3fd8-43ab-85b6-a3d37e218887\") " pod="openstack/glance-default-internal-api-0" Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.883511 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7ff5475cc9-9l788" event={"ID":"67457ff8-d89b-4f2b-8f49-b6323746b4b4","Type":"ContainerDied","Data":"06554e9c3d235541783f28313fcf3695b11706ce9c9bde7757a9ef53d4c595d4"} Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.883573 4824 scope.go:117] "RemoveContainer" containerID="f172c000d0b1e8b0c4db204fddd4668a7710b99c62386eff0a34b706c5d642c1" Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.883698 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7ff5475cc9-9l788" Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.886363 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.894741 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dh64c\" (UniqueName: \"kubernetes.io/projected/97b2f7d6-3fd8-43ab-85b6-a3d37e218887-kube-api-access-dh64c\") pod \"glance-default-internal-api-0\" (UID: \"97b2f7d6-3fd8-43ab-85b6-a3d37e218887\") " pod="openstack/glance-default-internal-api-0" Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.916246 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-ldn9v" event={"ID":"2382b80c-75fc-435f-be5f-0f1eb6a5cd17","Type":"ContainerStarted","Data":"dac702736fb8c27487a0f0433392419f3af8a038eb66fe6f9aa60c761cd56c89"} Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.929194 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-zcq55" event={"ID":"c607b17a-75ac-4e7f-8ff3-9d0edb23cc23","Type":"ContainerStarted","Data":"b897b689cf8a55f703dd1ff5a5a59a7583dfeb25aa27812cda69985492549892"} Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.947492 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 06 10:15:47 crc kubenswrapper[4824]: I1006 10:15:47.955058 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"97b2f7d6-3fd8-43ab-85b6-a3d37e218887\") " pod="openstack/glance-default-internal-api-0" Oct 06 10:15:47 crc kubenswrapper[4824]: W1006 10:15:47.993949 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod13781991_323c_49ae_9c5d_625e2ffd019b.slice/crio-bcefb09d69c2ac87b5481618d163857107777ec48e52728499f611ab6ba9ea44 WatchSource:0}: Error finding container bcefb09d69c2ac87b5481618d163857107777ec48e52728499f611ab6ba9ea44: Status 404 returned error can't find the container with id bcefb09d69c2ac87b5481618d163857107777ec48e52728499f611ab6ba9ea44 Oct 06 10:15:48 crc kubenswrapper[4824]: I1006 10:15:48.000168 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7ff5475cc9-9l788"] Oct 06 10:15:48 crc kubenswrapper[4824]: I1006 10:15:48.026293 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7ff5475cc9-9l788"] Oct 06 10:15:48 crc kubenswrapper[4824]: I1006 10:15:48.029177 4824 scope.go:117] "RemoveContainer" containerID="fcc28e95b9c1cf3a50d7a83b68d38848bb7e2b857eede943b4a727c3f8c0b23c" Oct 06 10:15:48 crc kubenswrapper[4824]: I1006 10:15:48.116875 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7749656757-6jdvp"] Oct 06 10:15:48 crc kubenswrapper[4824]: I1006 10:15:48.201970 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-fzvv8"] Oct 06 10:15:48 crc kubenswrapper[4824]: I1006 10:15:48.217302 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 06 10:15:48 crc kubenswrapper[4824]: I1006 10:15:48.230174 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-xsgsg"] Oct 06 10:15:48 crc kubenswrapper[4824]: I1006 10:15:48.342033 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-ccd5f54d9-dsnj5"] Oct 06 10:15:48 crc kubenswrapper[4824]: I1006 10:15:48.360130 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-ljqzg"] Oct 06 10:15:48 crc kubenswrapper[4824]: W1006 10:15:48.412920 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddfc3bcd2_9baa_4a3a_8588_295e692d8e3e.slice/crio-f67082f14f4158b88ea1fa1ca65841ad86067e9e8e4793f2a079d532cdac363d WatchSource:0}: Error finding container f67082f14f4158b88ea1fa1ca65841ad86067e9e8e4793f2a079d532cdac363d: Status 404 returned error can't find the container with id f67082f14f4158b88ea1fa1ca65841ad86067e9e8e4793f2a079d532cdac363d Oct 06 10:15:48 crc kubenswrapper[4824]: I1006 10:15:48.552192 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-9r9wx"] Oct 06 10:15:48 crc kubenswrapper[4824]: I1006 10:15:48.783457 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 06 10:15:48 crc kubenswrapper[4824]: I1006 10:15:48.967526 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-ljqzg" event={"ID":"dfc3bcd2-9baa-4a3a-8588-295e692d8e3e","Type":"ContainerStarted","Data":"f67082f14f4158b88ea1fa1ca65841ad86067e9e8e4793f2a079d532cdac363d"} Oct 06 10:15:48 crc kubenswrapper[4824]: I1006 10:15:48.975405 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-zcq55" event={"ID":"c607b17a-75ac-4e7f-8ff3-9d0edb23cc23","Type":"ContainerStarted","Data":"e21b3838b55eab22121026e4bcc94e71b1ddcd8a052894021b1b38e01c5f0f50"} Oct 06 10:15:48 crc kubenswrapper[4824]: I1006 10:15:48.981721 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-fzvv8" event={"ID":"554a3f37-c7c8-4d09-9da7-df5319dccecd","Type":"ContainerStarted","Data":"696fbfd81e9841b428ae3561c99fa89ee92932fa05c3d70b8d61e032e9d7cba3"} Oct 06 10:15:48 crc kubenswrapper[4824]: I1006 10:15:48.985259 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7749656757-6jdvp" event={"ID":"07c13b84-6c2a-4a6c-b23b-df1e14e077b5","Type":"ContainerStarted","Data":"73b4b5608cf04aa78bc07acbd6029d18e51c22f4afed2bfc0c560650ddbf00aa"} Oct 06 10:15:48 crc kubenswrapper[4824]: I1006 10:15:48.986200 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"13781991-323c-49ae-9c5d-625e2ffd019b","Type":"ContainerStarted","Data":"bcefb09d69c2ac87b5481618d163857107777ec48e52728499f611ab6ba9ea44"} Oct 06 10:15:49 crc kubenswrapper[4824]: I1006 10:15:48.991102 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-xsgsg" event={"ID":"3d78bdee-7130-48a9-ad24-6c8c6525784c","Type":"ContainerStarted","Data":"86eeda4af9d4a8081ccdcc7a0cceaeae21a27abc73451b614b28ba15c64ed645"} Oct 06 10:15:49 crc kubenswrapper[4824]: I1006 10:15:48.991129 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-xsgsg" event={"ID":"3d78bdee-7130-48a9-ad24-6c8c6525784c","Type":"ContainerStarted","Data":"05056eba008dc12ea46d094704e044242e8f4be07ceec5435fc1eea48dfad173"} Oct 06 10:15:49 crc kubenswrapper[4824]: I1006 10:15:49.002327 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"2e26a57e-6683-48b2-a0ea-3f8ff1027fdc","Type":"ContainerStarted","Data":"a28da5c24c055bfee9835c1c721a89e81c17ace9200bd8211d363675b148b925"} Oct 06 10:15:49 crc kubenswrapper[4824]: I1006 10:15:49.003709 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-zcq55" podStartSLOduration=4.003694229 podStartE2EDuration="4.003694229s" podCreationTimestamp="2025-10-06 10:15:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:15:49.001053354 +0000 UTC m=+1058.365476215" watchObservedRunningTime="2025-10-06 10:15:49.003694229 +0000 UTC m=+1058.368117090" Oct 06 10:15:49 crc kubenswrapper[4824]: I1006 10:15:49.027648 4824 generic.go:334] "Generic (PLEG): container finished" podID="f735f6b9-a885-4a39-b780-17d8d1b30a88" containerID="4529389a7035d38d7cda79c71395c4c7d5badfb9ea49bd926b2f342cde21be0b" exitCode=0 Oct 06 10:15:49 crc kubenswrapper[4824]: I1006 10:15:49.027741 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c5cc7c5ff-w7tn4" event={"ID":"f735f6b9-a885-4a39-b780-17d8d1b30a88","Type":"ContainerDied","Data":"4529389a7035d38d7cda79c71395c4c7d5badfb9ea49bd926b2f342cde21be0b"} Oct 06 10:15:49 crc kubenswrapper[4824]: I1006 10:15:49.045458 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-ccd5f54d9-dsnj5" event={"ID":"3d28a180-78e5-4977-bccb-9210f5fb0a7f","Type":"ContainerStarted","Data":"4122d854ffac7724984a0f667d6a5bc1650958517bb6edef3a883540de1a3236"} Oct 06 10:15:49 crc kubenswrapper[4824]: I1006 10:15:49.055772 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b5c85b87-9r9wx" event={"ID":"629a3cfa-5d3a-416a-9316-8d64594f34f1","Type":"ContainerStarted","Data":"9617a013ed1ae48301abd05bf70da4791eb94a06a50eb5a2480c688c149922a4"} Oct 06 10:15:49 crc kubenswrapper[4824]: I1006 10:15:49.055826 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b5c85b87-9r9wx" event={"ID":"629a3cfa-5d3a-416a-9316-8d64594f34f1","Type":"ContainerStarted","Data":"450870ccbe15cdeceee25a40f63b1b97cc7c7dccb95f272a11c1b2d4b55929a5"} Oct 06 10:15:49 crc kubenswrapper[4824]: I1006 10:15:49.067347 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 06 10:15:49 crc kubenswrapper[4824]: I1006 10:15:49.088371 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-xsgsg" podStartSLOduration=3.088344382 podStartE2EDuration="3.088344382s" podCreationTimestamp="2025-10-06 10:15:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:15:49.043666618 +0000 UTC m=+1058.408089479" watchObservedRunningTime="2025-10-06 10:15:49.088344382 +0000 UTC m=+1058.452767243" Oct 06 10:15:49 crc kubenswrapper[4824]: I1006 10:15:49.123208 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 06 10:15:49 crc kubenswrapper[4824]: I1006 10:15:49.199298 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7749656757-6jdvp"] Oct 06 10:15:49 crc kubenswrapper[4824]: I1006 10:15:49.263234 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 06 10:15:49 crc kubenswrapper[4824]: I1006 10:15:49.275066 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-5bb95d74fc-r95ns"] Oct 06 10:15:49 crc kubenswrapper[4824]: I1006 10:15:49.276889 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5bb95d74fc-r95ns" Oct 06 10:15:49 crc kubenswrapper[4824]: I1006 10:15:49.361377 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="67457ff8-d89b-4f2b-8f49-b6323746b4b4" path="/var/lib/kubelet/pods/67457ff8-d89b-4f2b-8f49-b6323746b4b4/volumes" Oct 06 10:15:49 crc kubenswrapper[4824]: I1006 10:15:49.362134 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 06 10:15:49 crc kubenswrapper[4824]: I1006 10:15:49.376242 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5bb95d74fc-r95ns"] Oct 06 10:15:49 crc kubenswrapper[4824]: I1006 10:15:49.415157 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/541d798b-323f-48c4-b0c1-a8e9d4333f6c-scripts\") pod \"horizon-5bb95d74fc-r95ns\" (UID: \"541d798b-323f-48c4-b0c1-a8e9d4333f6c\") " pod="openstack/horizon-5bb95d74fc-r95ns" Oct 06 10:15:49 crc kubenswrapper[4824]: I1006 10:15:49.415254 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/541d798b-323f-48c4-b0c1-a8e9d4333f6c-config-data\") pod \"horizon-5bb95d74fc-r95ns\" (UID: \"541d798b-323f-48c4-b0c1-a8e9d4333f6c\") " pod="openstack/horizon-5bb95d74fc-r95ns" Oct 06 10:15:49 crc kubenswrapper[4824]: I1006 10:15:49.415309 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/541d798b-323f-48c4-b0c1-a8e9d4333f6c-logs\") pod \"horizon-5bb95d74fc-r95ns\" (UID: \"541d798b-323f-48c4-b0c1-a8e9d4333f6c\") " pod="openstack/horizon-5bb95d74fc-r95ns" Oct 06 10:15:49 crc kubenswrapper[4824]: I1006 10:15:49.415361 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-frht4\" (UniqueName: \"kubernetes.io/projected/541d798b-323f-48c4-b0c1-a8e9d4333f6c-kube-api-access-frht4\") pod \"horizon-5bb95d74fc-r95ns\" (UID: \"541d798b-323f-48c4-b0c1-a8e9d4333f6c\") " pod="openstack/horizon-5bb95d74fc-r95ns" Oct 06 10:15:49 crc kubenswrapper[4824]: I1006 10:15:49.415416 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/541d798b-323f-48c4-b0c1-a8e9d4333f6c-horizon-secret-key\") pod \"horizon-5bb95d74fc-r95ns\" (UID: \"541d798b-323f-48c4-b0c1-a8e9d4333f6c\") " pod="openstack/horizon-5bb95d74fc-r95ns" Oct 06 10:15:49 crc kubenswrapper[4824]: I1006 10:15:49.517168 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/541d798b-323f-48c4-b0c1-a8e9d4333f6c-scripts\") pod \"horizon-5bb95d74fc-r95ns\" (UID: \"541d798b-323f-48c4-b0c1-a8e9d4333f6c\") " pod="openstack/horizon-5bb95d74fc-r95ns" Oct 06 10:15:49 crc kubenswrapper[4824]: I1006 10:15:49.517280 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/541d798b-323f-48c4-b0c1-a8e9d4333f6c-config-data\") pod \"horizon-5bb95d74fc-r95ns\" (UID: \"541d798b-323f-48c4-b0c1-a8e9d4333f6c\") " pod="openstack/horizon-5bb95d74fc-r95ns" Oct 06 10:15:49 crc kubenswrapper[4824]: I1006 10:15:49.517359 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/541d798b-323f-48c4-b0c1-a8e9d4333f6c-logs\") pod \"horizon-5bb95d74fc-r95ns\" (UID: \"541d798b-323f-48c4-b0c1-a8e9d4333f6c\") " pod="openstack/horizon-5bb95d74fc-r95ns" Oct 06 10:15:49 crc kubenswrapper[4824]: I1006 10:15:49.517433 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-frht4\" (UniqueName: \"kubernetes.io/projected/541d798b-323f-48c4-b0c1-a8e9d4333f6c-kube-api-access-frht4\") pod \"horizon-5bb95d74fc-r95ns\" (UID: \"541d798b-323f-48c4-b0c1-a8e9d4333f6c\") " pod="openstack/horizon-5bb95d74fc-r95ns" Oct 06 10:15:49 crc kubenswrapper[4824]: I1006 10:15:49.517504 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/541d798b-323f-48c4-b0c1-a8e9d4333f6c-horizon-secret-key\") pod \"horizon-5bb95d74fc-r95ns\" (UID: \"541d798b-323f-48c4-b0c1-a8e9d4333f6c\") " pod="openstack/horizon-5bb95d74fc-r95ns" Oct 06 10:15:49 crc kubenswrapper[4824]: I1006 10:15:49.518014 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/541d798b-323f-48c4-b0c1-a8e9d4333f6c-scripts\") pod \"horizon-5bb95d74fc-r95ns\" (UID: \"541d798b-323f-48c4-b0c1-a8e9d4333f6c\") " pod="openstack/horizon-5bb95d74fc-r95ns" Oct 06 10:15:49 crc kubenswrapper[4824]: I1006 10:15:49.518392 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/541d798b-323f-48c4-b0c1-a8e9d4333f6c-logs\") pod \"horizon-5bb95d74fc-r95ns\" (UID: \"541d798b-323f-48c4-b0c1-a8e9d4333f6c\") " pod="openstack/horizon-5bb95d74fc-r95ns" Oct 06 10:15:49 crc kubenswrapper[4824]: I1006 10:15:49.518501 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/541d798b-323f-48c4-b0c1-a8e9d4333f6c-config-data\") pod \"horizon-5bb95d74fc-r95ns\" (UID: \"541d798b-323f-48c4-b0c1-a8e9d4333f6c\") " pod="openstack/horizon-5bb95d74fc-r95ns" Oct 06 10:15:49 crc kubenswrapper[4824]: I1006 10:15:49.538256 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/541d798b-323f-48c4-b0c1-a8e9d4333f6c-horizon-secret-key\") pod \"horizon-5bb95d74fc-r95ns\" (UID: \"541d798b-323f-48c4-b0c1-a8e9d4333f6c\") " pod="openstack/horizon-5bb95d74fc-r95ns" Oct 06 10:15:49 crc kubenswrapper[4824]: I1006 10:15:49.573592 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-frht4\" (UniqueName: \"kubernetes.io/projected/541d798b-323f-48c4-b0c1-a8e9d4333f6c-kube-api-access-frht4\") pod \"horizon-5bb95d74fc-r95ns\" (UID: \"541d798b-323f-48c4-b0c1-a8e9d4333f6c\") " pod="openstack/horizon-5bb95d74fc-r95ns" Oct 06 10:15:49 crc kubenswrapper[4824]: I1006 10:15:49.619501 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5bb95d74fc-r95ns" Oct 06 10:15:49 crc kubenswrapper[4824]: I1006 10:15:49.689652 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c5cc7c5ff-w7tn4" Oct 06 10:15:49 crc kubenswrapper[4824]: I1006 10:15:49.822732 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f735f6b9-a885-4a39-b780-17d8d1b30a88-ovsdbserver-nb\") pod \"f735f6b9-a885-4a39-b780-17d8d1b30a88\" (UID: \"f735f6b9-a885-4a39-b780-17d8d1b30a88\") " Oct 06 10:15:49 crc kubenswrapper[4824]: I1006 10:15:49.822791 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7k85q\" (UniqueName: \"kubernetes.io/projected/f735f6b9-a885-4a39-b780-17d8d1b30a88-kube-api-access-7k85q\") pod \"f735f6b9-a885-4a39-b780-17d8d1b30a88\" (UID: \"f735f6b9-a885-4a39-b780-17d8d1b30a88\") " Oct 06 10:15:49 crc kubenswrapper[4824]: I1006 10:15:49.822866 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f735f6b9-a885-4a39-b780-17d8d1b30a88-dns-swift-storage-0\") pod \"f735f6b9-a885-4a39-b780-17d8d1b30a88\" (UID: \"f735f6b9-a885-4a39-b780-17d8d1b30a88\") " Oct 06 10:15:49 crc kubenswrapper[4824]: I1006 10:15:49.822896 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f735f6b9-a885-4a39-b780-17d8d1b30a88-config\") pod \"f735f6b9-a885-4a39-b780-17d8d1b30a88\" (UID: \"f735f6b9-a885-4a39-b780-17d8d1b30a88\") " Oct 06 10:15:49 crc kubenswrapper[4824]: I1006 10:15:49.822967 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f735f6b9-a885-4a39-b780-17d8d1b30a88-ovsdbserver-sb\") pod \"f735f6b9-a885-4a39-b780-17d8d1b30a88\" (UID: \"f735f6b9-a885-4a39-b780-17d8d1b30a88\") " Oct 06 10:15:49 crc kubenswrapper[4824]: I1006 10:15:49.823080 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f735f6b9-a885-4a39-b780-17d8d1b30a88-dns-svc\") pod \"f735f6b9-a885-4a39-b780-17d8d1b30a88\" (UID: \"f735f6b9-a885-4a39-b780-17d8d1b30a88\") " Oct 06 10:15:49 crc kubenswrapper[4824]: I1006 10:15:49.834217 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f735f6b9-a885-4a39-b780-17d8d1b30a88-kube-api-access-7k85q" (OuterVolumeSpecName: "kube-api-access-7k85q") pod "f735f6b9-a885-4a39-b780-17d8d1b30a88" (UID: "f735f6b9-a885-4a39-b780-17d8d1b30a88"). InnerVolumeSpecName "kube-api-access-7k85q". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:15:49 crc kubenswrapper[4824]: I1006 10:15:49.852683 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f735f6b9-a885-4a39-b780-17d8d1b30a88-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "f735f6b9-a885-4a39-b780-17d8d1b30a88" (UID: "f735f6b9-a885-4a39-b780-17d8d1b30a88"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:15:49 crc kubenswrapper[4824]: I1006 10:15:49.867520 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f735f6b9-a885-4a39-b780-17d8d1b30a88-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "f735f6b9-a885-4a39-b780-17d8d1b30a88" (UID: "f735f6b9-a885-4a39-b780-17d8d1b30a88"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:15:49 crc kubenswrapper[4824]: I1006 10:15:49.886841 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f735f6b9-a885-4a39-b780-17d8d1b30a88-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "f735f6b9-a885-4a39-b780-17d8d1b30a88" (UID: "f735f6b9-a885-4a39-b780-17d8d1b30a88"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:15:49 crc kubenswrapper[4824]: I1006 10:15:49.894135 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f735f6b9-a885-4a39-b780-17d8d1b30a88-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "f735f6b9-a885-4a39-b780-17d8d1b30a88" (UID: "f735f6b9-a885-4a39-b780-17d8d1b30a88"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:15:49 crc kubenswrapper[4824]: I1006 10:15:49.926222 4824 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f735f6b9-a885-4a39-b780-17d8d1b30a88-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 06 10:15:49 crc kubenswrapper[4824]: I1006 10:15:49.926250 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f735f6b9-a885-4a39-b780-17d8d1b30a88-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 06 10:15:49 crc kubenswrapper[4824]: I1006 10:15:49.926278 4824 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f735f6b9-a885-4a39-b780-17d8d1b30a88-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 06 10:15:49 crc kubenswrapper[4824]: I1006 10:15:49.926288 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f735f6b9-a885-4a39-b780-17d8d1b30a88-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 06 10:15:49 crc kubenswrapper[4824]: I1006 10:15:49.926297 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7k85q\" (UniqueName: \"kubernetes.io/projected/f735f6b9-a885-4a39-b780-17d8d1b30a88-kube-api-access-7k85q\") on node \"crc\" DevicePath \"\"" Oct 06 10:15:49 crc kubenswrapper[4824]: I1006 10:15:49.958912 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f735f6b9-a885-4a39-b780-17d8d1b30a88-config" (OuterVolumeSpecName: "config") pod "f735f6b9-a885-4a39-b780-17d8d1b30a88" (UID: "f735f6b9-a885-4a39-b780-17d8d1b30a88"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:15:50 crc kubenswrapper[4824]: I1006 10:15:50.029620 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f735f6b9-a885-4a39-b780-17d8d1b30a88-config\") on node \"crc\" DevicePath \"\"" Oct 06 10:15:50 crc kubenswrapper[4824]: I1006 10:15:50.097501 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c5cc7c5ff-w7tn4" event={"ID":"f735f6b9-a885-4a39-b780-17d8d1b30a88","Type":"ContainerDied","Data":"7a4a1791a985935514901bf3702467b7df3a6b688973a5a1c2e2ea00a0d079ee"} Oct 06 10:15:50 crc kubenswrapper[4824]: I1006 10:15:50.097559 4824 scope.go:117] "RemoveContainer" containerID="4529389a7035d38d7cda79c71395c4c7d5badfb9ea49bd926b2f342cde21be0b" Oct 06 10:15:50 crc kubenswrapper[4824]: I1006 10:15:50.097579 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c5cc7c5ff-w7tn4" Oct 06 10:15:50 crc kubenswrapper[4824]: I1006 10:15:50.108880 4824 generic.go:334] "Generic (PLEG): container finished" podID="629a3cfa-5d3a-416a-9316-8d64594f34f1" containerID="9617a013ed1ae48301abd05bf70da4791eb94a06a50eb5a2480c688c149922a4" exitCode=0 Oct 06 10:15:50 crc kubenswrapper[4824]: I1006 10:15:50.109345 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b5c85b87-9r9wx" event={"ID":"629a3cfa-5d3a-416a-9316-8d64594f34f1","Type":"ContainerDied","Data":"9617a013ed1ae48301abd05bf70da4791eb94a06a50eb5a2480c688c149922a4"} Oct 06 10:15:50 crc kubenswrapper[4824]: I1006 10:15:50.109377 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b5c85b87-9r9wx" event={"ID":"629a3cfa-5d3a-416a-9316-8d64594f34f1","Type":"ContainerStarted","Data":"0e86b0d6cbf1ecd7c499cc998c99d2bdd8a0c47ba05ae448195ee46a9eca3a32"} Oct 06 10:15:50 crc kubenswrapper[4824]: I1006 10:15:50.109424 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-8b5c85b87-9r9wx" Oct 06 10:15:50 crc kubenswrapper[4824]: I1006 10:15:50.118927 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"97b2f7d6-3fd8-43ab-85b6-a3d37e218887","Type":"ContainerStarted","Data":"927b0e04773f5ccabdce43dca2c18d30b75f19cf4fe15f6ca45c98b7944307b7"} Oct 06 10:15:50 crc kubenswrapper[4824]: I1006 10:15:50.135703 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-8b5c85b87-9r9wx" podStartSLOduration=4.135672412 podStartE2EDuration="4.135672412s" podCreationTimestamp="2025-10-06 10:15:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:15:50.133294024 +0000 UTC m=+1059.497716895" watchObservedRunningTime="2025-10-06 10:15:50.135672412 +0000 UTC m=+1059.500095273" Oct 06 10:15:50 crc kubenswrapper[4824]: I1006 10:15:50.288665 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c5cc7c5ff-w7tn4"] Oct 06 10:15:50 crc kubenswrapper[4824]: I1006 10:15:50.304523 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5c5cc7c5ff-w7tn4"] Oct 06 10:15:50 crc kubenswrapper[4824]: I1006 10:15:50.325246 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5bb95d74fc-r95ns"] Oct 06 10:15:51 crc kubenswrapper[4824]: I1006 10:15:51.139690 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"2e26a57e-6683-48b2-a0ea-3f8ff1027fdc","Type":"ContainerStarted","Data":"bd755795208f73b11771157f14d20e980d843dd07c235a308a1c5ef8ace0f43f"} Oct 06 10:15:51 crc kubenswrapper[4824]: I1006 10:15:51.145136 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5bb95d74fc-r95ns" event={"ID":"541d798b-323f-48c4-b0c1-a8e9d4333f6c","Type":"ContainerStarted","Data":"1c38bb2748e1f4ee0eb9c40b22bd9a887d55033eae5448a150da9aa776021f3c"} Oct 06 10:15:51 crc kubenswrapper[4824]: I1006 10:15:51.302598 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f735f6b9-a885-4a39-b780-17d8d1b30a88" path="/var/lib/kubelet/pods/f735f6b9-a885-4a39-b780-17d8d1b30a88/volumes" Oct 06 10:15:52 crc kubenswrapper[4824]: I1006 10:15:52.166099 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"2e26a57e-6683-48b2-a0ea-3f8ff1027fdc","Type":"ContainerStarted","Data":"d09824b8fc32e72c2d6d30bfea36532762fa0afa48b2d6885a6612dd89035ad0"} Oct 06 10:15:52 crc kubenswrapper[4824]: I1006 10:15:52.166307 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="2e26a57e-6683-48b2-a0ea-3f8ff1027fdc" containerName="glance-log" containerID="cri-o://bd755795208f73b11771157f14d20e980d843dd07c235a308a1c5ef8ace0f43f" gracePeriod=30 Oct 06 10:15:52 crc kubenswrapper[4824]: I1006 10:15:52.166458 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="2e26a57e-6683-48b2-a0ea-3f8ff1027fdc" containerName="glance-httpd" containerID="cri-o://d09824b8fc32e72c2d6d30bfea36532762fa0afa48b2d6885a6612dd89035ad0" gracePeriod=30 Oct 06 10:15:52 crc kubenswrapper[4824]: I1006 10:15:52.172877 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"97b2f7d6-3fd8-43ab-85b6-a3d37e218887","Type":"ContainerStarted","Data":"4b4abcc8eb3bb8c1ba125783f2c3e87528d26f2c2a2fae76b8c8c4f23d84ff1f"} Oct 06 10:15:52 crc kubenswrapper[4824]: I1006 10:15:52.196642 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=6.196619627 podStartE2EDuration="6.196619627s" podCreationTimestamp="2025-10-06 10:15:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:15:52.189946503 +0000 UTC m=+1061.554369364" watchObservedRunningTime="2025-10-06 10:15:52.196619627 +0000 UTC m=+1061.561042498" Oct 06 10:15:53 crc kubenswrapper[4824]: I1006 10:15:53.196588 4824 generic.go:334] "Generic (PLEG): container finished" podID="2e26a57e-6683-48b2-a0ea-3f8ff1027fdc" containerID="d09824b8fc32e72c2d6d30bfea36532762fa0afa48b2d6885a6612dd89035ad0" exitCode=0 Oct 06 10:15:53 crc kubenswrapper[4824]: I1006 10:15:53.196837 4824 generic.go:334] "Generic (PLEG): container finished" podID="2e26a57e-6683-48b2-a0ea-3f8ff1027fdc" containerID="bd755795208f73b11771157f14d20e980d843dd07c235a308a1c5ef8ace0f43f" exitCode=143 Oct 06 10:15:53 crc kubenswrapper[4824]: I1006 10:15:53.196780 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"2e26a57e-6683-48b2-a0ea-3f8ff1027fdc","Type":"ContainerDied","Data":"d09824b8fc32e72c2d6d30bfea36532762fa0afa48b2d6885a6612dd89035ad0"} Oct 06 10:15:53 crc kubenswrapper[4824]: I1006 10:15:53.196903 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"2e26a57e-6683-48b2-a0ea-3f8ff1027fdc","Type":"ContainerDied","Data":"bd755795208f73b11771157f14d20e980d843dd07c235a308a1c5ef8ace0f43f"} Oct 06 10:15:53 crc kubenswrapper[4824]: I1006 10:15:53.200819 4824 generic.go:334] "Generic (PLEG): container finished" podID="c607b17a-75ac-4e7f-8ff3-9d0edb23cc23" containerID="e21b3838b55eab22121026e4bcc94e71b1ddcd8a052894021b1b38e01c5f0f50" exitCode=0 Oct 06 10:15:53 crc kubenswrapper[4824]: I1006 10:15:53.200870 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-zcq55" event={"ID":"c607b17a-75ac-4e7f-8ff3-9d0edb23cc23","Type":"ContainerDied","Data":"e21b3838b55eab22121026e4bcc94e71b1ddcd8a052894021b1b38e01c5f0f50"} Oct 06 10:15:53 crc kubenswrapper[4824]: I1006 10:15:53.207749 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"97b2f7d6-3fd8-43ab-85b6-a3d37e218887","Type":"ContainerStarted","Data":"51657330175f5797f5507a724e5f16d8bc068fe1ab011ba8ea18d9f8f4298e0c"} Oct 06 10:15:53 crc kubenswrapper[4824]: I1006 10:15:53.207966 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="97b2f7d6-3fd8-43ab-85b6-a3d37e218887" containerName="glance-httpd" containerID="cri-o://51657330175f5797f5507a724e5f16d8bc068fe1ab011ba8ea18d9f8f4298e0c" gracePeriod=30 Oct 06 10:15:53 crc kubenswrapper[4824]: I1006 10:15:53.207935 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="97b2f7d6-3fd8-43ab-85b6-a3d37e218887" containerName="glance-log" containerID="cri-o://4b4abcc8eb3bb8c1ba125783f2c3e87528d26f2c2a2fae76b8c8c4f23d84ff1f" gracePeriod=30 Oct 06 10:15:53 crc kubenswrapper[4824]: I1006 10:15:53.254471 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=7.254434346 podStartE2EDuration="7.254434346s" podCreationTimestamp="2025-10-06 10:15:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:15:53.247056414 +0000 UTC m=+1062.611479275" watchObservedRunningTime="2025-10-06 10:15:53.254434346 +0000 UTC m=+1062.618857207" Oct 06 10:15:54 crc kubenswrapper[4824]: I1006 10:15:54.236681 4824 generic.go:334] "Generic (PLEG): container finished" podID="97b2f7d6-3fd8-43ab-85b6-a3d37e218887" containerID="51657330175f5797f5507a724e5f16d8bc068fe1ab011ba8ea18d9f8f4298e0c" exitCode=0 Oct 06 10:15:54 crc kubenswrapper[4824]: I1006 10:15:54.236972 4824 generic.go:334] "Generic (PLEG): container finished" podID="97b2f7d6-3fd8-43ab-85b6-a3d37e218887" containerID="4b4abcc8eb3bb8c1ba125783f2c3e87528d26f2c2a2fae76b8c8c4f23d84ff1f" exitCode=143 Oct 06 10:15:54 crc kubenswrapper[4824]: I1006 10:15:54.236809 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"97b2f7d6-3fd8-43ab-85b6-a3d37e218887","Type":"ContainerDied","Data":"51657330175f5797f5507a724e5f16d8bc068fe1ab011ba8ea18d9f8f4298e0c"} Oct 06 10:15:54 crc kubenswrapper[4824]: I1006 10:15:54.237152 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"97b2f7d6-3fd8-43ab-85b6-a3d37e218887","Type":"ContainerDied","Data":"4b4abcc8eb3bb8c1ba125783f2c3e87528d26f2c2a2fae76b8c8c4f23d84ff1f"} Oct 06 10:15:55 crc kubenswrapper[4824]: I1006 10:15:55.808226 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-ccd5f54d9-dsnj5"] Oct 06 10:15:55 crc kubenswrapper[4824]: I1006 10:15:55.856204 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-78c847f9f4-tv8zc"] Oct 06 10:15:55 crc kubenswrapper[4824]: E1006 10:15:55.864894 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f735f6b9-a885-4a39-b780-17d8d1b30a88" containerName="init" Oct 06 10:15:55 crc kubenswrapper[4824]: I1006 10:15:55.864948 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="f735f6b9-a885-4a39-b780-17d8d1b30a88" containerName="init" Oct 06 10:15:55 crc kubenswrapper[4824]: I1006 10:15:55.865750 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="f735f6b9-a885-4a39-b780-17d8d1b30a88" containerName="init" Oct 06 10:15:55 crc kubenswrapper[4824]: I1006 10:15:55.867306 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-78c847f9f4-tv8zc" Oct 06 10:15:55 crc kubenswrapper[4824]: I1006 10:15:55.888426 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-horizon-svc" Oct 06 10:15:55 crc kubenswrapper[4824]: I1006 10:15:55.941262 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-78c847f9f4-tv8zc"] Oct 06 10:15:55 crc kubenswrapper[4824]: I1006 10:15:55.971355 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-5bb95d74fc-r95ns"] Oct 06 10:15:55 crc kubenswrapper[4824]: I1006 10:15:55.977499 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-68fdc5dd48-gfq4d"] Oct 06 10:15:55 crc kubenswrapper[4824]: I1006 10:15:55.978955 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-68fdc5dd48-gfq4d" Oct 06 10:15:55 crc kubenswrapper[4824]: I1006 10:15:55.991731 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-68fdc5dd48-gfq4d"] Oct 06 10:15:56 crc kubenswrapper[4824]: I1006 10:15:56.000104 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/68ef92a4-c4bc-4acc-be58-e14010395b8e-horizon-secret-key\") pod \"horizon-78c847f9f4-tv8zc\" (UID: \"68ef92a4-c4bc-4acc-be58-e14010395b8e\") " pod="openstack/horizon-78c847f9f4-tv8zc" Oct 06 10:15:56 crc kubenswrapper[4824]: I1006 10:15:56.000193 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hxgpf\" (UniqueName: \"kubernetes.io/projected/68ef92a4-c4bc-4acc-be58-e14010395b8e-kube-api-access-hxgpf\") pod \"horizon-78c847f9f4-tv8zc\" (UID: \"68ef92a4-c4bc-4acc-be58-e14010395b8e\") " pod="openstack/horizon-78c847f9f4-tv8zc" Oct 06 10:15:56 crc kubenswrapper[4824]: I1006 10:15:56.000232 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/68ef92a4-c4bc-4acc-be58-e14010395b8e-logs\") pod \"horizon-78c847f9f4-tv8zc\" (UID: \"68ef92a4-c4bc-4acc-be58-e14010395b8e\") " pod="openstack/horizon-78c847f9f4-tv8zc" Oct 06 10:15:56 crc kubenswrapper[4824]: I1006 10:15:56.000259 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68ef92a4-c4bc-4acc-be58-e14010395b8e-combined-ca-bundle\") pod \"horizon-78c847f9f4-tv8zc\" (UID: \"68ef92a4-c4bc-4acc-be58-e14010395b8e\") " pod="openstack/horizon-78c847f9f4-tv8zc" Oct 06 10:15:56 crc kubenswrapper[4824]: I1006 10:15:56.000283 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/68ef92a4-c4bc-4acc-be58-e14010395b8e-horizon-tls-certs\") pod \"horizon-78c847f9f4-tv8zc\" (UID: \"68ef92a4-c4bc-4acc-be58-e14010395b8e\") " pod="openstack/horizon-78c847f9f4-tv8zc" Oct 06 10:15:56 crc kubenswrapper[4824]: I1006 10:15:56.000326 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/68ef92a4-c4bc-4acc-be58-e14010395b8e-config-data\") pod \"horizon-78c847f9f4-tv8zc\" (UID: \"68ef92a4-c4bc-4acc-be58-e14010395b8e\") " pod="openstack/horizon-78c847f9f4-tv8zc" Oct 06 10:15:56 crc kubenswrapper[4824]: I1006 10:15:56.000355 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/68ef92a4-c4bc-4acc-be58-e14010395b8e-scripts\") pod \"horizon-78c847f9f4-tv8zc\" (UID: \"68ef92a4-c4bc-4acc-be58-e14010395b8e\") " pod="openstack/horizon-78c847f9f4-tv8zc" Oct 06 10:15:56 crc kubenswrapper[4824]: I1006 10:15:56.102283 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/194bef17-98ac-413f-a91e-0abc9835f2f6-logs\") pod \"horizon-68fdc5dd48-gfq4d\" (UID: \"194bef17-98ac-413f-a91e-0abc9835f2f6\") " pod="openstack/horizon-68fdc5dd48-gfq4d" Oct 06 10:15:56 crc kubenswrapper[4824]: I1006 10:15:56.102345 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jqnzf\" (UniqueName: \"kubernetes.io/projected/194bef17-98ac-413f-a91e-0abc9835f2f6-kube-api-access-jqnzf\") pod \"horizon-68fdc5dd48-gfq4d\" (UID: \"194bef17-98ac-413f-a91e-0abc9835f2f6\") " pod="openstack/horizon-68fdc5dd48-gfq4d" Oct 06 10:15:56 crc kubenswrapper[4824]: I1006 10:15:56.102376 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/68ef92a4-c4bc-4acc-be58-e14010395b8e-horizon-secret-key\") pod \"horizon-78c847f9f4-tv8zc\" (UID: \"68ef92a4-c4bc-4acc-be58-e14010395b8e\") " pod="openstack/horizon-78c847f9f4-tv8zc" Oct 06 10:15:56 crc kubenswrapper[4824]: I1006 10:15:56.102613 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/194bef17-98ac-413f-a91e-0abc9835f2f6-horizon-tls-certs\") pod \"horizon-68fdc5dd48-gfq4d\" (UID: \"194bef17-98ac-413f-a91e-0abc9835f2f6\") " pod="openstack/horizon-68fdc5dd48-gfq4d" Oct 06 10:15:56 crc kubenswrapper[4824]: I1006 10:15:56.102679 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/194bef17-98ac-413f-a91e-0abc9835f2f6-horizon-secret-key\") pod \"horizon-68fdc5dd48-gfq4d\" (UID: \"194bef17-98ac-413f-a91e-0abc9835f2f6\") " pod="openstack/horizon-68fdc5dd48-gfq4d" Oct 06 10:15:56 crc kubenswrapper[4824]: I1006 10:15:56.102742 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hxgpf\" (UniqueName: \"kubernetes.io/projected/68ef92a4-c4bc-4acc-be58-e14010395b8e-kube-api-access-hxgpf\") pod \"horizon-78c847f9f4-tv8zc\" (UID: \"68ef92a4-c4bc-4acc-be58-e14010395b8e\") " pod="openstack/horizon-78c847f9f4-tv8zc" Oct 06 10:15:56 crc kubenswrapper[4824]: I1006 10:15:56.102809 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/194bef17-98ac-413f-a91e-0abc9835f2f6-scripts\") pod \"horizon-68fdc5dd48-gfq4d\" (UID: \"194bef17-98ac-413f-a91e-0abc9835f2f6\") " pod="openstack/horizon-68fdc5dd48-gfq4d" Oct 06 10:15:56 crc kubenswrapper[4824]: I1006 10:15:56.102889 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/68ef92a4-c4bc-4acc-be58-e14010395b8e-logs\") pod \"horizon-78c847f9f4-tv8zc\" (UID: \"68ef92a4-c4bc-4acc-be58-e14010395b8e\") " pod="openstack/horizon-78c847f9f4-tv8zc" Oct 06 10:15:56 crc kubenswrapper[4824]: I1006 10:15:56.103054 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68ef92a4-c4bc-4acc-be58-e14010395b8e-combined-ca-bundle\") pod \"horizon-78c847f9f4-tv8zc\" (UID: \"68ef92a4-c4bc-4acc-be58-e14010395b8e\") " pod="openstack/horizon-78c847f9f4-tv8zc" Oct 06 10:15:56 crc kubenswrapper[4824]: I1006 10:15:56.103168 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/68ef92a4-c4bc-4acc-be58-e14010395b8e-horizon-tls-certs\") pod \"horizon-78c847f9f4-tv8zc\" (UID: \"68ef92a4-c4bc-4acc-be58-e14010395b8e\") " pod="openstack/horizon-78c847f9f4-tv8zc" Oct 06 10:15:56 crc kubenswrapper[4824]: I1006 10:15:56.103301 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/194bef17-98ac-413f-a91e-0abc9835f2f6-config-data\") pod \"horizon-68fdc5dd48-gfq4d\" (UID: \"194bef17-98ac-413f-a91e-0abc9835f2f6\") " pod="openstack/horizon-68fdc5dd48-gfq4d" Oct 06 10:15:56 crc kubenswrapper[4824]: I1006 10:15:56.103337 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/194bef17-98ac-413f-a91e-0abc9835f2f6-combined-ca-bundle\") pod \"horizon-68fdc5dd48-gfq4d\" (UID: \"194bef17-98ac-413f-a91e-0abc9835f2f6\") " pod="openstack/horizon-68fdc5dd48-gfq4d" Oct 06 10:15:56 crc kubenswrapper[4824]: I1006 10:15:56.103497 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/68ef92a4-c4bc-4acc-be58-e14010395b8e-config-data\") pod \"horizon-78c847f9f4-tv8zc\" (UID: \"68ef92a4-c4bc-4acc-be58-e14010395b8e\") " pod="openstack/horizon-78c847f9f4-tv8zc" Oct 06 10:15:56 crc kubenswrapper[4824]: I1006 10:15:56.105743 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/68ef92a4-c4bc-4acc-be58-e14010395b8e-config-data\") pod \"horizon-78c847f9f4-tv8zc\" (UID: \"68ef92a4-c4bc-4acc-be58-e14010395b8e\") " pod="openstack/horizon-78c847f9f4-tv8zc" Oct 06 10:15:56 crc kubenswrapper[4824]: I1006 10:15:56.106144 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/68ef92a4-c4bc-4acc-be58-e14010395b8e-scripts\") pod \"horizon-78c847f9f4-tv8zc\" (UID: \"68ef92a4-c4bc-4acc-be58-e14010395b8e\") " pod="openstack/horizon-78c847f9f4-tv8zc" Oct 06 10:15:56 crc kubenswrapper[4824]: I1006 10:15:56.106880 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/68ef92a4-c4bc-4acc-be58-e14010395b8e-scripts\") pod \"horizon-78c847f9f4-tv8zc\" (UID: \"68ef92a4-c4bc-4acc-be58-e14010395b8e\") " pod="openstack/horizon-78c847f9f4-tv8zc" Oct 06 10:15:56 crc kubenswrapper[4824]: I1006 10:15:56.106957 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/68ef92a4-c4bc-4acc-be58-e14010395b8e-logs\") pod \"horizon-78c847f9f4-tv8zc\" (UID: \"68ef92a4-c4bc-4acc-be58-e14010395b8e\") " pod="openstack/horizon-78c847f9f4-tv8zc" Oct 06 10:15:56 crc kubenswrapper[4824]: I1006 10:15:56.109383 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/68ef92a4-c4bc-4acc-be58-e14010395b8e-horizon-secret-key\") pod \"horizon-78c847f9f4-tv8zc\" (UID: \"68ef92a4-c4bc-4acc-be58-e14010395b8e\") " pod="openstack/horizon-78c847f9f4-tv8zc" Oct 06 10:15:56 crc kubenswrapper[4824]: I1006 10:15:56.110609 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68ef92a4-c4bc-4acc-be58-e14010395b8e-combined-ca-bundle\") pod \"horizon-78c847f9f4-tv8zc\" (UID: \"68ef92a4-c4bc-4acc-be58-e14010395b8e\") " pod="openstack/horizon-78c847f9f4-tv8zc" Oct 06 10:15:56 crc kubenswrapper[4824]: I1006 10:15:56.115382 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/68ef92a4-c4bc-4acc-be58-e14010395b8e-horizon-tls-certs\") pod \"horizon-78c847f9f4-tv8zc\" (UID: \"68ef92a4-c4bc-4acc-be58-e14010395b8e\") " pod="openstack/horizon-78c847f9f4-tv8zc" Oct 06 10:15:56 crc kubenswrapper[4824]: I1006 10:15:56.121196 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hxgpf\" (UniqueName: \"kubernetes.io/projected/68ef92a4-c4bc-4acc-be58-e14010395b8e-kube-api-access-hxgpf\") pod \"horizon-78c847f9f4-tv8zc\" (UID: \"68ef92a4-c4bc-4acc-be58-e14010395b8e\") " pod="openstack/horizon-78c847f9f4-tv8zc" Oct 06 10:15:56 crc kubenswrapper[4824]: I1006 10:15:56.208443 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/194bef17-98ac-413f-a91e-0abc9835f2f6-scripts\") pod \"horizon-68fdc5dd48-gfq4d\" (UID: \"194bef17-98ac-413f-a91e-0abc9835f2f6\") " pod="openstack/horizon-68fdc5dd48-gfq4d" Oct 06 10:15:56 crc kubenswrapper[4824]: I1006 10:15:56.208560 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/194bef17-98ac-413f-a91e-0abc9835f2f6-config-data\") pod \"horizon-68fdc5dd48-gfq4d\" (UID: \"194bef17-98ac-413f-a91e-0abc9835f2f6\") " pod="openstack/horizon-68fdc5dd48-gfq4d" Oct 06 10:15:56 crc kubenswrapper[4824]: I1006 10:15:56.208588 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/194bef17-98ac-413f-a91e-0abc9835f2f6-combined-ca-bundle\") pod \"horizon-68fdc5dd48-gfq4d\" (UID: \"194bef17-98ac-413f-a91e-0abc9835f2f6\") " pod="openstack/horizon-68fdc5dd48-gfq4d" Oct 06 10:15:56 crc kubenswrapper[4824]: I1006 10:15:56.208651 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/194bef17-98ac-413f-a91e-0abc9835f2f6-logs\") pod \"horizon-68fdc5dd48-gfq4d\" (UID: \"194bef17-98ac-413f-a91e-0abc9835f2f6\") " pod="openstack/horizon-68fdc5dd48-gfq4d" Oct 06 10:15:56 crc kubenswrapper[4824]: I1006 10:15:56.208690 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jqnzf\" (UniqueName: \"kubernetes.io/projected/194bef17-98ac-413f-a91e-0abc9835f2f6-kube-api-access-jqnzf\") pod \"horizon-68fdc5dd48-gfq4d\" (UID: \"194bef17-98ac-413f-a91e-0abc9835f2f6\") " pod="openstack/horizon-68fdc5dd48-gfq4d" Oct 06 10:15:56 crc kubenswrapper[4824]: I1006 10:15:56.208733 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/194bef17-98ac-413f-a91e-0abc9835f2f6-horizon-tls-certs\") pod \"horizon-68fdc5dd48-gfq4d\" (UID: \"194bef17-98ac-413f-a91e-0abc9835f2f6\") " pod="openstack/horizon-68fdc5dd48-gfq4d" Oct 06 10:15:56 crc kubenswrapper[4824]: I1006 10:15:56.208836 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/194bef17-98ac-413f-a91e-0abc9835f2f6-horizon-secret-key\") pod \"horizon-68fdc5dd48-gfq4d\" (UID: \"194bef17-98ac-413f-a91e-0abc9835f2f6\") " pod="openstack/horizon-68fdc5dd48-gfq4d" Oct 06 10:15:56 crc kubenswrapper[4824]: I1006 10:15:56.209486 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/194bef17-98ac-413f-a91e-0abc9835f2f6-logs\") pod \"horizon-68fdc5dd48-gfq4d\" (UID: \"194bef17-98ac-413f-a91e-0abc9835f2f6\") " pod="openstack/horizon-68fdc5dd48-gfq4d" Oct 06 10:15:56 crc kubenswrapper[4824]: I1006 10:15:56.209759 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/194bef17-98ac-413f-a91e-0abc9835f2f6-scripts\") pod \"horizon-68fdc5dd48-gfq4d\" (UID: \"194bef17-98ac-413f-a91e-0abc9835f2f6\") " pod="openstack/horizon-68fdc5dd48-gfq4d" Oct 06 10:15:56 crc kubenswrapper[4824]: I1006 10:15:56.210538 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/194bef17-98ac-413f-a91e-0abc9835f2f6-config-data\") pod \"horizon-68fdc5dd48-gfq4d\" (UID: \"194bef17-98ac-413f-a91e-0abc9835f2f6\") " pod="openstack/horizon-68fdc5dd48-gfq4d" Oct 06 10:15:56 crc kubenswrapper[4824]: I1006 10:15:56.214026 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/194bef17-98ac-413f-a91e-0abc9835f2f6-horizon-tls-certs\") pod \"horizon-68fdc5dd48-gfq4d\" (UID: \"194bef17-98ac-413f-a91e-0abc9835f2f6\") " pod="openstack/horizon-68fdc5dd48-gfq4d" Oct 06 10:15:56 crc kubenswrapper[4824]: I1006 10:15:56.214410 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/194bef17-98ac-413f-a91e-0abc9835f2f6-horizon-secret-key\") pod \"horizon-68fdc5dd48-gfq4d\" (UID: \"194bef17-98ac-413f-a91e-0abc9835f2f6\") " pod="openstack/horizon-68fdc5dd48-gfq4d" Oct 06 10:15:56 crc kubenswrapper[4824]: I1006 10:15:56.216093 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/194bef17-98ac-413f-a91e-0abc9835f2f6-combined-ca-bundle\") pod \"horizon-68fdc5dd48-gfq4d\" (UID: \"194bef17-98ac-413f-a91e-0abc9835f2f6\") " pod="openstack/horizon-68fdc5dd48-gfq4d" Oct 06 10:15:56 crc kubenswrapper[4824]: I1006 10:15:56.229265 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jqnzf\" (UniqueName: \"kubernetes.io/projected/194bef17-98ac-413f-a91e-0abc9835f2f6-kube-api-access-jqnzf\") pod \"horizon-68fdc5dd48-gfq4d\" (UID: \"194bef17-98ac-413f-a91e-0abc9835f2f6\") " pod="openstack/horizon-68fdc5dd48-gfq4d" Oct 06 10:15:56 crc kubenswrapper[4824]: I1006 10:15:56.231075 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-78c847f9f4-tv8zc" Oct 06 10:15:56 crc kubenswrapper[4824]: I1006 10:15:56.307865 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-68fdc5dd48-gfq4d" Oct 06 10:15:57 crc kubenswrapper[4824]: I1006 10:15:57.865136 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-8b5c85b87-9r9wx" Oct 06 10:15:57 crc kubenswrapper[4824]: I1006 10:15:57.981593 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-wtjf5"] Oct 06 10:15:57 crc kubenswrapper[4824]: I1006 10:15:57.981929 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-77585f5f8c-wtjf5" podUID="51ecd7bf-8b94-4c8c-af6e-a7e74e7f7c1c" containerName="dnsmasq-dns" containerID="cri-o://e17f227e1abc0d9b1729053104cdd67ed12db77ffc53e5658e7c5da09ebc3383" gracePeriod=10 Oct 06 10:15:58 crc kubenswrapper[4824]: I1006 10:15:58.073451 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-77585f5f8c-wtjf5" podUID="51ecd7bf-8b94-4c8c-af6e-a7e74e7f7c1c" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.133:5353: connect: connection refused" Oct 06 10:15:58 crc kubenswrapper[4824]: I1006 10:15:58.300940 4824 generic.go:334] "Generic (PLEG): container finished" podID="51ecd7bf-8b94-4c8c-af6e-a7e74e7f7c1c" containerID="e17f227e1abc0d9b1729053104cdd67ed12db77ffc53e5658e7c5da09ebc3383" exitCode=0 Oct 06 10:15:58 crc kubenswrapper[4824]: I1006 10:15:58.301297 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77585f5f8c-wtjf5" event={"ID":"51ecd7bf-8b94-4c8c-af6e-a7e74e7f7c1c","Type":"ContainerDied","Data":"e17f227e1abc0d9b1729053104cdd67ed12db77ffc53e5658e7c5da09ebc3383"} Oct 06 10:15:59 crc kubenswrapper[4824]: I1006 10:15:59.018772 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-zcq55" Oct 06 10:15:59 crc kubenswrapper[4824]: I1006 10:15:59.100935 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c607b17a-75ac-4e7f-8ff3-9d0edb23cc23-fernet-keys\") pod \"c607b17a-75ac-4e7f-8ff3-9d0edb23cc23\" (UID: \"c607b17a-75ac-4e7f-8ff3-9d0edb23cc23\") " Oct 06 10:15:59 crc kubenswrapper[4824]: I1006 10:15:59.102230 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c607b17a-75ac-4e7f-8ff3-9d0edb23cc23-credential-keys\") pod \"c607b17a-75ac-4e7f-8ff3-9d0edb23cc23\" (UID: \"c607b17a-75ac-4e7f-8ff3-9d0edb23cc23\") " Oct 06 10:15:59 crc kubenswrapper[4824]: I1006 10:15:59.102422 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c607b17a-75ac-4e7f-8ff3-9d0edb23cc23-scripts\") pod \"c607b17a-75ac-4e7f-8ff3-9d0edb23cc23\" (UID: \"c607b17a-75ac-4e7f-8ff3-9d0edb23cc23\") " Oct 06 10:15:59 crc kubenswrapper[4824]: I1006 10:15:59.102594 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c607b17a-75ac-4e7f-8ff3-9d0edb23cc23-combined-ca-bundle\") pod \"c607b17a-75ac-4e7f-8ff3-9d0edb23cc23\" (UID: \"c607b17a-75ac-4e7f-8ff3-9d0edb23cc23\") " Oct 06 10:15:59 crc kubenswrapper[4824]: I1006 10:15:59.104235 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c607b17a-75ac-4e7f-8ff3-9d0edb23cc23-config-data\") pod \"c607b17a-75ac-4e7f-8ff3-9d0edb23cc23\" (UID: \"c607b17a-75ac-4e7f-8ff3-9d0edb23cc23\") " Oct 06 10:15:59 crc kubenswrapper[4824]: I1006 10:15:59.104462 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b6msl\" (UniqueName: \"kubernetes.io/projected/c607b17a-75ac-4e7f-8ff3-9d0edb23cc23-kube-api-access-b6msl\") pod \"c607b17a-75ac-4e7f-8ff3-9d0edb23cc23\" (UID: \"c607b17a-75ac-4e7f-8ff3-9d0edb23cc23\") " Oct 06 10:15:59 crc kubenswrapper[4824]: I1006 10:15:59.113140 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c607b17a-75ac-4e7f-8ff3-9d0edb23cc23-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "c607b17a-75ac-4e7f-8ff3-9d0edb23cc23" (UID: "c607b17a-75ac-4e7f-8ff3-9d0edb23cc23"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:15:59 crc kubenswrapper[4824]: I1006 10:15:59.115952 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c607b17a-75ac-4e7f-8ff3-9d0edb23cc23-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "c607b17a-75ac-4e7f-8ff3-9d0edb23cc23" (UID: "c607b17a-75ac-4e7f-8ff3-9d0edb23cc23"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:15:59 crc kubenswrapper[4824]: I1006 10:15:59.122254 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c607b17a-75ac-4e7f-8ff3-9d0edb23cc23-scripts" (OuterVolumeSpecName: "scripts") pod "c607b17a-75ac-4e7f-8ff3-9d0edb23cc23" (UID: "c607b17a-75ac-4e7f-8ff3-9d0edb23cc23"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:15:59 crc kubenswrapper[4824]: I1006 10:15:59.126931 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c607b17a-75ac-4e7f-8ff3-9d0edb23cc23-kube-api-access-b6msl" (OuterVolumeSpecName: "kube-api-access-b6msl") pod "c607b17a-75ac-4e7f-8ff3-9d0edb23cc23" (UID: "c607b17a-75ac-4e7f-8ff3-9d0edb23cc23"). InnerVolumeSpecName "kube-api-access-b6msl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:15:59 crc kubenswrapper[4824]: I1006 10:15:59.145566 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c607b17a-75ac-4e7f-8ff3-9d0edb23cc23-config-data" (OuterVolumeSpecName: "config-data") pod "c607b17a-75ac-4e7f-8ff3-9d0edb23cc23" (UID: "c607b17a-75ac-4e7f-8ff3-9d0edb23cc23"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:15:59 crc kubenswrapper[4824]: I1006 10:15:59.156119 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c607b17a-75ac-4e7f-8ff3-9d0edb23cc23-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c607b17a-75ac-4e7f-8ff3-9d0edb23cc23" (UID: "c607b17a-75ac-4e7f-8ff3-9d0edb23cc23"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:15:59 crc kubenswrapper[4824]: I1006 10:15:59.207805 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b6msl\" (UniqueName: \"kubernetes.io/projected/c607b17a-75ac-4e7f-8ff3-9d0edb23cc23-kube-api-access-b6msl\") on node \"crc\" DevicePath \"\"" Oct 06 10:15:59 crc kubenswrapper[4824]: I1006 10:15:59.208278 4824 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c607b17a-75ac-4e7f-8ff3-9d0edb23cc23-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 06 10:15:59 crc kubenswrapper[4824]: I1006 10:15:59.208407 4824 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c607b17a-75ac-4e7f-8ff3-9d0edb23cc23-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 06 10:15:59 crc kubenswrapper[4824]: I1006 10:15:59.208482 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c607b17a-75ac-4e7f-8ff3-9d0edb23cc23-scripts\") on node \"crc\" DevicePath \"\"" Oct 06 10:15:59 crc kubenswrapper[4824]: I1006 10:15:59.208551 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c607b17a-75ac-4e7f-8ff3-9d0edb23cc23-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 10:15:59 crc kubenswrapper[4824]: I1006 10:15:59.208617 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c607b17a-75ac-4e7f-8ff3-9d0edb23cc23-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 10:15:59 crc kubenswrapper[4824]: I1006 10:15:59.323387 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-zcq55" event={"ID":"c607b17a-75ac-4e7f-8ff3-9d0edb23cc23","Type":"ContainerDied","Data":"b897b689cf8a55f703dd1ff5a5a59a7583dfeb25aa27812cda69985492549892"} Oct 06 10:15:59 crc kubenswrapper[4824]: I1006 10:15:59.323447 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b897b689cf8a55f703dd1ff5a5a59a7583dfeb25aa27812cda69985492549892" Oct 06 10:15:59 crc kubenswrapper[4824]: I1006 10:15:59.323475 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-zcq55" Oct 06 10:16:00 crc kubenswrapper[4824]: I1006 10:16:00.126404 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-zcq55"] Oct 06 10:16:00 crc kubenswrapper[4824]: I1006 10:16:00.133582 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-zcq55"] Oct 06 10:16:00 crc kubenswrapper[4824]: I1006 10:16:00.226016 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-wlmn6"] Oct 06 10:16:00 crc kubenswrapper[4824]: E1006 10:16:00.226591 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c607b17a-75ac-4e7f-8ff3-9d0edb23cc23" containerName="keystone-bootstrap" Oct 06 10:16:00 crc kubenswrapper[4824]: I1006 10:16:00.226614 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="c607b17a-75ac-4e7f-8ff3-9d0edb23cc23" containerName="keystone-bootstrap" Oct 06 10:16:00 crc kubenswrapper[4824]: I1006 10:16:00.226868 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="c607b17a-75ac-4e7f-8ff3-9d0edb23cc23" containerName="keystone-bootstrap" Oct 06 10:16:00 crc kubenswrapper[4824]: I1006 10:16:00.228082 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-wlmn6" Oct 06 10:16:00 crc kubenswrapper[4824]: I1006 10:16:00.230252 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-lxhxz" Oct 06 10:16:00 crc kubenswrapper[4824]: I1006 10:16:00.230360 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 06 10:16:00 crc kubenswrapper[4824]: I1006 10:16:00.230462 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 06 10:16:00 crc kubenswrapper[4824]: I1006 10:16:00.230613 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 06 10:16:00 crc kubenswrapper[4824]: I1006 10:16:00.243916 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-wlmn6"] Oct 06 10:16:00 crc kubenswrapper[4824]: I1006 10:16:00.331426 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tpvs6\" (UniqueName: \"kubernetes.io/projected/73fb825b-ea3b-43c5-b143-e70310f8c1fd-kube-api-access-tpvs6\") pod \"keystone-bootstrap-wlmn6\" (UID: \"73fb825b-ea3b-43c5-b143-e70310f8c1fd\") " pod="openstack/keystone-bootstrap-wlmn6" Oct 06 10:16:00 crc kubenswrapper[4824]: I1006 10:16:00.331486 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/73fb825b-ea3b-43c5-b143-e70310f8c1fd-credential-keys\") pod \"keystone-bootstrap-wlmn6\" (UID: \"73fb825b-ea3b-43c5-b143-e70310f8c1fd\") " pod="openstack/keystone-bootstrap-wlmn6" Oct 06 10:16:00 crc kubenswrapper[4824]: I1006 10:16:00.331550 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/73fb825b-ea3b-43c5-b143-e70310f8c1fd-scripts\") pod \"keystone-bootstrap-wlmn6\" (UID: \"73fb825b-ea3b-43c5-b143-e70310f8c1fd\") " pod="openstack/keystone-bootstrap-wlmn6" Oct 06 10:16:00 crc kubenswrapper[4824]: I1006 10:16:00.331584 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/73fb825b-ea3b-43c5-b143-e70310f8c1fd-config-data\") pod \"keystone-bootstrap-wlmn6\" (UID: \"73fb825b-ea3b-43c5-b143-e70310f8c1fd\") " pod="openstack/keystone-bootstrap-wlmn6" Oct 06 10:16:00 crc kubenswrapper[4824]: I1006 10:16:00.331608 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73fb825b-ea3b-43c5-b143-e70310f8c1fd-combined-ca-bundle\") pod \"keystone-bootstrap-wlmn6\" (UID: \"73fb825b-ea3b-43c5-b143-e70310f8c1fd\") " pod="openstack/keystone-bootstrap-wlmn6" Oct 06 10:16:00 crc kubenswrapper[4824]: I1006 10:16:00.331649 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/73fb825b-ea3b-43c5-b143-e70310f8c1fd-fernet-keys\") pod \"keystone-bootstrap-wlmn6\" (UID: \"73fb825b-ea3b-43c5-b143-e70310f8c1fd\") " pod="openstack/keystone-bootstrap-wlmn6" Oct 06 10:16:00 crc kubenswrapper[4824]: I1006 10:16:00.439300 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/73fb825b-ea3b-43c5-b143-e70310f8c1fd-credential-keys\") pod \"keystone-bootstrap-wlmn6\" (UID: \"73fb825b-ea3b-43c5-b143-e70310f8c1fd\") " pod="openstack/keystone-bootstrap-wlmn6" Oct 06 10:16:00 crc kubenswrapper[4824]: I1006 10:16:00.440904 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/73fb825b-ea3b-43c5-b143-e70310f8c1fd-scripts\") pod \"keystone-bootstrap-wlmn6\" (UID: \"73fb825b-ea3b-43c5-b143-e70310f8c1fd\") " pod="openstack/keystone-bootstrap-wlmn6" Oct 06 10:16:00 crc kubenswrapper[4824]: I1006 10:16:00.440960 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/73fb825b-ea3b-43c5-b143-e70310f8c1fd-config-data\") pod \"keystone-bootstrap-wlmn6\" (UID: \"73fb825b-ea3b-43c5-b143-e70310f8c1fd\") " pod="openstack/keystone-bootstrap-wlmn6" Oct 06 10:16:00 crc kubenswrapper[4824]: I1006 10:16:00.441014 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73fb825b-ea3b-43c5-b143-e70310f8c1fd-combined-ca-bundle\") pod \"keystone-bootstrap-wlmn6\" (UID: \"73fb825b-ea3b-43c5-b143-e70310f8c1fd\") " pod="openstack/keystone-bootstrap-wlmn6" Oct 06 10:16:00 crc kubenswrapper[4824]: I1006 10:16:00.441079 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/73fb825b-ea3b-43c5-b143-e70310f8c1fd-fernet-keys\") pod \"keystone-bootstrap-wlmn6\" (UID: \"73fb825b-ea3b-43c5-b143-e70310f8c1fd\") " pod="openstack/keystone-bootstrap-wlmn6" Oct 06 10:16:00 crc kubenswrapper[4824]: I1006 10:16:00.441148 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tpvs6\" (UniqueName: \"kubernetes.io/projected/73fb825b-ea3b-43c5-b143-e70310f8c1fd-kube-api-access-tpvs6\") pod \"keystone-bootstrap-wlmn6\" (UID: \"73fb825b-ea3b-43c5-b143-e70310f8c1fd\") " pod="openstack/keystone-bootstrap-wlmn6" Oct 06 10:16:00 crc kubenswrapper[4824]: I1006 10:16:00.447455 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/73fb825b-ea3b-43c5-b143-e70310f8c1fd-fernet-keys\") pod \"keystone-bootstrap-wlmn6\" (UID: \"73fb825b-ea3b-43c5-b143-e70310f8c1fd\") " pod="openstack/keystone-bootstrap-wlmn6" Oct 06 10:16:00 crc kubenswrapper[4824]: I1006 10:16:00.447682 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/73fb825b-ea3b-43c5-b143-e70310f8c1fd-credential-keys\") pod \"keystone-bootstrap-wlmn6\" (UID: \"73fb825b-ea3b-43c5-b143-e70310f8c1fd\") " pod="openstack/keystone-bootstrap-wlmn6" Oct 06 10:16:00 crc kubenswrapper[4824]: I1006 10:16:00.464403 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/73fb825b-ea3b-43c5-b143-e70310f8c1fd-scripts\") pod \"keystone-bootstrap-wlmn6\" (UID: \"73fb825b-ea3b-43c5-b143-e70310f8c1fd\") " pod="openstack/keystone-bootstrap-wlmn6" Oct 06 10:16:00 crc kubenswrapper[4824]: I1006 10:16:00.464432 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73fb825b-ea3b-43c5-b143-e70310f8c1fd-combined-ca-bundle\") pod \"keystone-bootstrap-wlmn6\" (UID: \"73fb825b-ea3b-43c5-b143-e70310f8c1fd\") " pod="openstack/keystone-bootstrap-wlmn6" Oct 06 10:16:00 crc kubenswrapper[4824]: I1006 10:16:00.467747 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tpvs6\" (UniqueName: \"kubernetes.io/projected/73fb825b-ea3b-43c5-b143-e70310f8c1fd-kube-api-access-tpvs6\") pod \"keystone-bootstrap-wlmn6\" (UID: \"73fb825b-ea3b-43c5-b143-e70310f8c1fd\") " pod="openstack/keystone-bootstrap-wlmn6" Oct 06 10:16:00 crc kubenswrapper[4824]: I1006 10:16:00.475552 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/73fb825b-ea3b-43c5-b143-e70310f8c1fd-config-data\") pod \"keystone-bootstrap-wlmn6\" (UID: \"73fb825b-ea3b-43c5-b143-e70310f8c1fd\") " pod="openstack/keystone-bootstrap-wlmn6" Oct 06 10:16:00 crc kubenswrapper[4824]: I1006 10:16:00.571595 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-wlmn6" Oct 06 10:16:01 crc kubenswrapper[4824]: I1006 10:16:01.285994 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c607b17a-75ac-4e7f-8ff3-9d0edb23cc23" path="/var/lib/kubelet/pods/c607b17a-75ac-4e7f-8ff3-9d0edb23cc23/volumes" Oct 06 10:16:04 crc kubenswrapper[4824]: E1006 10:16:04.442144 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Oct 06 10:16:04 crc kubenswrapper[4824]: E1006 10:16:04.442964 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n5dch645h5f4h67h648h68bh57ch566hbch5cch574h5dch566h559h565h649h574h66ch56dh57ch6fh9dh655hb9h554h578h568h598hc4h5d7h665h5d8q,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-prrmr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-ccd5f54d9-dsnj5_openstack(3d28a180-78e5-4977-bccb-9210f5fb0a7f): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 06 10:16:04 crc kubenswrapper[4824]: E1006 10:16:04.444476 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Oct 06 10:16:04 crc kubenswrapper[4824]: E1006 10:16:04.444578 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n5ddh7h579h549h5bhbdh649hc6hb5h665h65fh5cch8dhbdh549hdfh568h8bh679h5f6h554h588hbch8fh654hbchfh55bh54chd4h67hc8q,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-vt2nw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-7749656757-6jdvp_openstack(07c13b84-6c2a-4a6c-b23b-df1e14e077b5): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 06 10:16:04 crc kubenswrapper[4824]: E1006 10:16:04.446025 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-ccd5f54d9-dsnj5" podUID="3d28a180-78e5-4977-bccb-9210f5fb0a7f" Oct 06 10:16:04 crc kubenswrapper[4824]: E1006 10:16:04.446751 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-7749656757-6jdvp" podUID="07c13b84-6c2a-4a6c-b23b-df1e14e077b5" Oct 06 10:16:06 crc kubenswrapper[4824]: E1006 10:16:06.958891 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Oct 06 10:16:06 crc kubenswrapper[4824]: E1006 10:16:06.959328 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n58dh54ch57bh557h589h655h98h5b7h5dfhc4hcbhfdh5fbh546h644h698h65fh666h5bhcdh64ch547h644h588h68fh665h677h59chd9h599h677h8dq,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-frht4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-5bb95d74fc-r95ns_openstack(541d798b-323f-48c4-b0c1-a8e9d4333f6c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 06 10:16:06 crc kubenswrapper[4824]: E1006 10:16:06.961609 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-5bb95d74fc-r95ns" podUID="541d798b-323f-48c4-b0c1-a8e9d4333f6c" Oct 06 10:16:08 crc kubenswrapper[4824]: I1006 10:16:08.073322 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-77585f5f8c-wtjf5" podUID="51ecd7bf-8b94-4c8c-af6e-a7e74e7f7c1c" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.133:5353: i/o timeout" Oct 06 10:16:08 crc kubenswrapper[4824]: I1006 10:16:08.421048 4824 generic.go:334] "Generic (PLEG): container finished" podID="3d78bdee-7130-48a9-ad24-6c8c6525784c" containerID="86eeda4af9d4a8081ccdcc7a0cceaeae21a27abc73451b614b28ba15c64ed645" exitCode=0 Oct 06 10:16:08 crc kubenswrapper[4824]: I1006 10:16:08.421104 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-xsgsg" event={"ID":"3d78bdee-7130-48a9-ad24-6c8c6525784c","Type":"ContainerDied","Data":"86eeda4af9d4a8081ccdcc7a0cceaeae21a27abc73451b614b28ba15c64ed645"} Oct 06 10:16:13 crc kubenswrapper[4824]: I1006 10:16:13.074719 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-77585f5f8c-wtjf5" podUID="51ecd7bf-8b94-4c8c-af6e-a7e74e7f7c1c" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.133:5353: i/o timeout" Oct 06 10:16:13 crc kubenswrapper[4824]: I1006 10:16:13.075706 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-77585f5f8c-wtjf5" Oct 06 10:16:15 crc kubenswrapper[4824]: E1006 10:16:15.413648 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified" Oct 06 10:16:15 crc kubenswrapper[4824]: E1006 10:16:15.413789 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ceilometer-central-agent,Image:quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nbch55bh5b8hbchbbh664h648h5cfh57dh84h5fbh695h69h5fh58bh675h664hd8h5cdh5c9h588hffhd5h5d8h5bdh55h565hcch9hf4h5fch5f8q,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:ceilometer-central-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-w6rqx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/python3 /var/lib/openstack/bin/centralhealth.py],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(13781991-323c-49ae-9c5d-625e2ffd019b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 06 10:16:15 crc kubenswrapper[4824]: I1006 10:16:15.510897 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"2e26a57e-6683-48b2-a0ea-3f8ff1027fdc","Type":"ContainerDied","Data":"a28da5c24c055bfee9835c1c721a89e81c17ace9200bd8211d363675b148b925"} Oct 06 10:16:15 crc kubenswrapper[4824]: I1006 10:16:15.510949 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a28da5c24c055bfee9835c1c721a89e81c17ace9200bd8211d363675b148b925" Oct 06 10:16:15 crc kubenswrapper[4824]: I1006 10:16:15.514285 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-ccd5f54d9-dsnj5" event={"ID":"3d28a180-78e5-4977-bccb-9210f5fb0a7f","Type":"ContainerDied","Data":"4122d854ffac7724984a0f667d6a5bc1650958517bb6edef3a883540de1a3236"} Oct 06 10:16:15 crc kubenswrapper[4824]: I1006 10:16:15.514382 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4122d854ffac7724984a0f667d6a5bc1650958517bb6edef3a883540de1a3236" Oct 06 10:16:15 crc kubenswrapper[4824]: I1006 10:16:15.516941 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77585f5f8c-wtjf5" event={"ID":"51ecd7bf-8b94-4c8c-af6e-a7e74e7f7c1c","Type":"ContainerDied","Data":"11d534cf0352c0c2d7dfea22ab9090584ee138420f40dc3e46ac9ddacd83f120"} Oct 06 10:16:15 crc kubenswrapper[4824]: I1006 10:16:15.517015 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="11d534cf0352c0c2d7dfea22ab9090584ee138420f40dc3e46ac9ddacd83f120" Oct 06 10:16:15 crc kubenswrapper[4824]: I1006 10:16:15.519074 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"97b2f7d6-3fd8-43ab-85b6-a3d37e218887","Type":"ContainerDied","Data":"927b0e04773f5ccabdce43dca2c18d30b75f19cf4fe15f6ca45c98b7944307b7"} Oct 06 10:16:15 crc kubenswrapper[4824]: I1006 10:16:15.519176 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="927b0e04773f5ccabdce43dca2c18d30b75f19cf4fe15f6ca45c98b7944307b7" Oct 06 10:16:15 crc kubenswrapper[4824]: I1006 10:16:15.520211 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7749656757-6jdvp" event={"ID":"07c13b84-6c2a-4a6c-b23b-df1e14e077b5","Type":"ContainerDied","Data":"73b4b5608cf04aa78bc07acbd6029d18e51c22f4afed2bfc0c560650ddbf00aa"} Oct 06 10:16:15 crc kubenswrapper[4824]: I1006 10:16:15.520260 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="73b4b5608cf04aa78bc07acbd6029d18e51c22f4afed2bfc0c560650ddbf00aa" Oct 06 10:16:15 crc kubenswrapper[4824]: I1006 10:16:15.567937 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77585f5f8c-wtjf5" Oct 06 10:16:15 crc kubenswrapper[4824]: I1006 10:16:15.577455 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 06 10:16:15 crc kubenswrapper[4824]: I1006 10:16:15.582501 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 06 10:16:15 crc kubenswrapper[4824]: I1006 10:16:15.611090 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2e26a57e-6683-48b2-a0ea-3f8ff1027fdc-logs\") pod \"2e26a57e-6683-48b2-a0ea-3f8ff1027fdc\" (UID: \"2e26a57e-6683-48b2-a0ea-3f8ff1027fdc\") " Oct 06 10:16:15 crc kubenswrapper[4824]: I1006 10:16:15.611148 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e26a57e-6683-48b2-a0ea-3f8ff1027fdc-config-data\") pod \"2e26a57e-6683-48b2-a0ea-3f8ff1027fdc\" (UID: \"2e26a57e-6683-48b2-a0ea-3f8ff1027fdc\") " Oct 06 10:16:15 crc kubenswrapper[4824]: I1006 10:16:15.611177 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e26a57e-6683-48b2-a0ea-3f8ff1027fdc-combined-ca-bundle\") pod \"2e26a57e-6683-48b2-a0ea-3f8ff1027fdc\" (UID: \"2e26a57e-6683-48b2-a0ea-3f8ff1027fdc\") " Oct 06 10:16:15 crc kubenswrapper[4824]: I1006 10:16:15.611200 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"97b2f7d6-3fd8-43ab-85b6-a3d37e218887\" (UID: \"97b2f7d6-3fd8-43ab-85b6-a3d37e218887\") " Oct 06 10:16:15 crc kubenswrapper[4824]: I1006 10:16:15.612796 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2e26a57e-6683-48b2-a0ea-3f8ff1027fdc-logs" (OuterVolumeSpecName: "logs") pod "2e26a57e-6683-48b2-a0ea-3f8ff1027fdc" (UID: "2e26a57e-6683-48b2-a0ea-3f8ff1027fdc"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:16:15 crc kubenswrapper[4824]: I1006 10:16:15.613055 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/51ecd7bf-8b94-4c8c-af6e-a7e74e7f7c1c-dns-swift-storage-0\") pod \"51ecd7bf-8b94-4c8c-af6e-a7e74e7f7c1c\" (UID: \"51ecd7bf-8b94-4c8c-af6e-a7e74e7f7c1c\") " Oct 06 10:16:15 crc kubenswrapper[4824]: I1006 10:16:15.613159 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dh64c\" (UniqueName: \"kubernetes.io/projected/97b2f7d6-3fd8-43ab-85b6-a3d37e218887-kube-api-access-dh64c\") pod \"97b2f7d6-3fd8-43ab-85b6-a3d37e218887\" (UID: \"97b2f7d6-3fd8-43ab-85b6-a3d37e218887\") " Oct 06 10:16:15 crc kubenswrapper[4824]: I1006 10:16:15.613213 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dl6mz\" (UniqueName: \"kubernetes.io/projected/51ecd7bf-8b94-4c8c-af6e-a7e74e7f7c1c-kube-api-access-dl6mz\") pod \"51ecd7bf-8b94-4c8c-af6e-a7e74e7f7c1c\" (UID: \"51ecd7bf-8b94-4c8c-af6e-a7e74e7f7c1c\") " Oct 06 10:16:15 crc kubenswrapper[4824]: I1006 10:16:15.613249 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/97b2f7d6-3fd8-43ab-85b6-a3d37e218887-config-data\") pod \"97b2f7d6-3fd8-43ab-85b6-a3d37e218887\" (UID: \"97b2f7d6-3fd8-43ab-85b6-a3d37e218887\") " Oct 06 10:16:15 crc kubenswrapper[4824]: I1006 10:16:15.613337 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/51ecd7bf-8b94-4c8c-af6e-a7e74e7f7c1c-config\") pod \"51ecd7bf-8b94-4c8c-af6e-a7e74e7f7c1c\" (UID: \"51ecd7bf-8b94-4c8c-af6e-a7e74e7f7c1c\") " Oct 06 10:16:15 crc kubenswrapper[4824]: I1006 10:16:15.613386 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nlv4q\" (UniqueName: \"kubernetes.io/projected/2e26a57e-6683-48b2-a0ea-3f8ff1027fdc-kube-api-access-nlv4q\") pod \"2e26a57e-6683-48b2-a0ea-3f8ff1027fdc\" (UID: \"2e26a57e-6683-48b2-a0ea-3f8ff1027fdc\") " Oct 06 10:16:15 crc kubenswrapper[4824]: I1006 10:16:15.613427 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/97b2f7d6-3fd8-43ab-85b6-a3d37e218887-scripts\") pod \"97b2f7d6-3fd8-43ab-85b6-a3d37e218887\" (UID: \"97b2f7d6-3fd8-43ab-85b6-a3d37e218887\") " Oct 06 10:16:15 crc kubenswrapper[4824]: I1006 10:16:15.613487 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/97b2f7d6-3fd8-43ab-85b6-a3d37e218887-internal-tls-certs\") pod \"97b2f7d6-3fd8-43ab-85b6-a3d37e218887\" (UID: \"97b2f7d6-3fd8-43ab-85b6-a3d37e218887\") " Oct 06 10:16:15 crc kubenswrapper[4824]: I1006 10:16:15.613539 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2e26a57e-6683-48b2-a0ea-3f8ff1027fdc-scripts\") pod \"2e26a57e-6683-48b2-a0ea-3f8ff1027fdc\" (UID: \"2e26a57e-6683-48b2-a0ea-3f8ff1027fdc\") " Oct 06 10:16:15 crc kubenswrapper[4824]: I1006 10:16:15.613649 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2e26a57e-6683-48b2-a0ea-3f8ff1027fdc-httpd-run\") pod \"2e26a57e-6683-48b2-a0ea-3f8ff1027fdc\" (UID: \"2e26a57e-6683-48b2-a0ea-3f8ff1027fdc\") " Oct 06 10:16:15 crc kubenswrapper[4824]: I1006 10:16:15.613698 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/51ecd7bf-8b94-4c8c-af6e-a7e74e7f7c1c-dns-svc\") pod \"51ecd7bf-8b94-4c8c-af6e-a7e74e7f7c1c\" (UID: \"51ecd7bf-8b94-4c8c-af6e-a7e74e7f7c1c\") " Oct 06 10:16:15 crc kubenswrapper[4824]: I1006 10:16:15.613744 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e26a57e-6683-48b2-a0ea-3f8ff1027fdc-public-tls-certs\") pod \"2e26a57e-6683-48b2-a0ea-3f8ff1027fdc\" (UID: \"2e26a57e-6683-48b2-a0ea-3f8ff1027fdc\") " Oct 06 10:16:15 crc kubenswrapper[4824]: I1006 10:16:15.613789 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97b2f7d6-3fd8-43ab-85b6-a3d37e218887-combined-ca-bundle\") pod \"97b2f7d6-3fd8-43ab-85b6-a3d37e218887\" (UID: \"97b2f7d6-3fd8-43ab-85b6-a3d37e218887\") " Oct 06 10:16:15 crc kubenswrapper[4824]: I1006 10:16:15.613818 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"2e26a57e-6683-48b2-a0ea-3f8ff1027fdc\" (UID: \"2e26a57e-6683-48b2-a0ea-3f8ff1027fdc\") " Oct 06 10:16:15 crc kubenswrapper[4824]: I1006 10:16:15.613874 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/97b2f7d6-3fd8-43ab-85b6-a3d37e218887-httpd-run\") pod \"97b2f7d6-3fd8-43ab-85b6-a3d37e218887\" (UID: \"97b2f7d6-3fd8-43ab-85b6-a3d37e218887\") " Oct 06 10:16:15 crc kubenswrapper[4824]: I1006 10:16:15.613915 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/97b2f7d6-3fd8-43ab-85b6-a3d37e218887-logs\") pod \"97b2f7d6-3fd8-43ab-85b6-a3d37e218887\" (UID: \"97b2f7d6-3fd8-43ab-85b6-a3d37e218887\") " Oct 06 10:16:15 crc kubenswrapper[4824]: I1006 10:16:15.613948 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/51ecd7bf-8b94-4c8c-af6e-a7e74e7f7c1c-ovsdbserver-nb\") pod \"51ecd7bf-8b94-4c8c-af6e-a7e74e7f7c1c\" (UID: \"51ecd7bf-8b94-4c8c-af6e-a7e74e7f7c1c\") " Oct 06 10:16:15 crc kubenswrapper[4824]: I1006 10:16:15.614013 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/51ecd7bf-8b94-4c8c-af6e-a7e74e7f7c1c-ovsdbserver-sb\") pod \"51ecd7bf-8b94-4c8c-af6e-a7e74e7f7c1c\" (UID: \"51ecd7bf-8b94-4c8c-af6e-a7e74e7f7c1c\") " Oct 06 10:16:15 crc kubenswrapper[4824]: I1006 10:16:15.614788 4824 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2e26a57e-6683-48b2-a0ea-3f8ff1027fdc-logs\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:15 crc kubenswrapper[4824]: I1006 10:16:15.631878 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "glance") pod "97b2f7d6-3fd8-43ab-85b6-a3d37e218887" (UID: "97b2f7d6-3fd8-43ab-85b6-a3d37e218887"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 06 10:16:15 crc kubenswrapper[4824]: I1006 10:16:15.632664 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e26a57e-6683-48b2-a0ea-3f8ff1027fdc-scripts" (OuterVolumeSpecName: "scripts") pod "2e26a57e-6683-48b2-a0ea-3f8ff1027fdc" (UID: "2e26a57e-6683-48b2-a0ea-3f8ff1027fdc"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:16:15 crc kubenswrapper[4824]: I1006 10:16:15.637329 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2e26a57e-6683-48b2-a0ea-3f8ff1027fdc-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "2e26a57e-6683-48b2-a0ea-3f8ff1027fdc" (UID: "2e26a57e-6683-48b2-a0ea-3f8ff1027fdc"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:16:15 crc kubenswrapper[4824]: I1006 10:16:15.651197 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/97b2f7d6-3fd8-43ab-85b6-a3d37e218887-logs" (OuterVolumeSpecName: "logs") pod "97b2f7d6-3fd8-43ab-85b6-a3d37e218887" (UID: "97b2f7d6-3fd8-43ab-85b6-a3d37e218887"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:16:15 crc kubenswrapper[4824]: I1006 10:16:15.665214 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "glance") pod "2e26a57e-6683-48b2-a0ea-3f8ff1027fdc" (UID: "2e26a57e-6683-48b2-a0ea-3f8ff1027fdc"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 06 10:16:15 crc kubenswrapper[4824]: I1006 10:16:15.689792 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/97b2f7d6-3fd8-43ab-85b6-a3d37e218887-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "97b2f7d6-3fd8-43ab-85b6-a3d37e218887" (UID: "97b2f7d6-3fd8-43ab-85b6-a3d37e218887"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:16:15 crc kubenswrapper[4824]: I1006 10:16:15.689961 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2e26a57e-6683-48b2-a0ea-3f8ff1027fdc-kube-api-access-nlv4q" (OuterVolumeSpecName: "kube-api-access-nlv4q") pod "2e26a57e-6683-48b2-a0ea-3f8ff1027fdc" (UID: "2e26a57e-6683-48b2-a0ea-3f8ff1027fdc"). InnerVolumeSpecName "kube-api-access-nlv4q". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:16:15 crc kubenswrapper[4824]: I1006 10:16:15.690059 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/51ecd7bf-8b94-4c8c-af6e-a7e74e7f7c1c-kube-api-access-dl6mz" (OuterVolumeSpecName: "kube-api-access-dl6mz") pod "51ecd7bf-8b94-4c8c-af6e-a7e74e7f7c1c" (UID: "51ecd7bf-8b94-4c8c-af6e-a7e74e7f7c1c"). InnerVolumeSpecName "kube-api-access-dl6mz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:16:15 crc kubenswrapper[4824]: I1006 10:16:15.718370 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2e26a57e-6683-48b2-a0ea-3f8ff1027fdc-scripts\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:15 crc kubenswrapper[4824]: I1006 10:16:15.718430 4824 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2e26a57e-6683-48b2-a0ea-3f8ff1027fdc-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:15 crc kubenswrapper[4824]: I1006 10:16:15.718470 4824 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Oct 06 10:16:15 crc kubenswrapper[4824]: I1006 10:16:15.718522 4824 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/97b2f7d6-3fd8-43ab-85b6-a3d37e218887-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:15 crc kubenswrapper[4824]: I1006 10:16:15.718536 4824 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/97b2f7d6-3fd8-43ab-85b6-a3d37e218887-logs\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:15 crc kubenswrapper[4824]: I1006 10:16:15.718560 4824 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Oct 06 10:16:15 crc kubenswrapper[4824]: I1006 10:16:15.718578 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dl6mz\" (UniqueName: \"kubernetes.io/projected/51ecd7bf-8b94-4c8c-af6e-a7e74e7f7c1c-kube-api-access-dl6mz\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:15 crc kubenswrapper[4824]: I1006 10:16:15.718602 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nlv4q\" (UniqueName: \"kubernetes.io/projected/2e26a57e-6683-48b2-a0ea-3f8ff1027fdc-kube-api-access-nlv4q\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:15 crc kubenswrapper[4824]: I1006 10:16:15.719212 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/97b2f7d6-3fd8-43ab-85b6-a3d37e218887-scripts" (OuterVolumeSpecName: "scripts") pod "97b2f7d6-3fd8-43ab-85b6-a3d37e218887" (UID: "97b2f7d6-3fd8-43ab-85b6-a3d37e218887"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:16:15 crc kubenswrapper[4824]: I1006 10:16:15.732408 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-ccd5f54d9-dsnj5" Oct 06 10:16:15 crc kubenswrapper[4824]: I1006 10:16:15.793285 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/97b2f7d6-3fd8-43ab-85b6-a3d37e218887-kube-api-access-dh64c" (OuterVolumeSpecName: "kube-api-access-dh64c") pod "97b2f7d6-3fd8-43ab-85b6-a3d37e218887" (UID: "97b2f7d6-3fd8-43ab-85b6-a3d37e218887"). InnerVolumeSpecName "kube-api-access-dh64c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:16:15 crc kubenswrapper[4824]: I1006 10:16:15.824484 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-prrmr\" (UniqueName: \"kubernetes.io/projected/3d28a180-78e5-4977-bccb-9210f5fb0a7f-kube-api-access-prrmr\") pod \"3d28a180-78e5-4977-bccb-9210f5fb0a7f\" (UID: \"3d28a180-78e5-4977-bccb-9210f5fb0a7f\") " Oct 06 10:16:15 crc kubenswrapper[4824]: I1006 10:16:15.824554 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3d28a180-78e5-4977-bccb-9210f5fb0a7f-scripts\") pod \"3d28a180-78e5-4977-bccb-9210f5fb0a7f\" (UID: \"3d28a180-78e5-4977-bccb-9210f5fb0a7f\") " Oct 06 10:16:15 crc kubenswrapper[4824]: I1006 10:16:15.824657 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3d28a180-78e5-4977-bccb-9210f5fb0a7f-config-data\") pod \"3d28a180-78e5-4977-bccb-9210f5fb0a7f\" (UID: \"3d28a180-78e5-4977-bccb-9210f5fb0a7f\") " Oct 06 10:16:15 crc kubenswrapper[4824]: I1006 10:16:15.824893 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3d28a180-78e5-4977-bccb-9210f5fb0a7f-logs\") pod \"3d28a180-78e5-4977-bccb-9210f5fb0a7f\" (UID: \"3d28a180-78e5-4977-bccb-9210f5fb0a7f\") " Oct 06 10:16:15 crc kubenswrapper[4824]: I1006 10:16:15.824953 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/3d28a180-78e5-4977-bccb-9210f5fb0a7f-horizon-secret-key\") pod \"3d28a180-78e5-4977-bccb-9210f5fb0a7f\" (UID: \"3d28a180-78e5-4977-bccb-9210f5fb0a7f\") " Oct 06 10:16:15 crc kubenswrapper[4824]: I1006 10:16:15.825365 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dh64c\" (UniqueName: \"kubernetes.io/projected/97b2f7d6-3fd8-43ab-85b6-a3d37e218887-kube-api-access-dh64c\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:15 crc kubenswrapper[4824]: I1006 10:16:15.825399 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/97b2f7d6-3fd8-43ab-85b6-a3d37e218887-scripts\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:15 crc kubenswrapper[4824]: I1006 10:16:15.836213 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3d28a180-78e5-4977-bccb-9210f5fb0a7f-scripts" (OuterVolumeSpecName: "scripts") pod "3d28a180-78e5-4977-bccb-9210f5fb0a7f" (UID: "3d28a180-78e5-4977-bccb-9210f5fb0a7f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:16:15 crc kubenswrapper[4824]: I1006 10:16:15.836289 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3d28a180-78e5-4977-bccb-9210f5fb0a7f-config-data" (OuterVolumeSpecName: "config-data") pod "3d28a180-78e5-4977-bccb-9210f5fb0a7f" (UID: "3d28a180-78e5-4977-bccb-9210f5fb0a7f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:16:15 crc kubenswrapper[4824]: I1006 10:16:15.836570 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3d28a180-78e5-4977-bccb-9210f5fb0a7f-logs" (OuterVolumeSpecName: "logs") pod "3d28a180-78e5-4977-bccb-9210f5fb0a7f" (UID: "3d28a180-78e5-4977-bccb-9210f5fb0a7f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:16:15 crc kubenswrapper[4824]: I1006 10:16:15.851861 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e26a57e-6683-48b2-a0ea-3f8ff1027fdc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2e26a57e-6683-48b2-a0ea-3f8ff1027fdc" (UID: "2e26a57e-6683-48b2-a0ea-3f8ff1027fdc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:16:15 crc kubenswrapper[4824]: I1006 10:16:15.859924 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3d28a180-78e5-4977-bccb-9210f5fb0a7f-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "3d28a180-78e5-4977-bccb-9210f5fb0a7f" (UID: "3d28a180-78e5-4977-bccb-9210f5fb0a7f"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:16:15 crc kubenswrapper[4824]: I1006 10:16:15.876963 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3d28a180-78e5-4977-bccb-9210f5fb0a7f-kube-api-access-prrmr" (OuterVolumeSpecName: "kube-api-access-prrmr") pod "3d28a180-78e5-4977-bccb-9210f5fb0a7f" (UID: "3d28a180-78e5-4977-bccb-9210f5fb0a7f"). InnerVolumeSpecName "kube-api-access-prrmr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:16:15 crc kubenswrapper[4824]: I1006 10:16:15.927422 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3d28a180-78e5-4977-bccb-9210f5fb0a7f-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:15 crc kubenswrapper[4824]: I1006 10:16:15.927456 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e26a57e-6683-48b2-a0ea-3f8ff1027fdc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:15 crc kubenswrapper[4824]: I1006 10:16:15.927470 4824 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3d28a180-78e5-4977-bccb-9210f5fb0a7f-logs\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:15 crc kubenswrapper[4824]: I1006 10:16:15.927480 4824 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/3d28a180-78e5-4977-bccb-9210f5fb0a7f-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:15 crc kubenswrapper[4824]: I1006 10:16:15.927490 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-prrmr\" (UniqueName: \"kubernetes.io/projected/3d28a180-78e5-4977-bccb-9210f5fb0a7f-kube-api-access-prrmr\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:15 crc kubenswrapper[4824]: I1006 10:16:15.927500 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3d28a180-78e5-4977-bccb-9210f5fb0a7f-scripts\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:15 crc kubenswrapper[4824]: I1006 10:16:15.930033 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7749656757-6jdvp" Oct 06 10:16:15 crc kubenswrapper[4824]: I1006 10:16:15.932429 4824 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Oct 06 10:16:15 crc kubenswrapper[4824]: I1006 10:16:15.932931 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/97b2f7d6-3fd8-43ab-85b6-a3d37e218887-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "97b2f7d6-3fd8-43ab-85b6-a3d37e218887" (UID: "97b2f7d6-3fd8-43ab-85b6-a3d37e218887"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:16:15 crc kubenswrapper[4824]: I1006 10:16:15.937351 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e26a57e-6683-48b2-a0ea-3f8ff1027fdc-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "2e26a57e-6683-48b2-a0ea-3f8ff1027fdc" (UID: "2e26a57e-6683-48b2-a0ea-3f8ff1027fdc"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:16:15 crc kubenswrapper[4824]: I1006 10:16:15.943101 4824 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Oct 06 10:16:15 crc kubenswrapper[4824]: I1006 10:16:15.946706 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/97b2f7d6-3fd8-43ab-85b6-a3d37e218887-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "97b2f7d6-3fd8-43ab-85b6-a3d37e218887" (UID: "97b2f7d6-3fd8-43ab-85b6-a3d37e218887"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:16:15 crc kubenswrapper[4824]: I1006 10:16:15.948208 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/51ecd7bf-8b94-4c8c-af6e-a7e74e7f7c1c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "51ecd7bf-8b94-4c8c-af6e-a7e74e7f7c1c" (UID: "51ecd7bf-8b94-4c8c-af6e-a7e74e7f7c1c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:16:15 crc kubenswrapper[4824]: I1006 10:16:15.956948 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/51ecd7bf-8b94-4c8c-af6e-a7e74e7f7c1c-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "51ecd7bf-8b94-4c8c-af6e-a7e74e7f7c1c" (UID: "51ecd7bf-8b94-4c8c-af6e-a7e74e7f7c1c"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:16:15 crc kubenswrapper[4824]: I1006 10:16:15.973454 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/51ecd7bf-8b94-4c8c-af6e-a7e74e7f7c1c-config" (OuterVolumeSpecName: "config") pod "51ecd7bf-8b94-4c8c-af6e-a7e74e7f7c1c" (UID: "51ecd7bf-8b94-4c8c-af6e-a7e74e7f7c1c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:16:15 crc kubenswrapper[4824]: I1006 10:16:15.976862 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e26a57e-6683-48b2-a0ea-3f8ff1027fdc-config-data" (OuterVolumeSpecName: "config-data") pod "2e26a57e-6683-48b2-a0ea-3f8ff1027fdc" (UID: "2e26a57e-6683-48b2-a0ea-3f8ff1027fdc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:16:15 crc kubenswrapper[4824]: I1006 10:16:15.982764 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/51ecd7bf-8b94-4c8c-af6e-a7e74e7f7c1c-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "51ecd7bf-8b94-4c8c-af6e-a7e74e7f7c1c" (UID: "51ecd7bf-8b94-4c8c-af6e-a7e74e7f7c1c"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:16:15 crc kubenswrapper[4824]: I1006 10:16:15.983947 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/51ecd7bf-8b94-4c8c-af6e-a7e74e7f7c1c-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "51ecd7bf-8b94-4c8c-af6e-a7e74e7f7c1c" (UID: "51ecd7bf-8b94-4c8c-af6e-a7e74e7f7c1c"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:16:15 crc kubenswrapper[4824]: I1006 10:16:15.997168 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/97b2f7d6-3fd8-43ab-85b6-a3d37e218887-config-data" (OuterVolumeSpecName: "config-data") pod "97b2f7d6-3fd8-43ab-85b6-a3d37e218887" (UID: "97b2f7d6-3fd8-43ab-85b6-a3d37e218887"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:16:16 crc kubenswrapper[4824]: I1006 10:16:16.028212 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt2nw\" (UniqueName: \"kubernetes.io/projected/07c13b84-6c2a-4a6c-b23b-df1e14e077b5-kube-api-access-vt2nw\") pod \"07c13b84-6c2a-4a6c-b23b-df1e14e077b5\" (UID: \"07c13b84-6c2a-4a6c-b23b-df1e14e077b5\") " Oct 06 10:16:16 crc kubenswrapper[4824]: I1006 10:16:16.028288 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/07c13b84-6c2a-4a6c-b23b-df1e14e077b5-horizon-secret-key\") pod \"07c13b84-6c2a-4a6c-b23b-df1e14e077b5\" (UID: \"07c13b84-6c2a-4a6c-b23b-df1e14e077b5\") " Oct 06 10:16:16 crc kubenswrapper[4824]: I1006 10:16:16.028397 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/07c13b84-6c2a-4a6c-b23b-df1e14e077b5-logs\") pod \"07c13b84-6c2a-4a6c-b23b-df1e14e077b5\" (UID: \"07c13b84-6c2a-4a6c-b23b-df1e14e077b5\") " Oct 06 10:16:16 crc kubenswrapper[4824]: I1006 10:16:16.028455 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/07c13b84-6c2a-4a6c-b23b-df1e14e077b5-scripts\") pod \"07c13b84-6c2a-4a6c-b23b-df1e14e077b5\" (UID: \"07c13b84-6c2a-4a6c-b23b-df1e14e077b5\") " Oct 06 10:16:16 crc kubenswrapper[4824]: I1006 10:16:16.028488 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/07c13b84-6c2a-4a6c-b23b-df1e14e077b5-config-data\") pod \"07c13b84-6c2a-4a6c-b23b-df1e14e077b5\" (UID: \"07c13b84-6c2a-4a6c-b23b-df1e14e077b5\") " Oct 06 10:16:16 crc kubenswrapper[4824]: I1006 10:16:16.028896 4824 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/51ecd7bf-8b94-4c8c-af6e-a7e74e7f7c1c-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:16 crc kubenswrapper[4824]: I1006 10:16:16.028920 4824 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e26a57e-6683-48b2-a0ea-3f8ff1027fdc-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:16 crc kubenswrapper[4824]: I1006 10:16:16.028934 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97b2f7d6-3fd8-43ab-85b6-a3d37e218887-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:16 crc kubenswrapper[4824]: I1006 10:16:16.028945 4824 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:16 crc kubenswrapper[4824]: I1006 10:16:16.028955 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/51ecd7bf-8b94-4c8c-af6e-a7e74e7f7c1c-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:16 crc kubenswrapper[4824]: I1006 10:16:16.028966 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/51ecd7bf-8b94-4c8c-af6e-a7e74e7f7c1c-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:16 crc kubenswrapper[4824]: I1006 10:16:16.029063 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e26a57e-6683-48b2-a0ea-3f8ff1027fdc-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:16 crc kubenswrapper[4824]: I1006 10:16:16.029079 4824 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:16 crc kubenswrapper[4824]: I1006 10:16:16.029090 4824 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/51ecd7bf-8b94-4c8c-af6e-a7e74e7f7c1c-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:16 crc kubenswrapper[4824]: I1006 10:16:16.029103 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/97b2f7d6-3fd8-43ab-85b6-a3d37e218887-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:16 crc kubenswrapper[4824]: I1006 10:16:16.029113 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/51ecd7bf-8b94-4c8c-af6e-a7e74e7f7c1c-config\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:16 crc kubenswrapper[4824]: I1006 10:16:16.029126 4824 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/97b2f7d6-3fd8-43ab-85b6-a3d37e218887-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:16 crc kubenswrapper[4824]: I1006 10:16:16.029858 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/07c13b84-6c2a-4a6c-b23b-df1e14e077b5-config-data" (OuterVolumeSpecName: "config-data") pod "07c13b84-6c2a-4a6c-b23b-df1e14e077b5" (UID: "07c13b84-6c2a-4a6c-b23b-df1e14e077b5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:16:16 crc kubenswrapper[4824]: I1006 10:16:16.030350 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/07c13b84-6c2a-4a6c-b23b-df1e14e077b5-logs" (OuterVolumeSpecName: "logs") pod "07c13b84-6c2a-4a6c-b23b-df1e14e077b5" (UID: "07c13b84-6c2a-4a6c-b23b-df1e14e077b5"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:16:16 crc kubenswrapper[4824]: I1006 10:16:16.031001 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/07c13b84-6c2a-4a6c-b23b-df1e14e077b5-scripts" (OuterVolumeSpecName: "scripts") pod "07c13b84-6c2a-4a6c-b23b-df1e14e077b5" (UID: "07c13b84-6c2a-4a6c-b23b-df1e14e077b5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:16:16 crc kubenswrapper[4824]: I1006 10:16:16.031245 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/07c13b84-6c2a-4a6c-b23b-df1e14e077b5-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "07c13b84-6c2a-4a6c-b23b-df1e14e077b5" (UID: "07c13b84-6c2a-4a6c-b23b-df1e14e077b5"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:16:16 crc kubenswrapper[4824]: I1006 10:16:16.035614 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/07c13b84-6c2a-4a6c-b23b-df1e14e077b5-kube-api-access-vt2nw" (OuterVolumeSpecName: "kube-api-access-vt2nw") pod "07c13b84-6c2a-4a6c-b23b-df1e14e077b5" (UID: "07c13b84-6c2a-4a6c-b23b-df1e14e077b5"). InnerVolumeSpecName "kube-api-access-vt2nw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:16:16 crc kubenswrapper[4824]: I1006 10:16:16.131474 4824 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/07c13b84-6c2a-4a6c-b23b-df1e14e077b5-logs\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:16 crc kubenswrapper[4824]: I1006 10:16:16.131599 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/07c13b84-6c2a-4a6c-b23b-df1e14e077b5-scripts\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:16 crc kubenswrapper[4824]: I1006 10:16:16.131618 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/07c13b84-6c2a-4a6c-b23b-df1e14e077b5-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:16 crc kubenswrapper[4824]: I1006 10:16:16.131633 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt2nw\" (UniqueName: \"kubernetes.io/projected/07c13b84-6c2a-4a6c-b23b-df1e14e077b5-kube-api-access-vt2nw\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:16 crc kubenswrapper[4824]: I1006 10:16:16.131646 4824 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/07c13b84-6c2a-4a6c-b23b-df1e14e077b5-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:16 crc kubenswrapper[4824]: E1006 10:16:16.375478 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified" Oct 06 10:16:16 crc kubenswrapper[4824]: E1006 10:16:16.375708 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:barbican-db-sync,Image:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,Command:[/bin/bash],Args:[-c barbican-manage db upgrade],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/barbican/barbican.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-vcs8n,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42403,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42403,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-db-sync-ljqzg_openstack(dfc3bcd2-9baa-4a3a-8588-295e692d8e3e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 06 10:16:16 crc kubenswrapper[4824]: E1006 10:16:16.377030 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/barbican-db-sync-ljqzg" podUID="dfc3bcd2-9baa-4a3a-8588-295e692d8e3e" Oct 06 10:16:16 crc kubenswrapper[4824]: I1006 10:16:16.380070 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5bb95d74fc-r95ns" Oct 06 10:16:16 crc kubenswrapper[4824]: I1006 10:16:16.396863 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-xsgsg" Oct 06 10:16:16 crc kubenswrapper[4824]: I1006 10:16:16.437424 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/541d798b-323f-48c4-b0c1-a8e9d4333f6c-horizon-secret-key\") pod \"541d798b-323f-48c4-b0c1-a8e9d4333f6c\" (UID: \"541d798b-323f-48c4-b0c1-a8e9d4333f6c\") " Oct 06 10:16:16 crc kubenswrapper[4824]: I1006 10:16:16.437523 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/3d78bdee-7130-48a9-ad24-6c8c6525784c-config\") pod \"3d78bdee-7130-48a9-ad24-6c8c6525784c\" (UID: \"3d78bdee-7130-48a9-ad24-6c8c6525784c\") " Oct 06 10:16:16 crc kubenswrapper[4824]: I1006 10:16:16.440916 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8h584\" (UniqueName: \"kubernetes.io/projected/3d78bdee-7130-48a9-ad24-6c8c6525784c-kube-api-access-8h584\") pod \"3d78bdee-7130-48a9-ad24-6c8c6525784c\" (UID: \"3d78bdee-7130-48a9-ad24-6c8c6525784c\") " Oct 06 10:16:16 crc kubenswrapper[4824]: I1006 10:16:16.440999 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/541d798b-323f-48c4-b0c1-a8e9d4333f6c-scripts\") pod \"541d798b-323f-48c4-b0c1-a8e9d4333f6c\" (UID: \"541d798b-323f-48c4-b0c1-a8e9d4333f6c\") " Oct 06 10:16:16 crc kubenswrapper[4824]: I1006 10:16:16.441026 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d78bdee-7130-48a9-ad24-6c8c6525784c-combined-ca-bundle\") pod \"3d78bdee-7130-48a9-ad24-6c8c6525784c\" (UID: \"3d78bdee-7130-48a9-ad24-6c8c6525784c\") " Oct 06 10:16:16 crc kubenswrapper[4824]: I1006 10:16:16.441055 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/541d798b-323f-48c4-b0c1-a8e9d4333f6c-logs\") pod \"541d798b-323f-48c4-b0c1-a8e9d4333f6c\" (UID: \"541d798b-323f-48c4-b0c1-a8e9d4333f6c\") " Oct 06 10:16:16 crc kubenswrapper[4824]: I1006 10:16:16.441080 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-frht4\" (UniqueName: \"kubernetes.io/projected/541d798b-323f-48c4-b0c1-a8e9d4333f6c-kube-api-access-frht4\") pod \"541d798b-323f-48c4-b0c1-a8e9d4333f6c\" (UID: \"541d798b-323f-48c4-b0c1-a8e9d4333f6c\") " Oct 06 10:16:16 crc kubenswrapper[4824]: I1006 10:16:16.441124 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/541d798b-323f-48c4-b0c1-a8e9d4333f6c-config-data\") pod \"541d798b-323f-48c4-b0c1-a8e9d4333f6c\" (UID: \"541d798b-323f-48c4-b0c1-a8e9d4333f6c\") " Oct 06 10:16:16 crc kubenswrapper[4824]: I1006 10:16:16.441647 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/541d798b-323f-48c4-b0c1-a8e9d4333f6c-logs" (OuterVolumeSpecName: "logs") pod "541d798b-323f-48c4-b0c1-a8e9d4333f6c" (UID: "541d798b-323f-48c4-b0c1-a8e9d4333f6c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:16:16 crc kubenswrapper[4824]: I1006 10:16:16.441961 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/541d798b-323f-48c4-b0c1-a8e9d4333f6c-scripts" (OuterVolumeSpecName: "scripts") pod "541d798b-323f-48c4-b0c1-a8e9d4333f6c" (UID: "541d798b-323f-48c4-b0c1-a8e9d4333f6c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:16:16 crc kubenswrapper[4824]: I1006 10:16:16.442999 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/541d798b-323f-48c4-b0c1-a8e9d4333f6c-config-data" (OuterVolumeSpecName: "config-data") pod "541d798b-323f-48c4-b0c1-a8e9d4333f6c" (UID: "541d798b-323f-48c4-b0c1-a8e9d4333f6c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:16:16 crc kubenswrapper[4824]: I1006 10:16:16.446119 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/541d798b-323f-48c4-b0c1-a8e9d4333f6c-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "541d798b-323f-48c4-b0c1-a8e9d4333f6c" (UID: "541d798b-323f-48c4-b0c1-a8e9d4333f6c"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:16:16 crc kubenswrapper[4824]: I1006 10:16:16.446911 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3d78bdee-7130-48a9-ad24-6c8c6525784c-kube-api-access-8h584" (OuterVolumeSpecName: "kube-api-access-8h584") pod "3d78bdee-7130-48a9-ad24-6c8c6525784c" (UID: "3d78bdee-7130-48a9-ad24-6c8c6525784c"). InnerVolumeSpecName "kube-api-access-8h584". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:16:16 crc kubenswrapper[4824]: I1006 10:16:16.448480 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/541d798b-323f-48c4-b0c1-a8e9d4333f6c-kube-api-access-frht4" (OuterVolumeSpecName: "kube-api-access-frht4") pod "541d798b-323f-48c4-b0c1-a8e9d4333f6c" (UID: "541d798b-323f-48c4-b0c1-a8e9d4333f6c"). InnerVolumeSpecName "kube-api-access-frht4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:16:16 crc kubenswrapper[4824]: I1006 10:16:16.466142 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3d78bdee-7130-48a9-ad24-6c8c6525784c-config" (OuterVolumeSpecName: "config") pod "3d78bdee-7130-48a9-ad24-6c8c6525784c" (UID: "3d78bdee-7130-48a9-ad24-6c8c6525784c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:16:16 crc kubenswrapper[4824]: I1006 10:16:16.472168 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3d78bdee-7130-48a9-ad24-6c8c6525784c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3d78bdee-7130-48a9-ad24-6c8c6525784c" (UID: "3d78bdee-7130-48a9-ad24-6c8c6525784c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:16:16 crc kubenswrapper[4824]: I1006 10:16:16.530033 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5bb95d74fc-r95ns" Oct 06 10:16:16 crc kubenswrapper[4824]: I1006 10:16:16.530029 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5bb95d74fc-r95ns" event={"ID":"541d798b-323f-48c4-b0c1-a8e9d4333f6c","Type":"ContainerDied","Data":"1c38bb2748e1f4ee0eb9c40b22bd9a887d55033eae5448a150da9aa776021f3c"} Oct 06 10:16:16 crc kubenswrapper[4824]: I1006 10:16:16.532384 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77585f5f8c-wtjf5" Oct 06 10:16:16 crc kubenswrapper[4824]: I1006 10:16:16.532731 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-ccd5f54d9-dsnj5" Oct 06 10:16:16 crc kubenswrapper[4824]: I1006 10:16:16.532916 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 06 10:16:16 crc kubenswrapper[4824]: I1006 10:16:16.533541 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-xsgsg" event={"ID":"3d78bdee-7130-48a9-ad24-6c8c6525784c","Type":"ContainerDied","Data":"05056eba008dc12ea46d094704e044242e8f4be07ceec5435fc1eea48dfad173"} Oct 06 10:16:16 crc kubenswrapper[4824]: I1006 10:16:16.533623 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="05056eba008dc12ea46d094704e044242e8f4be07ceec5435fc1eea48dfad173" Oct 06 10:16:16 crc kubenswrapper[4824]: I1006 10:16:16.533922 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7749656757-6jdvp" Oct 06 10:16:16 crc kubenswrapper[4824]: I1006 10:16:16.533965 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-xsgsg" Oct 06 10:16:16 crc kubenswrapper[4824]: I1006 10:16:16.534104 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 06 10:16:16 crc kubenswrapper[4824]: I1006 10:16:16.544074 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/3d78bdee-7130-48a9-ad24-6c8c6525784c-config\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:16 crc kubenswrapper[4824]: I1006 10:16:16.544103 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8h584\" (UniqueName: \"kubernetes.io/projected/3d78bdee-7130-48a9-ad24-6c8c6525784c-kube-api-access-8h584\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:16 crc kubenswrapper[4824]: I1006 10:16:16.544115 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/541d798b-323f-48c4-b0c1-a8e9d4333f6c-scripts\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:16 crc kubenswrapper[4824]: I1006 10:16:16.544126 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d78bdee-7130-48a9-ad24-6c8c6525784c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:16 crc kubenswrapper[4824]: I1006 10:16:16.544136 4824 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/541d798b-323f-48c4-b0c1-a8e9d4333f6c-logs\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:16 crc kubenswrapper[4824]: I1006 10:16:16.544144 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-frht4\" (UniqueName: \"kubernetes.io/projected/541d798b-323f-48c4-b0c1-a8e9d4333f6c-kube-api-access-frht4\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:16 crc kubenswrapper[4824]: I1006 10:16:16.544153 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/541d798b-323f-48c4-b0c1-a8e9d4333f6c-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:16 crc kubenswrapper[4824]: I1006 10:16:16.544162 4824 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/541d798b-323f-48c4-b0c1-a8e9d4333f6c-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:16 crc kubenswrapper[4824]: E1006 10:16:16.544706 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified\\\"\"" pod="openstack/barbican-db-sync-ljqzg" podUID="dfc3bcd2-9baa-4a3a-8588-295e692d8e3e" Oct 06 10:16:16 crc kubenswrapper[4824]: I1006 10:16:16.646366 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-ccd5f54d9-dsnj5"] Oct 06 10:16:16 crc kubenswrapper[4824]: I1006 10:16:16.672402 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-ccd5f54d9-dsnj5"] Oct 06 10:16:16 crc kubenswrapper[4824]: I1006 10:16:16.686227 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-wtjf5"] Oct 06 10:16:16 crc kubenswrapper[4824]: I1006 10:16:16.694345 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-wtjf5"] Oct 06 10:16:16 crc kubenswrapper[4824]: I1006 10:16:16.743733 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-5bb95d74fc-r95ns"] Oct 06 10:16:16 crc kubenswrapper[4824]: I1006 10:16:16.748305 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-5bb95d74fc-r95ns"] Oct 06 10:16:16 crc kubenswrapper[4824]: I1006 10:16:16.759211 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 06 10:16:16 crc kubenswrapper[4824]: I1006 10:16:16.767364 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 06 10:16:16 crc kubenswrapper[4824]: I1006 10:16:16.787037 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 06 10:16:16 crc kubenswrapper[4824]: E1006 10:16:16.787444 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e26a57e-6683-48b2-a0ea-3f8ff1027fdc" containerName="glance-httpd" Oct 06 10:16:16 crc kubenswrapper[4824]: I1006 10:16:16.787463 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e26a57e-6683-48b2-a0ea-3f8ff1027fdc" containerName="glance-httpd" Oct 06 10:16:16 crc kubenswrapper[4824]: E1006 10:16:16.787478 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51ecd7bf-8b94-4c8c-af6e-a7e74e7f7c1c" containerName="init" Oct 06 10:16:16 crc kubenswrapper[4824]: I1006 10:16:16.787484 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="51ecd7bf-8b94-4c8c-af6e-a7e74e7f7c1c" containerName="init" Oct 06 10:16:16 crc kubenswrapper[4824]: E1006 10:16:16.787501 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97b2f7d6-3fd8-43ab-85b6-a3d37e218887" containerName="glance-log" Oct 06 10:16:16 crc kubenswrapper[4824]: I1006 10:16:16.787507 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="97b2f7d6-3fd8-43ab-85b6-a3d37e218887" containerName="glance-log" Oct 06 10:16:16 crc kubenswrapper[4824]: E1006 10:16:16.787517 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51ecd7bf-8b94-4c8c-af6e-a7e74e7f7c1c" containerName="dnsmasq-dns" Oct 06 10:16:16 crc kubenswrapper[4824]: I1006 10:16:16.787524 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="51ecd7bf-8b94-4c8c-af6e-a7e74e7f7c1c" containerName="dnsmasq-dns" Oct 06 10:16:16 crc kubenswrapper[4824]: E1006 10:16:16.787535 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d78bdee-7130-48a9-ad24-6c8c6525784c" containerName="neutron-db-sync" Oct 06 10:16:16 crc kubenswrapper[4824]: I1006 10:16:16.787542 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d78bdee-7130-48a9-ad24-6c8c6525784c" containerName="neutron-db-sync" Oct 06 10:16:16 crc kubenswrapper[4824]: E1006 10:16:16.787560 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97b2f7d6-3fd8-43ab-85b6-a3d37e218887" containerName="glance-httpd" Oct 06 10:16:16 crc kubenswrapper[4824]: I1006 10:16:16.787569 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="97b2f7d6-3fd8-43ab-85b6-a3d37e218887" containerName="glance-httpd" Oct 06 10:16:16 crc kubenswrapper[4824]: E1006 10:16:16.787593 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e26a57e-6683-48b2-a0ea-3f8ff1027fdc" containerName="glance-log" Oct 06 10:16:16 crc kubenswrapper[4824]: I1006 10:16:16.787599 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e26a57e-6683-48b2-a0ea-3f8ff1027fdc" containerName="glance-log" Oct 06 10:16:16 crc kubenswrapper[4824]: I1006 10:16:16.787741 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e26a57e-6683-48b2-a0ea-3f8ff1027fdc" containerName="glance-httpd" Oct 06 10:16:16 crc kubenswrapper[4824]: I1006 10:16:16.787752 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="97b2f7d6-3fd8-43ab-85b6-a3d37e218887" containerName="glance-log" Oct 06 10:16:16 crc kubenswrapper[4824]: I1006 10:16:16.787769 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="3d78bdee-7130-48a9-ad24-6c8c6525784c" containerName="neutron-db-sync" Oct 06 10:16:16 crc kubenswrapper[4824]: I1006 10:16:16.787778 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="51ecd7bf-8b94-4c8c-af6e-a7e74e7f7c1c" containerName="dnsmasq-dns" Oct 06 10:16:16 crc kubenswrapper[4824]: I1006 10:16:16.787790 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="97b2f7d6-3fd8-43ab-85b6-a3d37e218887" containerName="glance-httpd" Oct 06 10:16:16 crc kubenswrapper[4824]: I1006 10:16:16.787807 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e26a57e-6683-48b2-a0ea-3f8ff1027fdc" containerName="glance-log" Oct 06 10:16:16 crc kubenswrapper[4824]: I1006 10:16:16.788733 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 06 10:16:16 crc kubenswrapper[4824]: I1006 10:16:16.795941 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 06 10:16:16 crc kubenswrapper[4824]: I1006 10:16:16.796098 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Oct 06 10:16:16 crc kubenswrapper[4824]: I1006 10:16:16.796212 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-ctsvg" Oct 06 10:16:16 crc kubenswrapper[4824]: I1006 10:16:16.796321 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Oct 06 10:16:16 crc kubenswrapper[4824]: I1006 10:16:16.801045 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7749656757-6jdvp"] Oct 06 10:16:16 crc kubenswrapper[4824]: I1006 10:16:16.807762 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 06 10:16:16 crc kubenswrapper[4824]: I1006 10:16:16.819897 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-7749656757-6jdvp"] Oct 06 10:16:16 crc kubenswrapper[4824]: I1006 10:16:16.826304 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 06 10:16:16 crc kubenswrapper[4824]: I1006 10:16:16.832685 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 06 10:16:16 crc kubenswrapper[4824]: I1006 10:16:16.839083 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 06 10:16:16 crc kubenswrapper[4824]: I1006 10:16:16.840918 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 06 10:16:16 crc kubenswrapper[4824]: I1006 10:16:16.842728 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Oct 06 10:16:16 crc kubenswrapper[4824]: I1006 10:16:16.844317 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 06 10:16:16 crc kubenswrapper[4824]: I1006 10:16:16.845142 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 06 10:16:16 crc kubenswrapper[4824]: I1006 10:16:16.854067 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/508d58be-fcaa-4360-ab57-fbaedb728761-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"508d58be-fcaa-4360-ab57-fbaedb728761\") " pod="openstack/glance-default-external-api-0" Oct 06 10:16:16 crc kubenswrapper[4824]: I1006 10:16:16.854159 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/508d58be-fcaa-4360-ab57-fbaedb728761-logs\") pod \"glance-default-external-api-0\" (UID: \"508d58be-fcaa-4360-ab57-fbaedb728761\") " pod="openstack/glance-default-external-api-0" Oct 06 10:16:16 crc kubenswrapper[4824]: I1006 10:16:16.854188 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/508d58be-fcaa-4360-ab57-fbaedb728761-config-data\") pod \"glance-default-external-api-0\" (UID: \"508d58be-fcaa-4360-ab57-fbaedb728761\") " pod="openstack/glance-default-external-api-0" Oct 06 10:16:16 crc kubenswrapper[4824]: I1006 10:16:16.854306 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/508d58be-fcaa-4360-ab57-fbaedb728761-scripts\") pod \"glance-default-external-api-0\" (UID: \"508d58be-fcaa-4360-ab57-fbaedb728761\") " pod="openstack/glance-default-external-api-0" Oct 06 10:16:16 crc kubenswrapper[4824]: I1006 10:16:16.854329 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/508d58be-fcaa-4360-ab57-fbaedb728761-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"508d58be-fcaa-4360-ab57-fbaedb728761\") " pod="openstack/glance-default-external-api-0" Oct 06 10:16:16 crc kubenswrapper[4824]: I1006 10:16:16.854397 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"508d58be-fcaa-4360-ab57-fbaedb728761\") " pod="openstack/glance-default-external-api-0" Oct 06 10:16:16 crc kubenswrapper[4824]: I1006 10:16:16.854546 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/508d58be-fcaa-4360-ab57-fbaedb728761-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"508d58be-fcaa-4360-ab57-fbaedb728761\") " pod="openstack/glance-default-external-api-0" Oct 06 10:16:16 crc kubenswrapper[4824]: I1006 10:16:16.854585 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f5hbv\" (UniqueName: \"kubernetes.io/projected/508d58be-fcaa-4360-ab57-fbaedb728761-kube-api-access-f5hbv\") pod \"glance-default-external-api-0\" (UID: \"508d58be-fcaa-4360-ab57-fbaedb728761\") " pod="openstack/glance-default-external-api-0" Oct 06 10:16:16 crc kubenswrapper[4824]: I1006 10:16:16.961419 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a85f7f99-c70b-40a6-8ea2-839d91a53407-scripts\") pod \"glance-default-internal-api-0\" (UID: \"a85f7f99-c70b-40a6-8ea2-839d91a53407\") " pod="openstack/glance-default-internal-api-0" Oct 06 10:16:16 crc kubenswrapper[4824]: I1006 10:16:16.961466 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/508d58be-fcaa-4360-ab57-fbaedb728761-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"508d58be-fcaa-4360-ab57-fbaedb728761\") " pod="openstack/glance-default-external-api-0" Oct 06 10:16:16 crc kubenswrapper[4824]: I1006 10:16:16.961510 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/508d58be-fcaa-4360-ab57-fbaedb728761-logs\") pod \"glance-default-external-api-0\" (UID: \"508d58be-fcaa-4360-ab57-fbaedb728761\") " pod="openstack/glance-default-external-api-0" Oct 06 10:16:16 crc kubenswrapper[4824]: I1006 10:16:16.961591 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a85f7f99-c70b-40a6-8ea2-839d91a53407-config-data\") pod \"glance-default-internal-api-0\" (UID: \"a85f7f99-c70b-40a6-8ea2-839d91a53407\") " pod="openstack/glance-default-internal-api-0" Oct 06 10:16:16 crc kubenswrapper[4824]: I1006 10:16:16.961653 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/508d58be-fcaa-4360-ab57-fbaedb728761-config-data\") pod \"glance-default-external-api-0\" (UID: \"508d58be-fcaa-4360-ab57-fbaedb728761\") " pod="openstack/glance-default-external-api-0" Oct 06 10:16:16 crc kubenswrapper[4824]: I1006 10:16:16.962040 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/508d58be-fcaa-4360-ab57-fbaedb728761-logs\") pod \"glance-default-external-api-0\" (UID: \"508d58be-fcaa-4360-ab57-fbaedb728761\") " pod="openstack/glance-default-external-api-0" Oct 06 10:16:16 crc kubenswrapper[4824]: I1006 10:16:16.962090 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/508d58be-fcaa-4360-ab57-fbaedb728761-scripts\") pod \"glance-default-external-api-0\" (UID: \"508d58be-fcaa-4360-ab57-fbaedb728761\") " pod="openstack/glance-default-external-api-0" Oct 06 10:16:16 crc kubenswrapper[4824]: I1006 10:16:16.962206 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/508d58be-fcaa-4360-ab57-fbaedb728761-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"508d58be-fcaa-4360-ab57-fbaedb728761\") " pod="openstack/glance-default-external-api-0" Oct 06 10:16:16 crc kubenswrapper[4824]: I1006 10:16:16.962512 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/508d58be-fcaa-4360-ab57-fbaedb728761-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"508d58be-fcaa-4360-ab57-fbaedb728761\") " pod="openstack/glance-default-external-api-0" Oct 06 10:16:16 crc kubenswrapper[4824]: I1006 10:16:16.963930 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a85f7f99-c70b-40a6-8ea2-839d91a53407-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"a85f7f99-c70b-40a6-8ea2-839d91a53407\") " pod="openstack/glance-default-internal-api-0" Oct 06 10:16:16 crc kubenswrapper[4824]: I1006 10:16:16.964109 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"508d58be-fcaa-4360-ab57-fbaedb728761\") " pod="openstack/glance-default-external-api-0" Oct 06 10:16:16 crc kubenswrapper[4824]: I1006 10:16:16.964802 4824 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"508d58be-fcaa-4360-ab57-fbaedb728761\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/glance-default-external-api-0" Oct 06 10:16:16 crc kubenswrapper[4824]: I1006 10:16:16.966507 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"a85f7f99-c70b-40a6-8ea2-839d91a53407\") " pod="openstack/glance-default-internal-api-0" Oct 06 10:16:16 crc kubenswrapper[4824]: I1006 10:16:16.966637 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a85f7f99-c70b-40a6-8ea2-839d91a53407-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"a85f7f99-c70b-40a6-8ea2-839d91a53407\") " pod="openstack/glance-default-internal-api-0" Oct 06 10:16:16 crc kubenswrapper[4824]: I1006 10:16:16.966781 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a85f7f99-c70b-40a6-8ea2-839d91a53407-logs\") pod \"glance-default-internal-api-0\" (UID: \"a85f7f99-c70b-40a6-8ea2-839d91a53407\") " pod="openstack/glance-default-internal-api-0" Oct 06 10:16:16 crc kubenswrapper[4824]: I1006 10:16:16.966851 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a85f7f99-c70b-40a6-8ea2-839d91a53407-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"a85f7f99-c70b-40a6-8ea2-839d91a53407\") " pod="openstack/glance-default-internal-api-0" Oct 06 10:16:16 crc kubenswrapper[4824]: I1006 10:16:16.967061 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/508d58be-fcaa-4360-ab57-fbaedb728761-config-data\") pod \"glance-default-external-api-0\" (UID: \"508d58be-fcaa-4360-ab57-fbaedb728761\") " pod="openstack/glance-default-external-api-0" Oct 06 10:16:16 crc kubenswrapper[4824]: I1006 10:16:16.967138 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/508d58be-fcaa-4360-ab57-fbaedb728761-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"508d58be-fcaa-4360-ab57-fbaedb728761\") " pod="openstack/glance-default-external-api-0" Oct 06 10:16:16 crc kubenswrapper[4824]: I1006 10:16:16.973324 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/508d58be-fcaa-4360-ab57-fbaedb728761-scripts\") pod \"glance-default-external-api-0\" (UID: \"508d58be-fcaa-4360-ab57-fbaedb728761\") " pod="openstack/glance-default-external-api-0" Oct 06 10:16:16 crc kubenswrapper[4824]: I1006 10:16:16.976460 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n5zpc\" (UniqueName: \"kubernetes.io/projected/a85f7f99-c70b-40a6-8ea2-839d91a53407-kube-api-access-n5zpc\") pod \"glance-default-internal-api-0\" (UID: \"a85f7f99-c70b-40a6-8ea2-839d91a53407\") " pod="openstack/glance-default-internal-api-0" Oct 06 10:16:16 crc kubenswrapper[4824]: I1006 10:16:16.976643 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f5hbv\" (UniqueName: \"kubernetes.io/projected/508d58be-fcaa-4360-ab57-fbaedb728761-kube-api-access-f5hbv\") pod \"glance-default-external-api-0\" (UID: \"508d58be-fcaa-4360-ab57-fbaedb728761\") " pod="openstack/glance-default-external-api-0" Oct 06 10:16:16 crc kubenswrapper[4824]: I1006 10:16:16.976929 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/508d58be-fcaa-4360-ab57-fbaedb728761-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"508d58be-fcaa-4360-ab57-fbaedb728761\") " pod="openstack/glance-default-external-api-0" Oct 06 10:16:16 crc kubenswrapper[4824]: I1006 10:16:16.997920 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f5hbv\" (UniqueName: \"kubernetes.io/projected/508d58be-fcaa-4360-ab57-fbaedb728761-kube-api-access-f5hbv\") pod \"glance-default-external-api-0\" (UID: \"508d58be-fcaa-4360-ab57-fbaedb728761\") " pod="openstack/glance-default-external-api-0" Oct 06 10:16:17 crc kubenswrapper[4824]: I1006 10:16:17.013532 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"508d58be-fcaa-4360-ab57-fbaedb728761\") " pod="openstack/glance-default-external-api-0" Oct 06 10:16:17 crc kubenswrapper[4824]: I1006 10:16:17.014131 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/508d58be-fcaa-4360-ab57-fbaedb728761-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"508d58be-fcaa-4360-ab57-fbaedb728761\") " pod="openstack/glance-default-external-api-0" Oct 06 10:16:17 crc kubenswrapper[4824]: I1006 10:16:17.078380 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a85f7f99-c70b-40a6-8ea2-839d91a53407-scripts\") pod \"glance-default-internal-api-0\" (UID: \"a85f7f99-c70b-40a6-8ea2-839d91a53407\") " pod="openstack/glance-default-internal-api-0" Oct 06 10:16:17 crc kubenswrapper[4824]: I1006 10:16:17.078459 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a85f7f99-c70b-40a6-8ea2-839d91a53407-config-data\") pod \"glance-default-internal-api-0\" (UID: \"a85f7f99-c70b-40a6-8ea2-839d91a53407\") " pod="openstack/glance-default-internal-api-0" Oct 06 10:16:17 crc kubenswrapper[4824]: I1006 10:16:17.078938 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a85f7f99-c70b-40a6-8ea2-839d91a53407-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"a85f7f99-c70b-40a6-8ea2-839d91a53407\") " pod="openstack/glance-default-internal-api-0" Oct 06 10:16:17 crc kubenswrapper[4824]: I1006 10:16:17.078969 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"a85f7f99-c70b-40a6-8ea2-839d91a53407\") " pod="openstack/glance-default-internal-api-0" Oct 06 10:16:17 crc kubenswrapper[4824]: I1006 10:16:17.079024 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a85f7f99-c70b-40a6-8ea2-839d91a53407-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"a85f7f99-c70b-40a6-8ea2-839d91a53407\") " pod="openstack/glance-default-internal-api-0" Oct 06 10:16:17 crc kubenswrapper[4824]: I1006 10:16:17.079052 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a85f7f99-c70b-40a6-8ea2-839d91a53407-logs\") pod \"glance-default-internal-api-0\" (UID: \"a85f7f99-c70b-40a6-8ea2-839d91a53407\") " pod="openstack/glance-default-internal-api-0" Oct 06 10:16:17 crc kubenswrapper[4824]: I1006 10:16:17.079073 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a85f7f99-c70b-40a6-8ea2-839d91a53407-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"a85f7f99-c70b-40a6-8ea2-839d91a53407\") " pod="openstack/glance-default-internal-api-0" Oct 06 10:16:17 crc kubenswrapper[4824]: I1006 10:16:17.079120 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n5zpc\" (UniqueName: \"kubernetes.io/projected/a85f7f99-c70b-40a6-8ea2-839d91a53407-kube-api-access-n5zpc\") pod \"glance-default-internal-api-0\" (UID: \"a85f7f99-c70b-40a6-8ea2-839d91a53407\") " pod="openstack/glance-default-internal-api-0" Oct 06 10:16:17 crc kubenswrapper[4824]: I1006 10:16:17.079635 4824 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"a85f7f99-c70b-40a6-8ea2-839d91a53407\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/glance-default-internal-api-0" Oct 06 10:16:17 crc kubenswrapper[4824]: I1006 10:16:17.081387 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a85f7f99-c70b-40a6-8ea2-839d91a53407-logs\") pod \"glance-default-internal-api-0\" (UID: \"a85f7f99-c70b-40a6-8ea2-839d91a53407\") " pod="openstack/glance-default-internal-api-0" Oct 06 10:16:17 crc kubenswrapper[4824]: I1006 10:16:17.082558 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a85f7f99-c70b-40a6-8ea2-839d91a53407-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"a85f7f99-c70b-40a6-8ea2-839d91a53407\") " pod="openstack/glance-default-internal-api-0" Oct 06 10:16:17 crc kubenswrapper[4824]: I1006 10:16:17.085093 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a85f7f99-c70b-40a6-8ea2-839d91a53407-config-data\") pod \"glance-default-internal-api-0\" (UID: \"a85f7f99-c70b-40a6-8ea2-839d91a53407\") " pod="openstack/glance-default-internal-api-0" Oct 06 10:16:17 crc kubenswrapper[4824]: I1006 10:16:17.086960 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a85f7f99-c70b-40a6-8ea2-839d91a53407-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"a85f7f99-c70b-40a6-8ea2-839d91a53407\") " pod="openstack/glance-default-internal-api-0" Oct 06 10:16:17 crc kubenswrapper[4824]: I1006 10:16:17.094182 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a85f7f99-c70b-40a6-8ea2-839d91a53407-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"a85f7f99-c70b-40a6-8ea2-839d91a53407\") " pod="openstack/glance-default-internal-api-0" Oct 06 10:16:17 crc kubenswrapper[4824]: I1006 10:16:17.097274 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n5zpc\" (UniqueName: \"kubernetes.io/projected/a85f7f99-c70b-40a6-8ea2-839d91a53407-kube-api-access-n5zpc\") pod \"glance-default-internal-api-0\" (UID: \"a85f7f99-c70b-40a6-8ea2-839d91a53407\") " pod="openstack/glance-default-internal-api-0" Oct 06 10:16:17 crc kubenswrapper[4824]: I1006 10:16:17.104996 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 06 10:16:17 crc kubenswrapper[4824]: I1006 10:16:17.108365 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a85f7f99-c70b-40a6-8ea2-839d91a53407-scripts\") pod \"glance-default-internal-api-0\" (UID: \"a85f7f99-c70b-40a6-8ea2-839d91a53407\") " pod="openstack/glance-default-internal-api-0" Oct 06 10:16:17 crc kubenswrapper[4824]: I1006 10:16:17.119324 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"a85f7f99-c70b-40a6-8ea2-839d91a53407\") " pod="openstack/glance-default-internal-api-0" Oct 06 10:16:17 crc kubenswrapper[4824]: I1006 10:16:17.160099 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 06 10:16:17 crc kubenswrapper[4824]: I1006 10:16:17.288553 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="07c13b84-6c2a-4a6c-b23b-df1e14e077b5" path="/var/lib/kubelet/pods/07c13b84-6c2a-4a6c-b23b-df1e14e077b5/volumes" Oct 06 10:16:17 crc kubenswrapper[4824]: I1006 10:16:17.289080 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2e26a57e-6683-48b2-a0ea-3f8ff1027fdc" path="/var/lib/kubelet/pods/2e26a57e-6683-48b2-a0ea-3f8ff1027fdc/volumes" Oct 06 10:16:17 crc kubenswrapper[4824]: I1006 10:16:17.289930 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3d28a180-78e5-4977-bccb-9210f5fb0a7f" path="/var/lib/kubelet/pods/3d28a180-78e5-4977-bccb-9210f5fb0a7f/volumes" Oct 06 10:16:17 crc kubenswrapper[4824]: I1006 10:16:17.290339 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="51ecd7bf-8b94-4c8c-af6e-a7e74e7f7c1c" path="/var/lib/kubelet/pods/51ecd7bf-8b94-4c8c-af6e-a7e74e7f7c1c/volumes" Oct 06 10:16:17 crc kubenswrapper[4824]: I1006 10:16:17.291808 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="541d798b-323f-48c4-b0c1-a8e9d4333f6c" path="/var/lib/kubelet/pods/541d798b-323f-48c4-b0c1-a8e9d4333f6c/volumes" Oct 06 10:16:17 crc kubenswrapper[4824]: I1006 10:16:17.292434 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="97b2f7d6-3fd8-43ab-85b6-a3d37e218887" path="/var/lib/kubelet/pods/97b2f7d6-3fd8-43ab-85b6-a3d37e218887/volumes" Oct 06 10:16:17 crc kubenswrapper[4824]: I1006 10:16:17.661922 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-84b966f6c9-qwzmc"] Oct 06 10:16:17 crc kubenswrapper[4824]: I1006 10:16:17.663481 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84b966f6c9-qwzmc" Oct 06 10:16:17 crc kubenswrapper[4824]: I1006 10:16:17.685150 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-84b966f6c9-qwzmc"] Oct 06 10:16:17 crc kubenswrapper[4824]: I1006 10:16:17.691214 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/edde940b-8e99-477c-bdd3-9e767af9a4bc-ovsdbserver-nb\") pod \"dnsmasq-dns-84b966f6c9-qwzmc\" (UID: \"edde940b-8e99-477c-bdd3-9e767af9a4bc\") " pod="openstack/dnsmasq-dns-84b966f6c9-qwzmc" Oct 06 10:16:17 crc kubenswrapper[4824]: I1006 10:16:17.691265 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8qhq9\" (UniqueName: \"kubernetes.io/projected/edde940b-8e99-477c-bdd3-9e767af9a4bc-kube-api-access-8qhq9\") pod \"dnsmasq-dns-84b966f6c9-qwzmc\" (UID: \"edde940b-8e99-477c-bdd3-9e767af9a4bc\") " pod="openstack/dnsmasq-dns-84b966f6c9-qwzmc" Oct 06 10:16:17 crc kubenswrapper[4824]: I1006 10:16:17.691293 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/edde940b-8e99-477c-bdd3-9e767af9a4bc-dns-swift-storage-0\") pod \"dnsmasq-dns-84b966f6c9-qwzmc\" (UID: \"edde940b-8e99-477c-bdd3-9e767af9a4bc\") " pod="openstack/dnsmasq-dns-84b966f6c9-qwzmc" Oct 06 10:16:17 crc kubenswrapper[4824]: I1006 10:16:17.691392 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/edde940b-8e99-477c-bdd3-9e767af9a4bc-ovsdbserver-sb\") pod \"dnsmasq-dns-84b966f6c9-qwzmc\" (UID: \"edde940b-8e99-477c-bdd3-9e767af9a4bc\") " pod="openstack/dnsmasq-dns-84b966f6c9-qwzmc" Oct 06 10:16:17 crc kubenswrapper[4824]: I1006 10:16:17.691413 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/edde940b-8e99-477c-bdd3-9e767af9a4bc-dns-svc\") pod \"dnsmasq-dns-84b966f6c9-qwzmc\" (UID: \"edde940b-8e99-477c-bdd3-9e767af9a4bc\") " pod="openstack/dnsmasq-dns-84b966f6c9-qwzmc" Oct 06 10:16:17 crc kubenswrapper[4824]: I1006 10:16:17.691444 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/edde940b-8e99-477c-bdd3-9e767af9a4bc-config\") pod \"dnsmasq-dns-84b966f6c9-qwzmc\" (UID: \"edde940b-8e99-477c-bdd3-9e767af9a4bc\") " pod="openstack/dnsmasq-dns-84b966f6c9-qwzmc" Oct 06 10:16:17 crc kubenswrapper[4824]: I1006 10:16:17.796131 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/edde940b-8e99-477c-bdd3-9e767af9a4bc-dns-svc\") pod \"dnsmasq-dns-84b966f6c9-qwzmc\" (UID: \"edde940b-8e99-477c-bdd3-9e767af9a4bc\") " pod="openstack/dnsmasq-dns-84b966f6c9-qwzmc" Oct 06 10:16:17 crc kubenswrapper[4824]: I1006 10:16:17.796217 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/edde940b-8e99-477c-bdd3-9e767af9a4bc-config\") pod \"dnsmasq-dns-84b966f6c9-qwzmc\" (UID: \"edde940b-8e99-477c-bdd3-9e767af9a4bc\") " pod="openstack/dnsmasq-dns-84b966f6c9-qwzmc" Oct 06 10:16:17 crc kubenswrapper[4824]: I1006 10:16:17.796272 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/edde940b-8e99-477c-bdd3-9e767af9a4bc-ovsdbserver-nb\") pod \"dnsmasq-dns-84b966f6c9-qwzmc\" (UID: \"edde940b-8e99-477c-bdd3-9e767af9a4bc\") " pod="openstack/dnsmasq-dns-84b966f6c9-qwzmc" Oct 06 10:16:17 crc kubenswrapper[4824]: I1006 10:16:17.796302 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8qhq9\" (UniqueName: \"kubernetes.io/projected/edde940b-8e99-477c-bdd3-9e767af9a4bc-kube-api-access-8qhq9\") pod \"dnsmasq-dns-84b966f6c9-qwzmc\" (UID: \"edde940b-8e99-477c-bdd3-9e767af9a4bc\") " pod="openstack/dnsmasq-dns-84b966f6c9-qwzmc" Oct 06 10:16:17 crc kubenswrapper[4824]: I1006 10:16:17.796325 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/edde940b-8e99-477c-bdd3-9e767af9a4bc-dns-swift-storage-0\") pod \"dnsmasq-dns-84b966f6c9-qwzmc\" (UID: \"edde940b-8e99-477c-bdd3-9e767af9a4bc\") " pod="openstack/dnsmasq-dns-84b966f6c9-qwzmc" Oct 06 10:16:17 crc kubenswrapper[4824]: I1006 10:16:17.796477 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/edde940b-8e99-477c-bdd3-9e767af9a4bc-ovsdbserver-sb\") pod \"dnsmasq-dns-84b966f6c9-qwzmc\" (UID: \"edde940b-8e99-477c-bdd3-9e767af9a4bc\") " pod="openstack/dnsmasq-dns-84b966f6c9-qwzmc" Oct 06 10:16:17 crc kubenswrapper[4824]: I1006 10:16:17.797738 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/edde940b-8e99-477c-bdd3-9e767af9a4bc-ovsdbserver-sb\") pod \"dnsmasq-dns-84b966f6c9-qwzmc\" (UID: \"edde940b-8e99-477c-bdd3-9e767af9a4bc\") " pod="openstack/dnsmasq-dns-84b966f6c9-qwzmc" Oct 06 10:16:17 crc kubenswrapper[4824]: I1006 10:16:17.797739 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/edde940b-8e99-477c-bdd3-9e767af9a4bc-dns-svc\") pod \"dnsmasq-dns-84b966f6c9-qwzmc\" (UID: \"edde940b-8e99-477c-bdd3-9e767af9a4bc\") " pod="openstack/dnsmasq-dns-84b966f6c9-qwzmc" Oct 06 10:16:17 crc kubenswrapper[4824]: I1006 10:16:17.798490 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/edde940b-8e99-477c-bdd3-9e767af9a4bc-ovsdbserver-nb\") pod \"dnsmasq-dns-84b966f6c9-qwzmc\" (UID: \"edde940b-8e99-477c-bdd3-9e767af9a4bc\") " pod="openstack/dnsmasq-dns-84b966f6c9-qwzmc" Oct 06 10:16:17 crc kubenswrapper[4824]: I1006 10:16:17.799559 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/edde940b-8e99-477c-bdd3-9e767af9a4bc-config\") pod \"dnsmasq-dns-84b966f6c9-qwzmc\" (UID: \"edde940b-8e99-477c-bdd3-9e767af9a4bc\") " pod="openstack/dnsmasq-dns-84b966f6c9-qwzmc" Oct 06 10:16:17 crc kubenswrapper[4824]: I1006 10:16:17.800598 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/edde940b-8e99-477c-bdd3-9e767af9a4bc-dns-swift-storage-0\") pod \"dnsmasq-dns-84b966f6c9-qwzmc\" (UID: \"edde940b-8e99-477c-bdd3-9e767af9a4bc\") " pod="openstack/dnsmasq-dns-84b966f6c9-qwzmc" Oct 06 10:16:17 crc kubenswrapper[4824]: I1006 10:16:17.821281 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-65448d5ddd-mfzdl"] Oct 06 10:16:17 crc kubenswrapper[4824]: I1006 10:16:17.825857 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-65448d5ddd-mfzdl" Oct 06 10:16:17 crc kubenswrapper[4824]: I1006 10:16:17.833904 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8qhq9\" (UniqueName: \"kubernetes.io/projected/edde940b-8e99-477c-bdd3-9e767af9a4bc-kube-api-access-8qhq9\") pod \"dnsmasq-dns-84b966f6c9-qwzmc\" (UID: \"edde940b-8e99-477c-bdd3-9e767af9a4bc\") " pod="openstack/dnsmasq-dns-84b966f6c9-qwzmc" Oct 06 10:16:17 crc kubenswrapper[4824]: I1006 10:16:17.842040 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Oct 06 10:16:17 crc kubenswrapper[4824]: I1006 10:16:17.842265 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-wx2zn" Oct 06 10:16:17 crc kubenswrapper[4824]: I1006 10:16:17.842446 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Oct 06 10:16:17 crc kubenswrapper[4824]: I1006 10:16:17.842562 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Oct 06 10:16:17 crc kubenswrapper[4824]: I1006 10:16:17.862535 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-65448d5ddd-mfzdl"] Oct 06 10:16:17 crc kubenswrapper[4824]: I1006 10:16:17.902438 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/299892ea-e7b0-455b-9277-865866f47b63-config\") pod \"neutron-65448d5ddd-mfzdl\" (UID: \"299892ea-e7b0-455b-9277-865866f47b63\") " pod="openstack/neutron-65448d5ddd-mfzdl" Oct 06 10:16:17 crc kubenswrapper[4824]: I1006 10:16:17.902518 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hwlc2\" (UniqueName: \"kubernetes.io/projected/299892ea-e7b0-455b-9277-865866f47b63-kube-api-access-hwlc2\") pod \"neutron-65448d5ddd-mfzdl\" (UID: \"299892ea-e7b0-455b-9277-865866f47b63\") " pod="openstack/neutron-65448d5ddd-mfzdl" Oct 06 10:16:17 crc kubenswrapper[4824]: I1006 10:16:17.902600 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/299892ea-e7b0-455b-9277-865866f47b63-httpd-config\") pod \"neutron-65448d5ddd-mfzdl\" (UID: \"299892ea-e7b0-455b-9277-865866f47b63\") " pod="openstack/neutron-65448d5ddd-mfzdl" Oct 06 10:16:17 crc kubenswrapper[4824]: I1006 10:16:17.902818 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/299892ea-e7b0-455b-9277-865866f47b63-combined-ca-bundle\") pod \"neutron-65448d5ddd-mfzdl\" (UID: \"299892ea-e7b0-455b-9277-865866f47b63\") " pod="openstack/neutron-65448d5ddd-mfzdl" Oct 06 10:16:17 crc kubenswrapper[4824]: I1006 10:16:17.902843 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/299892ea-e7b0-455b-9277-865866f47b63-ovndb-tls-certs\") pod \"neutron-65448d5ddd-mfzdl\" (UID: \"299892ea-e7b0-455b-9277-865866f47b63\") " pod="openstack/neutron-65448d5ddd-mfzdl" Oct 06 10:16:17 crc kubenswrapper[4824]: I1006 10:16:17.992001 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84b966f6c9-qwzmc" Oct 06 10:16:18 crc kubenswrapper[4824]: I1006 10:16:18.004713 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/299892ea-e7b0-455b-9277-865866f47b63-combined-ca-bundle\") pod \"neutron-65448d5ddd-mfzdl\" (UID: \"299892ea-e7b0-455b-9277-865866f47b63\") " pod="openstack/neutron-65448d5ddd-mfzdl" Oct 06 10:16:18 crc kubenswrapper[4824]: I1006 10:16:18.004782 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/299892ea-e7b0-455b-9277-865866f47b63-ovndb-tls-certs\") pod \"neutron-65448d5ddd-mfzdl\" (UID: \"299892ea-e7b0-455b-9277-865866f47b63\") " pod="openstack/neutron-65448d5ddd-mfzdl" Oct 06 10:16:18 crc kubenswrapper[4824]: I1006 10:16:18.004826 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/299892ea-e7b0-455b-9277-865866f47b63-config\") pod \"neutron-65448d5ddd-mfzdl\" (UID: \"299892ea-e7b0-455b-9277-865866f47b63\") " pod="openstack/neutron-65448d5ddd-mfzdl" Oct 06 10:16:18 crc kubenswrapper[4824]: I1006 10:16:18.004866 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hwlc2\" (UniqueName: \"kubernetes.io/projected/299892ea-e7b0-455b-9277-865866f47b63-kube-api-access-hwlc2\") pod \"neutron-65448d5ddd-mfzdl\" (UID: \"299892ea-e7b0-455b-9277-865866f47b63\") " pod="openstack/neutron-65448d5ddd-mfzdl" Oct 06 10:16:18 crc kubenswrapper[4824]: I1006 10:16:18.004952 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/299892ea-e7b0-455b-9277-865866f47b63-httpd-config\") pod \"neutron-65448d5ddd-mfzdl\" (UID: \"299892ea-e7b0-455b-9277-865866f47b63\") " pod="openstack/neutron-65448d5ddd-mfzdl" Oct 06 10:16:18 crc kubenswrapper[4824]: I1006 10:16:18.012085 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/299892ea-e7b0-455b-9277-865866f47b63-config\") pod \"neutron-65448d5ddd-mfzdl\" (UID: \"299892ea-e7b0-455b-9277-865866f47b63\") " pod="openstack/neutron-65448d5ddd-mfzdl" Oct 06 10:16:18 crc kubenswrapper[4824]: I1006 10:16:18.015262 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/299892ea-e7b0-455b-9277-865866f47b63-httpd-config\") pod \"neutron-65448d5ddd-mfzdl\" (UID: \"299892ea-e7b0-455b-9277-865866f47b63\") " pod="openstack/neutron-65448d5ddd-mfzdl" Oct 06 10:16:18 crc kubenswrapper[4824]: I1006 10:16:18.015350 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/299892ea-e7b0-455b-9277-865866f47b63-combined-ca-bundle\") pod \"neutron-65448d5ddd-mfzdl\" (UID: \"299892ea-e7b0-455b-9277-865866f47b63\") " pod="openstack/neutron-65448d5ddd-mfzdl" Oct 06 10:16:18 crc kubenswrapper[4824]: I1006 10:16:18.021229 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/299892ea-e7b0-455b-9277-865866f47b63-ovndb-tls-certs\") pod \"neutron-65448d5ddd-mfzdl\" (UID: \"299892ea-e7b0-455b-9277-865866f47b63\") " pod="openstack/neutron-65448d5ddd-mfzdl" Oct 06 10:16:18 crc kubenswrapper[4824]: I1006 10:16:18.029083 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hwlc2\" (UniqueName: \"kubernetes.io/projected/299892ea-e7b0-455b-9277-865866f47b63-kube-api-access-hwlc2\") pod \"neutron-65448d5ddd-mfzdl\" (UID: \"299892ea-e7b0-455b-9277-865866f47b63\") " pod="openstack/neutron-65448d5ddd-mfzdl" Oct 06 10:16:18 crc kubenswrapper[4824]: I1006 10:16:18.076486 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-77585f5f8c-wtjf5" podUID="51ecd7bf-8b94-4c8c-af6e-a7e74e7f7c1c" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.133:5353: i/o timeout" Oct 06 10:16:18 crc kubenswrapper[4824]: I1006 10:16:18.206724 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-65448d5ddd-mfzdl" Oct 06 10:16:18 crc kubenswrapper[4824]: E1006 10:16:18.747297 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified" Oct 06 10:16:18 crc kubenswrapper[4824]: E1006 10:16:18.747441 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-fqk7s,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-ldn9v_openstack(2382b80c-75fc-435f-be5f-0f1eb6a5cd17): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 06 10:16:18 crc kubenswrapper[4824]: E1006 10:16:18.752188 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-ldn9v" podUID="2382b80c-75fc-435f-be5f-0f1eb6a5cd17" Oct 06 10:16:19 crc kubenswrapper[4824]: I1006 10:16:19.204137 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-78c847f9f4-tv8zc"] Oct 06 10:16:19 crc kubenswrapper[4824]: I1006 10:16:19.286316 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-68fdc5dd48-gfq4d"] Oct 06 10:16:19 crc kubenswrapper[4824]: I1006 10:16:19.590291 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-78c847f9f4-tv8zc" event={"ID":"68ef92a4-c4bc-4acc-be58-e14010395b8e","Type":"ContainerStarted","Data":"7a182aa0ff510239fa3e786b04999f60551ad03be619039603b91a15a219780e"} Oct 06 10:16:19 crc kubenswrapper[4824]: I1006 10:16:19.599135 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-68fdc5dd48-gfq4d" event={"ID":"194bef17-98ac-413f-a91e-0abc9835f2f6","Type":"ContainerStarted","Data":"1a60a38a8a0c0769b946e6137e3835734d33a346829cbc994562a0eb6233f7ec"} Oct 06 10:16:19 crc kubenswrapper[4824]: E1006 10:16:19.602385 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified\\\"\"" pod="openstack/cinder-db-sync-ldn9v" podUID="2382b80c-75fc-435f-be5f-0f1eb6a5cd17" Oct 06 10:16:19 crc kubenswrapper[4824]: I1006 10:16:19.774119 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-c7c794649-hcxzr"] Oct 06 10:16:19 crc kubenswrapper[4824]: I1006 10:16:19.775794 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-c7c794649-hcxzr" Oct 06 10:16:19 crc kubenswrapper[4824]: I1006 10:16:19.781554 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Oct 06 10:16:19 crc kubenswrapper[4824]: I1006 10:16:19.782008 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Oct 06 10:16:19 crc kubenswrapper[4824]: I1006 10:16:19.798633 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-c7c794649-hcxzr"] Oct 06 10:16:19 crc kubenswrapper[4824]: I1006 10:16:19.872873 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zzwrw\" (UniqueName: \"kubernetes.io/projected/2919256c-ce36-4092-898d-024f10692bef-kube-api-access-zzwrw\") pod \"neutron-c7c794649-hcxzr\" (UID: \"2919256c-ce36-4092-898d-024f10692bef\") " pod="openstack/neutron-c7c794649-hcxzr" Oct 06 10:16:19 crc kubenswrapper[4824]: I1006 10:16:19.872967 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2919256c-ce36-4092-898d-024f10692bef-combined-ca-bundle\") pod \"neutron-c7c794649-hcxzr\" (UID: \"2919256c-ce36-4092-898d-024f10692bef\") " pod="openstack/neutron-c7c794649-hcxzr" Oct 06 10:16:19 crc kubenswrapper[4824]: I1006 10:16:19.873009 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/2919256c-ce36-4092-898d-024f10692bef-httpd-config\") pod \"neutron-c7c794649-hcxzr\" (UID: \"2919256c-ce36-4092-898d-024f10692bef\") " pod="openstack/neutron-c7c794649-hcxzr" Oct 06 10:16:19 crc kubenswrapper[4824]: I1006 10:16:19.873029 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/2919256c-ce36-4092-898d-024f10692bef-config\") pod \"neutron-c7c794649-hcxzr\" (UID: \"2919256c-ce36-4092-898d-024f10692bef\") " pod="openstack/neutron-c7c794649-hcxzr" Oct 06 10:16:19 crc kubenswrapper[4824]: I1006 10:16:19.873054 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2919256c-ce36-4092-898d-024f10692bef-internal-tls-certs\") pod \"neutron-c7c794649-hcxzr\" (UID: \"2919256c-ce36-4092-898d-024f10692bef\") " pod="openstack/neutron-c7c794649-hcxzr" Oct 06 10:16:19 crc kubenswrapper[4824]: I1006 10:16:19.873181 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/2919256c-ce36-4092-898d-024f10692bef-ovndb-tls-certs\") pod \"neutron-c7c794649-hcxzr\" (UID: \"2919256c-ce36-4092-898d-024f10692bef\") " pod="openstack/neutron-c7c794649-hcxzr" Oct 06 10:16:19 crc kubenswrapper[4824]: I1006 10:16:19.873269 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2919256c-ce36-4092-898d-024f10692bef-public-tls-certs\") pod \"neutron-c7c794649-hcxzr\" (UID: \"2919256c-ce36-4092-898d-024f10692bef\") " pod="openstack/neutron-c7c794649-hcxzr" Oct 06 10:16:19 crc kubenswrapper[4824]: I1006 10:16:19.956652 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-wlmn6"] Oct 06 10:16:19 crc kubenswrapper[4824]: I1006 10:16:19.974776 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2919256c-ce36-4092-898d-024f10692bef-public-tls-certs\") pod \"neutron-c7c794649-hcxzr\" (UID: \"2919256c-ce36-4092-898d-024f10692bef\") " pod="openstack/neutron-c7c794649-hcxzr" Oct 06 10:16:19 crc kubenswrapper[4824]: I1006 10:16:19.974868 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zzwrw\" (UniqueName: \"kubernetes.io/projected/2919256c-ce36-4092-898d-024f10692bef-kube-api-access-zzwrw\") pod \"neutron-c7c794649-hcxzr\" (UID: \"2919256c-ce36-4092-898d-024f10692bef\") " pod="openstack/neutron-c7c794649-hcxzr" Oct 06 10:16:19 crc kubenswrapper[4824]: I1006 10:16:19.974928 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2919256c-ce36-4092-898d-024f10692bef-combined-ca-bundle\") pod \"neutron-c7c794649-hcxzr\" (UID: \"2919256c-ce36-4092-898d-024f10692bef\") " pod="openstack/neutron-c7c794649-hcxzr" Oct 06 10:16:19 crc kubenswrapper[4824]: I1006 10:16:19.974946 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/2919256c-ce36-4092-898d-024f10692bef-httpd-config\") pod \"neutron-c7c794649-hcxzr\" (UID: \"2919256c-ce36-4092-898d-024f10692bef\") " pod="openstack/neutron-c7c794649-hcxzr" Oct 06 10:16:19 crc kubenswrapper[4824]: I1006 10:16:19.974968 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/2919256c-ce36-4092-898d-024f10692bef-config\") pod \"neutron-c7c794649-hcxzr\" (UID: \"2919256c-ce36-4092-898d-024f10692bef\") " pod="openstack/neutron-c7c794649-hcxzr" Oct 06 10:16:19 crc kubenswrapper[4824]: I1006 10:16:19.975005 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2919256c-ce36-4092-898d-024f10692bef-internal-tls-certs\") pod \"neutron-c7c794649-hcxzr\" (UID: \"2919256c-ce36-4092-898d-024f10692bef\") " pod="openstack/neutron-c7c794649-hcxzr" Oct 06 10:16:19 crc kubenswrapper[4824]: I1006 10:16:19.975030 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/2919256c-ce36-4092-898d-024f10692bef-ovndb-tls-certs\") pod \"neutron-c7c794649-hcxzr\" (UID: \"2919256c-ce36-4092-898d-024f10692bef\") " pod="openstack/neutron-c7c794649-hcxzr" Oct 06 10:16:19 crc kubenswrapper[4824]: I1006 10:16:19.984921 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/2919256c-ce36-4092-898d-024f10692bef-config\") pod \"neutron-c7c794649-hcxzr\" (UID: \"2919256c-ce36-4092-898d-024f10692bef\") " pod="openstack/neutron-c7c794649-hcxzr" Oct 06 10:16:19 crc kubenswrapper[4824]: I1006 10:16:19.988605 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/2919256c-ce36-4092-898d-024f10692bef-ovndb-tls-certs\") pod \"neutron-c7c794649-hcxzr\" (UID: \"2919256c-ce36-4092-898d-024f10692bef\") " pod="openstack/neutron-c7c794649-hcxzr" Oct 06 10:16:19 crc kubenswrapper[4824]: I1006 10:16:19.989085 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/2919256c-ce36-4092-898d-024f10692bef-httpd-config\") pod \"neutron-c7c794649-hcxzr\" (UID: \"2919256c-ce36-4092-898d-024f10692bef\") " pod="openstack/neutron-c7c794649-hcxzr" Oct 06 10:16:19 crc kubenswrapper[4824]: I1006 10:16:19.989264 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2919256c-ce36-4092-898d-024f10692bef-combined-ca-bundle\") pod \"neutron-c7c794649-hcxzr\" (UID: \"2919256c-ce36-4092-898d-024f10692bef\") " pod="openstack/neutron-c7c794649-hcxzr" Oct 06 10:16:19 crc kubenswrapper[4824]: I1006 10:16:19.991493 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2919256c-ce36-4092-898d-024f10692bef-internal-tls-certs\") pod \"neutron-c7c794649-hcxzr\" (UID: \"2919256c-ce36-4092-898d-024f10692bef\") " pod="openstack/neutron-c7c794649-hcxzr" Oct 06 10:16:20 crc kubenswrapper[4824]: I1006 10:16:19.998294 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2919256c-ce36-4092-898d-024f10692bef-public-tls-certs\") pod \"neutron-c7c794649-hcxzr\" (UID: \"2919256c-ce36-4092-898d-024f10692bef\") " pod="openstack/neutron-c7c794649-hcxzr" Oct 06 10:16:20 crc kubenswrapper[4824]: I1006 10:16:20.016853 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zzwrw\" (UniqueName: \"kubernetes.io/projected/2919256c-ce36-4092-898d-024f10692bef-kube-api-access-zzwrw\") pod \"neutron-c7c794649-hcxzr\" (UID: \"2919256c-ce36-4092-898d-024f10692bef\") " pod="openstack/neutron-c7c794649-hcxzr" Oct 06 10:16:20 crc kubenswrapper[4824]: I1006 10:16:20.078190 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-65448d5ddd-mfzdl"] Oct 06 10:16:20 crc kubenswrapper[4824]: I1006 10:16:20.099269 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-c7c794649-hcxzr" Oct 06 10:16:20 crc kubenswrapper[4824]: I1006 10:16:20.175687 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-84b966f6c9-qwzmc"] Oct 06 10:16:20 crc kubenswrapper[4824]: I1006 10:16:20.271488 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 06 10:16:20 crc kubenswrapper[4824]: I1006 10:16:20.626759 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-65448d5ddd-mfzdl" event={"ID":"299892ea-e7b0-455b-9277-865866f47b63","Type":"ContainerStarted","Data":"2cb51964aca5d0c9ee841b955b8b7366929eb144c4e2e4aca4376e930b9de5cf"} Oct 06 10:16:20 crc kubenswrapper[4824]: I1006 10:16:20.627134 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-65448d5ddd-mfzdl" event={"ID":"299892ea-e7b0-455b-9277-865866f47b63","Type":"ContainerStarted","Data":"28a0a8ae4d05860d5fe7df5368cd78b30c60ffcde19db5dff14cce8952b9d3aa"} Oct 06 10:16:20 crc kubenswrapper[4824]: I1006 10:16:20.634252 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"508d58be-fcaa-4360-ab57-fbaedb728761","Type":"ContainerStarted","Data":"66a2d1202ba3ab9a6dcb54d2034450460df4d403cba38005cc0584e842b32557"} Oct 06 10:16:20 crc kubenswrapper[4824]: I1006 10:16:20.653405 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-fzvv8" event={"ID":"554a3f37-c7c8-4d09-9da7-df5319dccecd","Type":"ContainerStarted","Data":"a787dec7530bf4c4cbfc1e0888aebec0261071050df377f25ba2aebb9d060923"} Oct 06 10:16:20 crc kubenswrapper[4824]: I1006 10:16:20.672569 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-68fdc5dd48-gfq4d" event={"ID":"194bef17-98ac-413f-a91e-0abc9835f2f6","Type":"ContainerStarted","Data":"a27c1a94b5468ee311fd30f4aaa5bc8f4a2b7aaa3dc5c92cd9726c9e15e7ee80"} Oct 06 10:16:20 crc kubenswrapper[4824]: I1006 10:16:20.675446 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-wlmn6" event={"ID":"73fb825b-ea3b-43c5-b143-e70310f8c1fd","Type":"ContainerStarted","Data":"90ca09e2651e49075cef289ad0aa7e54e3882d823344ef1a9a6a28355a58ff56"} Oct 06 10:16:20 crc kubenswrapper[4824]: I1006 10:16:20.675477 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-wlmn6" event={"ID":"73fb825b-ea3b-43c5-b143-e70310f8c1fd","Type":"ContainerStarted","Data":"46f6c47a06c544dcb4e13ef4f1f4782faa48cc9b568008f5dc60118b956d5e27"} Oct 06 10:16:20 crc kubenswrapper[4824]: I1006 10:16:20.691313 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84b966f6c9-qwzmc" event={"ID":"edde940b-8e99-477c-bdd3-9e767af9a4bc","Type":"ContainerStarted","Data":"2de991af05c79688d84d822398a984767c382aabfb3e0b7684f1fe348a6926bc"} Oct 06 10:16:20 crc kubenswrapper[4824]: I1006 10:16:20.691742 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-fzvv8" podStartSLOduration=7.493531589 podStartE2EDuration="34.691716355s" podCreationTimestamp="2025-10-06 10:15:46 +0000 UTC" firstStartedPulling="2025-10-06 10:15:48.234316243 +0000 UTC m=+1057.598739104" lastFinishedPulling="2025-10-06 10:16:15.432501009 +0000 UTC m=+1084.796923870" observedRunningTime="2025-10-06 10:16:20.677479583 +0000 UTC m=+1090.041902454" watchObservedRunningTime="2025-10-06 10:16:20.691716355 +0000 UTC m=+1090.056139386" Oct 06 10:16:20 crc kubenswrapper[4824]: I1006 10:16:20.702001 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"13781991-323c-49ae-9c5d-625e2ffd019b","Type":"ContainerStarted","Data":"c75d8022f4b4f06e5bd8a64705b9bfbcd9ee5e6a7edb3e34e026c46d0f72d761"} Oct 06 10:16:20 crc kubenswrapper[4824]: I1006 10:16:20.727247 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-wlmn6" podStartSLOduration=20.727221532 podStartE2EDuration="20.727221532s" podCreationTimestamp="2025-10-06 10:16:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:16:20.701119537 +0000 UTC m=+1090.065542398" watchObservedRunningTime="2025-10-06 10:16:20.727221532 +0000 UTC m=+1090.091644393" Oct 06 10:16:20 crc kubenswrapper[4824]: I1006 10:16:20.733650 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-78c847f9f4-tv8zc" event={"ID":"68ef92a4-c4bc-4acc-be58-e14010395b8e","Type":"ContainerStarted","Data":"e347cab97c01ec84b5735cd302c782ea65d188c9d2c13baf5983c3a2843b3758"} Oct 06 10:16:20 crc kubenswrapper[4824]: I1006 10:16:20.803681 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-c7c794649-hcxzr"] Oct 06 10:16:21 crc kubenswrapper[4824]: I1006 10:16:21.675417 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 06 10:16:21 crc kubenswrapper[4824]: I1006 10:16:21.824019 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-78c847f9f4-tv8zc" event={"ID":"68ef92a4-c4bc-4acc-be58-e14010395b8e","Type":"ContainerStarted","Data":"62bfee4f7781b10bfd4f3b2388238d938ea44501a5d20380123b779025d75b32"} Oct 06 10:16:21 crc kubenswrapper[4824]: I1006 10:16:21.836746 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-65448d5ddd-mfzdl" event={"ID":"299892ea-e7b0-455b-9277-865866f47b63","Type":"ContainerStarted","Data":"d3552bb35c20a93f32e2830bccd410137d4b0c1ca566b67532d960b5ca86d33a"} Oct 06 10:16:21 crc kubenswrapper[4824]: I1006 10:16:21.836889 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-65448d5ddd-mfzdl" Oct 06 10:16:21 crc kubenswrapper[4824]: I1006 10:16:21.841239 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"a85f7f99-c70b-40a6-8ea2-839d91a53407","Type":"ContainerStarted","Data":"681b14ea91aa7b1bdb3c7c3c06994964c1bef09cf51181d7dca3cf76ae83bcd7"} Oct 06 10:16:21 crc kubenswrapper[4824]: I1006 10:16:21.847094 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-78c847f9f4-tv8zc" podStartSLOduration=26.265098143 podStartE2EDuration="26.847077435s" podCreationTimestamp="2025-10-06 10:15:55 +0000 UTC" firstStartedPulling="2025-10-06 10:16:19.328214456 +0000 UTC m=+1088.692637317" lastFinishedPulling="2025-10-06 10:16:19.910193748 +0000 UTC m=+1089.274616609" observedRunningTime="2025-10-06 10:16:21.844500001 +0000 UTC m=+1091.208922862" watchObservedRunningTime="2025-10-06 10:16:21.847077435 +0000 UTC m=+1091.211500286" Oct 06 10:16:21 crc kubenswrapper[4824]: I1006 10:16:21.855297 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"508d58be-fcaa-4360-ab57-fbaedb728761","Type":"ContainerStarted","Data":"2bf864c0a9c8f342accc699a7313528172c1c81ba4c82a5d1cab8dbbfdbe803b"} Oct 06 10:16:21 crc kubenswrapper[4824]: I1006 10:16:21.881156 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-65448d5ddd-mfzdl" podStartSLOduration=4.881137677 podStartE2EDuration="4.881137677s" podCreationTimestamp="2025-10-06 10:16:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:16:21.871477929 +0000 UTC m=+1091.235900800" watchObservedRunningTime="2025-10-06 10:16:21.881137677 +0000 UTC m=+1091.245560538" Oct 06 10:16:21 crc kubenswrapper[4824]: I1006 10:16:21.886733 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-68fdc5dd48-gfq4d" event={"ID":"194bef17-98ac-413f-a91e-0abc9835f2f6","Type":"ContainerStarted","Data":"5652a0212da0dc132c12a8453925efb6473500c241ef4d8871792bc0c878514a"} Oct 06 10:16:21 crc kubenswrapper[4824]: I1006 10:16:21.900192 4824 generic.go:334] "Generic (PLEG): container finished" podID="edde940b-8e99-477c-bdd3-9e767af9a4bc" containerID="8d3c656a9a067207dfb41d0c1771e02a3f4d418075ce19eecfff8a23b5e7f9e3" exitCode=0 Oct 06 10:16:21 crc kubenswrapper[4824]: I1006 10:16:21.900277 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84b966f6c9-qwzmc" event={"ID":"edde940b-8e99-477c-bdd3-9e767af9a4bc","Type":"ContainerDied","Data":"8d3c656a9a067207dfb41d0c1771e02a3f4d418075ce19eecfff8a23b5e7f9e3"} Oct 06 10:16:21 crc kubenswrapper[4824]: I1006 10:16:21.900307 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84b966f6c9-qwzmc" event={"ID":"edde940b-8e99-477c-bdd3-9e767af9a4bc","Type":"ContainerStarted","Data":"91658e4810109b74e59028cee75dd39cbad5b51837f934261d99c5a6fdd8496c"} Oct 06 10:16:21 crc kubenswrapper[4824]: I1006 10:16:21.902346 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-84b966f6c9-qwzmc" Oct 06 10:16:21 crc kubenswrapper[4824]: I1006 10:16:21.918114 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-68fdc5dd48-gfq4d" podStartSLOduration=26.464536066 podStartE2EDuration="26.918094792s" podCreationTimestamp="2025-10-06 10:15:55 +0000 UTC" firstStartedPulling="2025-10-06 10:16:19.392568798 +0000 UTC m=+1088.756991659" lastFinishedPulling="2025-10-06 10:16:19.846127514 +0000 UTC m=+1089.210550385" observedRunningTime="2025-10-06 10:16:21.911068298 +0000 UTC m=+1091.275491159" watchObservedRunningTime="2025-10-06 10:16:21.918094792 +0000 UTC m=+1091.282517653" Oct 06 10:16:21 crc kubenswrapper[4824]: I1006 10:16:21.934442 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-c7c794649-hcxzr" event={"ID":"2919256c-ce36-4092-898d-024f10692bef","Type":"ContainerStarted","Data":"bc52c02b9098fced80b49edcd3c81712fcada904a84f461bcbb7d41f94b8055b"} Oct 06 10:16:21 crc kubenswrapper[4824]: I1006 10:16:21.934484 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-c7c794649-hcxzr" Oct 06 10:16:21 crc kubenswrapper[4824]: I1006 10:16:21.934494 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-c7c794649-hcxzr" event={"ID":"2919256c-ce36-4092-898d-024f10692bef","Type":"ContainerStarted","Data":"ed34426e2ca905741ee4b08ec88a1ad08c0526e4b878c08f757db85274e1aa4d"} Oct 06 10:16:21 crc kubenswrapper[4824]: I1006 10:16:21.967510 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-84b966f6c9-qwzmc" podStartSLOduration=4.967474783 podStartE2EDuration="4.967474783s" podCreationTimestamp="2025-10-06 10:16:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:16:21.958703216 +0000 UTC m=+1091.323126077" watchObservedRunningTime="2025-10-06 10:16:21.967474783 +0000 UTC m=+1091.331897644" Oct 06 10:16:21 crc kubenswrapper[4824]: I1006 10:16:21.981950 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-c7c794649-hcxzr" podStartSLOduration=2.98193372 podStartE2EDuration="2.98193372s" podCreationTimestamp="2025-10-06 10:16:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:16:21.980159296 +0000 UTC m=+1091.344582157" watchObservedRunningTime="2025-10-06 10:16:21.98193372 +0000 UTC m=+1091.346356581" Oct 06 10:16:22 crc kubenswrapper[4824]: I1006 10:16:22.947240 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-c7c794649-hcxzr" event={"ID":"2919256c-ce36-4092-898d-024f10692bef","Type":"ContainerStarted","Data":"1b850f3e625d6163ed3f456a5a98f0ed1f75b7b46c9b8fc5f2b23b1ee930ab46"} Oct 06 10:16:25 crc kubenswrapper[4824]: I1006 10:16:25.020284 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"a85f7f99-c70b-40a6-8ea2-839d91a53407","Type":"ContainerStarted","Data":"0a813108b412d32c6f3f573b57157b9dca2f4c7447b82953ab11535213306d0a"} Oct 06 10:16:25 crc kubenswrapper[4824]: I1006 10:16:25.031384 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"508d58be-fcaa-4360-ab57-fbaedb728761","Type":"ContainerStarted","Data":"e0cc57ebccd50c51a2c1ef0eccf9882e4f8ce8ab483e28f660cc3be288f38f95"} Oct 06 10:16:25 crc kubenswrapper[4824]: I1006 10:16:25.058489 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=9.057385163 podStartE2EDuration="9.057385163s" podCreationTimestamp="2025-10-06 10:16:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:16:25.053176379 +0000 UTC m=+1094.417599240" watchObservedRunningTime="2025-10-06 10:16:25.057385163 +0000 UTC m=+1094.421808024" Oct 06 10:16:26 crc kubenswrapper[4824]: I1006 10:16:26.045947 4824 generic.go:334] "Generic (PLEG): container finished" podID="73fb825b-ea3b-43c5-b143-e70310f8c1fd" containerID="90ca09e2651e49075cef289ad0aa7e54e3882d823344ef1a9a6a28355a58ff56" exitCode=0 Oct 06 10:16:26 crc kubenswrapper[4824]: I1006 10:16:26.046124 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-wlmn6" event={"ID":"73fb825b-ea3b-43c5-b143-e70310f8c1fd","Type":"ContainerDied","Data":"90ca09e2651e49075cef289ad0aa7e54e3882d823344ef1a9a6a28355a58ff56"} Oct 06 10:16:26 crc kubenswrapper[4824]: I1006 10:16:26.051441 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"a85f7f99-c70b-40a6-8ea2-839d91a53407","Type":"ContainerStarted","Data":"866e8b22170a7926f5516f9626a0711f498d48e51135ad184d51704db33d88a6"} Oct 06 10:16:26 crc kubenswrapper[4824]: I1006 10:16:26.053802 4824 generic.go:334] "Generic (PLEG): container finished" podID="554a3f37-c7c8-4d09-9da7-df5319dccecd" containerID="a787dec7530bf4c4cbfc1e0888aebec0261071050df377f25ba2aebb9d060923" exitCode=0 Oct 06 10:16:26 crc kubenswrapper[4824]: I1006 10:16:26.054038 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-fzvv8" event={"ID":"554a3f37-c7c8-4d09-9da7-df5319dccecd","Type":"ContainerDied","Data":"a787dec7530bf4c4cbfc1e0888aebec0261071050df377f25ba2aebb9d060923"} Oct 06 10:16:26 crc kubenswrapper[4824]: I1006 10:16:26.130228 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=10.130199093 podStartE2EDuration="10.130199093s" podCreationTimestamp="2025-10-06 10:16:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:16:26.119265842 +0000 UTC m=+1095.483688703" watchObservedRunningTime="2025-10-06 10:16:26.130199093 +0000 UTC m=+1095.494621964" Oct 06 10:16:26 crc kubenswrapper[4824]: I1006 10:16:26.231852 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-78c847f9f4-tv8zc" Oct 06 10:16:26 crc kubenswrapper[4824]: I1006 10:16:26.231928 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-78c847f9f4-tv8zc" Oct 06 10:16:26 crc kubenswrapper[4824]: I1006 10:16:26.309512 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-68fdc5dd48-gfq4d" Oct 06 10:16:26 crc kubenswrapper[4824]: I1006 10:16:26.310449 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-68fdc5dd48-gfq4d" Oct 06 10:16:27 crc kubenswrapper[4824]: I1006 10:16:27.105226 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 06 10:16:27 crc kubenswrapper[4824]: I1006 10:16:27.105276 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 06 10:16:27 crc kubenswrapper[4824]: I1006 10:16:27.139784 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 06 10:16:27 crc kubenswrapper[4824]: I1006 10:16:27.160965 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 06 10:16:27 crc kubenswrapper[4824]: I1006 10:16:27.161027 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 06 10:16:27 crc kubenswrapper[4824]: I1006 10:16:27.173020 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 06 10:16:27 crc kubenswrapper[4824]: I1006 10:16:27.225036 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 06 10:16:27 crc kubenswrapper[4824]: I1006 10:16:27.234414 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 06 10:16:27 crc kubenswrapper[4824]: I1006 10:16:27.994190 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-84b966f6c9-qwzmc" Oct 06 10:16:28 crc kubenswrapper[4824]: I1006 10:16:28.081073 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-9r9wx"] Oct 06 10:16:28 crc kubenswrapper[4824]: I1006 10:16:28.081767 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-8b5c85b87-9r9wx" podUID="629a3cfa-5d3a-416a-9316-8d64594f34f1" containerName="dnsmasq-dns" containerID="cri-o://0e86b0d6cbf1ecd7c499cc998c99d2bdd8a0c47ba05ae448195ee46a9eca3a32" gracePeriod=10 Oct 06 10:16:28 crc kubenswrapper[4824]: I1006 10:16:28.141674 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 06 10:16:28 crc kubenswrapper[4824]: I1006 10:16:28.141750 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 06 10:16:28 crc kubenswrapper[4824]: I1006 10:16:28.141772 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 06 10:16:28 crc kubenswrapper[4824]: I1006 10:16:28.141781 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 06 10:16:29 crc kubenswrapper[4824]: I1006 10:16:29.173371 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-fzvv8" Oct 06 10:16:29 crc kubenswrapper[4824]: I1006 10:16:29.191344 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-wlmn6" Oct 06 10:16:29 crc kubenswrapper[4824]: I1006 10:16:29.209528 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-wlmn6" event={"ID":"73fb825b-ea3b-43c5-b143-e70310f8c1fd","Type":"ContainerDied","Data":"46f6c47a06c544dcb4e13ef4f1f4782faa48cc9b568008f5dc60118b956d5e27"} Oct 06 10:16:29 crc kubenswrapper[4824]: I1006 10:16:29.209577 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="46f6c47a06c544dcb4e13ef4f1f4782faa48cc9b568008f5dc60118b956d5e27" Oct 06 10:16:29 crc kubenswrapper[4824]: I1006 10:16:29.238288 4824 generic.go:334] "Generic (PLEG): container finished" podID="629a3cfa-5d3a-416a-9316-8d64594f34f1" containerID="0e86b0d6cbf1ecd7c499cc998c99d2bdd8a0c47ba05ae448195ee46a9eca3a32" exitCode=0 Oct 06 10:16:29 crc kubenswrapper[4824]: I1006 10:16:29.238402 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b5c85b87-9r9wx" event={"ID":"629a3cfa-5d3a-416a-9316-8d64594f34f1","Type":"ContainerDied","Data":"0e86b0d6cbf1ecd7c499cc998c99d2bdd8a0c47ba05ae448195ee46a9eca3a32"} Oct 06 10:16:29 crc kubenswrapper[4824]: I1006 10:16:29.250755 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-fzvv8" Oct 06 10:16:29 crc kubenswrapper[4824]: I1006 10:16:29.250904 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-fzvv8" event={"ID":"554a3f37-c7c8-4d09-9da7-df5319dccecd","Type":"ContainerDied","Data":"696fbfd81e9841b428ae3561c99fa89ee92932fa05c3d70b8d61e032e9d7cba3"} Oct 06 10:16:29 crc kubenswrapper[4824]: I1006 10:16:29.250927 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="696fbfd81e9841b428ae3561c99fa89ee92932fa05c3d70b8d61e032e9d7cba3" Oct 06 10:16:29 crc kubenswrapper[4824]: I1006 10:16:29.255079 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/554a3f37-c7c8-4d09-9da7-df5319dccecd-logs\") pod \"554a3f37-c7c8-4d09-9da7-df5319dccecd\" (UID: \"554a3f37-c7c8-4d09-9da7-df5319dccecd\") " Oct 06 10:16:29 crc kubenswrapper[4824]: I1006 10:16:29.255159 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/554a3f37-c7c8-4d09-9da7-df5319dccecd-scripts\") pod \"554a3f37-c7c8-4d09-9da7-df5319dccecd\" (UID: \"554a3f37-c7c8-4d09-9da7-df5319dccecd\") " Oct 06 10:16:29 crc kubenswrapper[4824]: I1006 10:16:29.255482 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/554a3f37-c7c8-4d09-9da7-df5319dccecd-config-data\") pod \"554a3f37-c7c8-4d09-9da7-df5319dccecd\" (UID: \"554a3f37-c7c8-4d09-9da7-df5319dccecd\") " Oct 06 10:16:29 crc kubenswrapper[4824]: I1006 10:16:29.255654 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/554a3f37-c7c8-4d09-9da7-df5319dccecd-combined-ca-bundle\") pod \"554a3f37-c7c8-4d09-9da7-df5319dccecd\" (UID: \"554a3f37-c7c8-4d09-9da7-df5319dccecd\") " Oct 06 10:16:29 crc kubenswrapper[4824]: I1006 10:16:29.255696 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zhrg9\" (UniqueName: \"kubernetes.io/projected/554a3f37-c7c8-4d09-9da7-df5319dccecd-kube-api-access-zhrg9\") pod \"554a3f37-c7c8-4d09-9da7-df5319dccecd\" (UID: \"554a3f37-c7c8-4d09-9da7-df5319dccecd\") " Oct 06 10:16:29 crc kubenswrapper[4824]: I1006 10:16:29.257327 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/554a3f37-c7c8-4d09-9da7-df5319dccecd-logs" (OuterVolumeSpecName: "logs") pod "554a3f37-c7c8-4d09-9da7-df5319dccecd" (UID: "554a3f37-c7c8-4d09-9da7-df5319dccecd"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:16:29 crc kubenswrapper[4824]: I1006 10:16:29.265424 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/554a3f37-c7c8-4d09-9da7-df5319dccecd-kube-api-access-zhrg9" (OuterVolumeSpecName: "kube-api-access-zhrg9") pod "554a3f37-c7c8-4d09-9da7-df5319dccecd" (UID: "554a3f37-c7c8-4d09-9da7-df5319dccecd"). InnerVolumeSpecName "kube-api-access-zhrg9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:16:29 crc kubenswrapper[4824]: I1006 10:16:29.266260 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/554a3f37-c7c8-4d09-9da7-df5319dccecd-scripts" (OuterVolumeSpecName: "scripts") pod "554a3f37-c7c8-4d09-9da7-df5319dccecd" (UID: "554a3f37-c7c8-4d09-9da7-df5319dccecd"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:16:29 crc kubenswrapper[4824]: I1006 10:16:29.326064 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/554a3f37-c7c8-4d09-9da7-df5319dccecd-config-data" (OuterVolumeSpecName: "config-data") pod "554a3f37-c7c8-4d09-9da7-df5319dccecd" (UID: "554a3f37-c7c8-4d09-9da7-df5319dccecd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:16:29 crc kubenswrapper[4824]: I1006 10:16:29.332999 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/554a3f37-c7c8-4d09-9da7-df5319dccecd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "554a3f37-c7c8-4d09-9da7-df5319dccecd" (UID: "554a3f37-c7c8-4d09-9da7-df5319dccecd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:16:29 crc kubenswrapper[4824]: I1006 10:16:29.357746 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tpvs6\" (UniqueName: \"kubernetes.io/projected/73fb825b-ea3b-43c5-b143-e70310f8c1fd-kube-api-access-tpvs6\") pod \"73fb825b-ea3b-43c5-b143-e70310f8c1fd\" (UID: \"73fb825b-ea3b-43c5-b143-e70310f8c1fd\") " Oct 06 10:16:29 crc kubenswrapper[4824]: I1006 10:16:29.357827 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/73fb825b-ea3b-43c5-b143-e70310f8c1fd-credential-keys\") pod \"73fb825b-ea3b-43c5-b143-e70310f8c1fd\" (UID: \"73fb825b-ea3b-43c5-b143-e70310f8c1fd\") " Oct 06 10:16:29 crc kubenswrapper[4824]: I1006 10:16:29.357848 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/73fb825b-ea3b-43c5-b143-e70310f8c1fd-scripts\") pod \"73fb825b-ea3b-43c5-b143-e70310f8c1fd\" (UID: \"73fb825b-ea3b-43c5-b143-e70310f8c1fd\") " Oct 06 10:16:29 crc kubenswrapper[4824]: I1006 10:16:29.357882 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73fb825b-ea3b-43c5-b143-e70310f8c1fd-combined-ca-bundle\") pod \"73fb825b-ea3b-43c5-b143-e70310f8c1fd\" (UID: \"73fb825b-ea3b-43c5-b143-e70310f8c1fd\") " Oct 06 10:16:29 crc kubenswrapper[4824]: I1006 10:16:29.357918 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/73fb825b-ea3b-43c5-b143-e70310f8c1fd-config-data\") pod \"73fb825b-ea3b-43c5-b143-e70310f8c1fd\" (UID: \"73fb825b-ea3b-43c5-b143-e70310f8c1fd\") " Oct 06 10:16:29 crc kubenswrapper[4824]: I1006 10:16:29.358053 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/73fb825b-ea3b-43c5-b143-e70310f8c1fd-fernet-keys\") pod \"73fb825b-ea3b-43c5-b143-e70310f8c1fd\" (UID: \"73fb825b-ea3b-43c5-b143-e70310f8c1fd\") " Oct 06 10:16:29 crc kubenswrapper[4824]: I1006 10:16:29.359716 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/554a3f37-c7c8-4d09-9da7-df5319dccecd-scripts\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:29 crc kubenswrapper[4824]: I1006 10:16:29.359736 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/554a3f37-c7c8-4d09-9da7-df5319dccecd-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:29 crc kubenswrapper[4824]: I1006 10:16:29.359747 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/554a3f37-c7c8-4d09-9da7-df5319dccecd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:29 crc kubenswrapper[4824]: I1006 10:16:29.359944 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zhrg9\" (UniqueName: \"kubernetes.io/projected/554a3f37-c7c8-4d09-9da7-df5319dccecd-kube-api-access-zhrg9\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:29 crc kubenswrapper[4824]: I1006 10:16:29.359953 4824 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/554a3f37-c7c8-4d09-9da7-df5319dccecd-logs\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:29 crc kubenswrapper[4824]: I1006 10:16:29.374732 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/73fb825b-ea3b-43c5-b143-e70310f8c1fd-scripts" (OuterVolumeSpecName: "scripts") pod "73fb825b-ea3b-43c5-b143-e70310f8c1fd" (UID: "73fb825b-ea3b-43c5-b143-e70310f8c1fd"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:16:29 crc kubenswrapper[4824]: I1006 10:16:29.385283 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/73fb825b-ea3b-43c5-b143-e70310f8c1fd-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "73fb825b-ea3b-43c5-b143-e70310f8c1fd" (UID: "73fb825b-ea3b-43c5-b143-e70310f8c1fd"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:16:29 crc kubenswrapper[4824]: I1006 10:16:29.386716 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8b5c85b87-9r9wx" Oct 06 10:16:29 crc kubenswrapper[4824]: I1006 10:16:29.392547 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/73fb825b-ea3b-43c5-b143-e70310f8c1fd-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "73fb825b-ea3b-43c5-b143-e70310f8c1fd" (UID: "73fb825b-ea3b-43c5-b143-e70310f8c1fd"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:16:29 crc kubenswrapper[4824]: I1006 10:16:29.401266 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/73fb825b-ea3b-43c5-b143-e70310f8c1fd-kube-api-access-tpvs6" (OuterVolumeSpecName: "kube-api-access-tpvs6") pod "73fb825b-ea3b-43c5-b143-e70310f8c1fd" (UID: "73fb825b-ea3b-43c5-b143-e70310f8c1fd"). InnerVolumeSpecName "kube-api-access-tpvs6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:16:29 crc kubenswrapper[4824]: I1006 10:16:29.417195 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/73fb825b-ea3b-43c5-b143-e70310f8c1fd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "73fb825b-ea3b-43c5-b143-e70310f8c1fd" (UID: "73fb825b-ea3b-43c5-b143-e70310f8c1fd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:16:29 crc kubenswrapper[4824]: I1006 10:16:29.457365 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/73fb825b-ea3b-43c5-b143-e70310f8c1fd-config-data" (OuterVolumeSpecName: "config-data") pod "73fb825b-ea3b-43c5-b143-e70310f8c1fd" (UID: "73fb825b-ea3b-43c5-b143-e70310f8c1fd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:16:29 crc kubenswrapper[4824]: I1006 10:16:29.460634 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/629a3cfa-5d3a-416a-9316-8d64594f34f1-ovsdbserver-nb\") pod \"629a3cfa-5d3a-416a-9316-8d64594f34f1\" (UID: \"629a3cfa-5d3a-416a-9316-8d64594f34f1\") " Oct 06 10:16:29 crc kubenswrapper[4824]: I1006 10:16:29.460717 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/629a3cfa-5d3a-416a-9316-8d64594f34f1-dns-swift-storage-0\") pod \"629a3cfa-5d3a-416a-9316-8d64594f34f1\" (UID: \"629a3cfa-5d3a-416a-9316-8d64594f34f1\") " Oct 06 10:16:29 crc kubenswrapper[4824]: I1006 10:16:29.460855 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/629a3cfa-5d3a-416a-9316-8d64594f34f1-config\") pod \"629a3cfa-5d3a-416a-9316-8d64594f34f1\" (UID: \"629a3cfa-5d3a-416a-9316-8d64594f34f1\") " Oct 06 10:16:29 crc kubenswrapper[4824]: I1006 10:16:29.461038 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dk7h7\" (UniqueName: \"kubernetes.io/projected/629a3cfa-5d3a-416a-9316-8d64594f34f1-kube-api-access-dk7h7\") pod \"629a3cfa-5d3a-416a-9316-8d64594f34f1\" (UID: \"629a3cfa-5d3a-416a-9316-8d64594f34f1\") " Oct 06 10:16:29 crc kubenswrapper[4824]: I1006 10:16:29.461071 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/629a3cfa-5d3a-416a-9316-8d64594f34f1-dns-svc\") pod \"629a3cfa-5d3a-416a-9316-8d64594f34f1\" (UID: \"629a3cfa-5d3a-416a-9316-8d64594f34f1\") " Oct 06 10:16:29 crc kubenswrapper[4824]: I1006 10:16:29.461099 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/629a3cfa-5d3a-416a-9316-8d64594f34f1-ovsdbserver-sb\") pod \"629a3cfa-5d3a-416a-9316-8d64594f34f1\" (UID: \"629a3cfa-5d3a-416a-9316-8d64594f34f1\") " Oct 06 10:16:29 crc kubenswrapper[4824]: I1006 10:16:29.461614 4824 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/73fb825b-ea3b-43c5-b143-e70310f8c1fd-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:29 crc kubenswrapper[4824]: I1006 10:16:29.461649 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tpvs6\" (UniqueName: \"kubernetes.io/projected/73fb825b-ea3b-43c5-b143-e70310f8c1fd-kube-api-access-tpvs6\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:29 crc kubenswrapper[4824]: I1006 10:16:29.461664 4824 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/73fb825b-ea3b-43c5-b143-e70310f8c1fd-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:29 crc kubenswrapper[4824]: I1006 10:16:29.461676 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/73fb825b-ea3b-43c5-b143-e70310f8c1fd-scripts\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:29 crc kubenswrapper[4824]: I1006 10:16:29.461687 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73fb825b-ea3b-43c5-b143-e70310f8c1fd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:29 crc kubenswrapper[4824]: I1006 10:16:29.461699 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/73fb825b-ea3b-43c5-b143-e70310f8c1fd-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:29 crc kubenswrapper[4824]: I1006 10:16:29.482316 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/629a3cfa-5d3a-416a-9316-8d64594f34f1-kube-api-access-dk7h7" (OuterVolumeSpecName: "kube-api-access-dk7h7") pod "629a3cfa-5d3a-416a-9316-8d64594f34f1" (UID: "629a3cfa-5d3a-416a-9316-8d64594f34f1"). InnerVolumeSpecName "kube-api-access-dk7h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:16:29 crc kubenswrapper[4824]: I1006 10:16:29.565609 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dk7h7\" (UniqueName: \"kubernetes.io/projected/629a3cfa-5d3a-416a-9316-8d64594f34f1-kube-api-access-dk7h7\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:29 crc kubenswrapper[4824]: I1006 10:16:29.570632 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/629a3cfa-5d3a-416a-9316-8d64594f34f1-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "629a3cfa-5d3a-416a-9316-8d64594f34f1" (UID: "629a3cfa-5d3a-416a-9316-8d64594f34f1"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:16:29 crc kubenswrapper[4824]: I1006 10:16:29.618674 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/629a3cfa-5d3a-416a-9316-8d64594f34f1-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "629a3cfa-5d3a-416a-9316-8d64594f34f1" (UID: "629a3cfa-5d3a-416a-9316-8d64594f34f1"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:16:29 crc kubenswrapper[4824]: I1006 10:16:29.633577 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/629a3cfa-5d3a-416a-9316-8d64594f34f1-config" (OuterVolumeSpecName: "config") pod "629a3cfa-5d3a-416a-9316-8d64594f34f1" (UID: "629a3cfa-5d3a-416a-9316-8d64594f34f1"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:16:29 crc kubenswrapper[4824]: I1006 10:16:29.638728 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/629a3cfa-5d3a-416a-9316-8d64594f34f1-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "629a3cfa-5d3a-416a-9316-8d64594f34f1" (UID: "629a3cfa-5d3a-416a-9316-8d64594f34f1"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:16:29 crc kubenswrapper[4824]: I1006 10:16:29.646917 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/629a3cfa-5d3a-416a-9316-8d64594f34f1-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "629a3cfa-5d3a-416a-9316-8d64594f34f1" (UID: "629a3cfa-5d3a-416a-9316-8d64594f34f1"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:16:29 crc kubenswrapper[4824]: I1006 10:16:29.666954 4824 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/629a3cfa-5d3a-416a-9316-8d64594f34f1-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:29 crc kubenswrapper[4824]: I1006 10:16:29.667010 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/629a3cfa-5d3a-416a-9316-8d64594f34f1-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:29 crc kubenswrapper[4824]: I1006 10:16:29.667021 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/629a3cfa-5d3a-416a-9316-8d64594f34f1-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:29 crc kubenswrapper[4824]: I1006 10:16:29.667032 4824 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/629a3cfa-5d3a-416a-9316-8d64594f34f1-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:29 crc kubenswrapper[4824]: I1006 10:16:29.667042 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/629a3cfa-5d3a-416a-9316-8d64594f34f1-config\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:30 crc kubenswrapper[4824]: I1006 10:16:30.276248 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"13781991-323c-49ae-9c5d-625e2ffd019b","Type":"ContainerStarted","Data":"68acaf0ce4e83b374f4f0fabe242c7a9dd535c1809d4f2f3f0cd8f1a64d60fe5"} Oct 06 10:16:30 crc kubenswrapper[4824]: I1006 10:16:30.282375 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8b5c85b87-9r9wx" Oct 06 10:16:30 crc kubenswrapper[4824]: I1006 10:16:30.282398 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-wlmn6" Oct 06 10:16:30 crc kubenswrapper[4824]: I1006 10:16:30.282358 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b5c85b87-9r9wx" event={"ID":"629a3cfa-5d3a-416a-9316-8d64594f34f1","Type":"ContainerDied","Data":"450870ccbe15cdeceee25a40f63b1b97cc7c7dccb95f272a11c1b2d4b55929a5"} Oct 06 10:16:30 crc kubenswrapper[4824]: I1006 10:16:30.282545 4824 scope.go:117] "RemoveContainer" containerID="0e86b0d6cbf1ecd7c499cc998c99d2bdd8a0c47ba05ae448195ee46a9eca3a32" Oct 06 10:16:30 crc kubenswrapper[4824]: I1006 10:16:30.329159 4824 scope.go:117] "RemoveContainer" containerID="9617a013ed1ae48301abd05bf70da4791eb94a06a50eb5a2480c688c149922a4" Oct 06 10:16:30 crc kubenswrapper[4824]: I1006 10:16:30.358780 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-5f446cffd4-v2pxw"] Oct 06 10:16:30 crc kubenswrapper[4824]: E1006 10:16:30.359192 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73fb825b-ea3b-43c5-b143-e70310f8c1fd" containerName="keystone-bootstrap" Oct 06 10:16:30 crc kubenswrapper[4824]: I1006 10:16:30.359208 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="73fb825b-ea3b-43c5-b143-e70310f8c1fd" containerName="keystone-bootstrap" Oct 06 10:16:30 crc kubenswrapper[4824]: E1006 10:16:30.359225 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="554a3f37-c7c8-4d09-9da7-df5319dccecd" containerName="placement-db-sync" Oct 06 10:16:30 crc kubenswrapper[4824]: I1006 10:16:30.359230 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="554a3f37-c7c8-4d09-9da7-df5319dccecd" containerName="placement-db-sync" Oct 06 10:16:30 crc kubenswrapper[4824]: E1006 10:16:30.359242 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="629a3cfa-5d3a-416a-9316-8d64594f34f1" containerName="init" Oct 06 10:16:30 crc kubenswrapper[4824]: I1006 10:16:30.359248 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="629a3cfa-5d3a-416a-9316-8d64594f34f1" containerName="init" Oct 06 10:16:30 crc kubenswrapper[4824]: E1006 10:16:30.359276 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="629a3cfa-5d3a-416a-9316-8d64594f34f1" containerName="dnsmasq-dns" Oct 06 10:16:30 crc kubenswrapper[4824]: I1006 10:16:30.359282 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="629a3cfa-5d3a-416a-9316-8d64594f34f1" containerName="dnsmasq-dns" Oct 06 10:16:30 crc kubenswrapper[4824]: I1006 10:16:30.359436 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="554a3f37-c7c8-4d09-9da7-df5319dccecd" containerName="placement-db-sync" Oct 06 10:16:30 crc kubenswrapper[4824]: I1006 10:16:30.359456 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="73fb825b-ea3b-43c5-b143-e70310f8c1fd" containerName="keystone-bootstrap" Oct 06 10:16:30 crc kubenswrapper[4824]: I1006 10:16:30.359468 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="629a3cfa-5d3a-416a-9316-8d64594f34f1" containerName="dnsmasq-dns" Oct 06 10:16:30 crc kubenswrapper[4824]: I1006 10:16:30.360370 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-5f446cffd4-v2pxw" Oct 06 10:16:30 crc kubenswrapper[4824]: I1006 10:16:30.368943 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-j6t79" Oct 06 10:16:30 crc kubenswrapper[4824]: I1006 10:16:30.369226 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Oct 06 10:16:30 crc kubenswrapper[4824]: I1006 10:16:30.369333 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Oct 06 10:16:30 crc kubenswrapper[4824]: I1006 10:16:30.369509 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Oct 06 10:16:30 crc kubenswrapper[4824]: I1006 10:16:30.369294 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Oct 06 10:16:30 crc kubenswrapper[4824]: I1006 10:16:30.397800 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-9r9wx"] Oct 06 10:16:30 crc kubenswrapper[4824]: I1006 10:16:30.423653 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-9r9wx"] Oct 06 10:16:30 crc kubenswrapper[4824]: I1006 10:16:30.436282 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-5f446cffd4-v2pxw"] Oct 06 10:16:30 crc kubenswrapper[4824]: I1006 10:16:30.459501 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-fbf8cb574-lrj2t"] Oct 06 10:16:30 crc kubenswrapper[4824]: I1006 10:16:30.460894 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-fbf8cb574-lrj2t" Oct 06 10:16:30 crc kubenswrapper[4824]: I1006 10:16:30.469034 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Oct 06 10:16:30 crc kubenswrapper[4824]: I1006 10:16:30.469403 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Oct 06 10:16:30 crc kubenswrapper[4824]: I1006 10:16:30.469427 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 06 10:16:30 crc kubenswrapper[4824]: I1006 10:16:30.469451 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-lxhxz" Oct 06 10:16:30 crc kubenswrapper[4824]: I1006 10:16:30.469477 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 06 10:16:30 crc kubenswrapper[4824]: I1006 10:16:30.470349 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 06 10:16:30 crc kubenswrapper[4824]: I1006 10:16:30.502503 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4cv85\" (UniqueName: \"kubernetes.io/projected/1d3b99b4-b383-4c33-ae51-b61c611a5b3d-kube-api-access-4cv85\") pod \"placement-5f446cffd4-v2pxw\" (UID: \"1d3b99b4-b383-4c33-ae51-b61c611a5b3d\") " pod="openstack/placement-5f446cffd4-v2pxw" Oct 06 10:16:30 crc kubenswrapper[4824]: I1006 10:16:30.502822 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1d3b99b4-b383-4c33-ae51-b61c611a5b3d-scripts\") pod \"placement-5f446cffd4-v2pxw\" (UID: \"1d3b99b4-b383-4c33-ae51-b61c611a5b3d\") " pod="openstack/placement-5f446cffd4-v2pxw" Oct 06 10:16:30 crc kubenswrapper[4824]: I1006 10:16:30.503018 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d3b99b4-b383-4c33-ae51-b61c611a5b3d-config-data\") pod \"placement-5f446cffd4-v2pxw\" (UID: \"1d3b99b4-b383-4c33-ae51-b61c611a5b3d\") " pod="openstack/placement-5f446cffd4-v2pxw" Oct 06 10:16:30 crc kubenswrapper[4824]: I1006 10:16:30.503341 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1d3b99b4-b383-4c33-ae51-b61c611a5b3d-logs\") pod \"placement-5f446cffd4-v2pxw\" (UID: \"1d3b99b4-b383-4c33-ae51-b61c611a5b3d\") " pod="openstack/placement-5f446cffd4-v2pxw" Oct 06 10:16:30 crc kubenswrapper[4824]: I1006 10:16:30.503578 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1d3b99b4-b383-4c33-ae51-b61c611a5b3d-public-tls-certs\") pod \"placement-5f446cffd4-v2pxw\" (UID: \"1d3b99b4-b383-4c33-ae51-b61c611a5b3d\") " pod="openstack/placement-5f446cffd4-v2pxw" Oct 06 10:16:30 crc kubenswrapper[4824]: I1006 10:16:30.503704 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1d3b99b4-b383-4c33-ae51-b61c611a5b3d-internal-tls-certs\") pod \"placement-5f446cffd4-v2pxw\" (UID: \"1d3b99b4-b383-4c33-ae51-b61c611a5b3d\") " pod="openstack/placement-5f446cffd4-v2pxw" Oct 06 10:16:30 crc kubenswrapper[4824]: I1006 10:16:30.503778 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d3b99b4-b383-4c33-ae51-b61c611a5b3d-combined-ca-bundle\") pod \"placement-5f446cffd4-v2pxw\" (UID: \"1d3b99b4-b383-4c33-ae51-b61c611a5b3d\") " pod="openstack/placement-5f446cffd4-v2pxw" Oct 06 10:16:30 crc kubenswrapper[4824]: I1006 10:16:30.543453 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-fbf8cb574-lrj2t"] Oct 06 10:16:30 crc kubenswrapper[4824]: I1006 10:16:30.604846 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1d3b99b4-b383-4c33-ae51-b61c611a5b3d-public-tls-certs\") pod \"placement-5f446cffd4-v2pxw\" (UID: \"1d3b99b4-b383-4c33-ae51-b61c611a5b3d\") " pod="openstack/placement-5f446cffd4-v2pxw" Oct 06 10:16:30 crc kubenswrapper[4824]: I1006 10:16:30.604902 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1d3b99b4-b383-4c33-ae51-b61c611a5b3d-internal-tls-certs\") pod \"placement-5f446cffd4-v2pxw\" (UID: \"1d3b99b4-b383-4c33-ae51-b61c611a5b3d\") " pod="openstack/placement-5f446cffd4-v2pxw" Oct 06 10:16:30 crc kubenswrapper[4824]: I1006 10:16:30.604928 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d3b99b4-b383-4c33-ae51-b61c611a5b3d-combined-ca-bundle\") pod \"placement-5f446cffd4-v2pxw\" (UID: \"1d3b99b4-b383-4c33-ae51-b61c611a5b3d\") " pod="openstack/placement-5f446cffd4-v2pxw" Oct 06 10:16:30 crc kubenswrapper[4824]: I1006 10:16:30.604955 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1015a252-b07f-4398-9423-41daf44c33b6-combined-ca-bundle\") pod \"keystone-fbf8cb574-lrj2t\" (UID: \"1015a252-b07f-4398-9423-41daf44c33b6\") " pod="openstack/keystone-fbf8cb574-lrj2t" Oct 06 10:16:30 crc kubenswrapper[4824]: I1006 10:16:30.604998 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4cv85\" (UniqueName: \"kubernetes.io/projected/1d3b99b4-b383-4c33-ae51-b61c611a5b3d-kube-api-access-4cv85\") pod \"placement-5f446cffd4-v2pxw\" (UID: \"1d3b99b4-b383-4c33-ae51-b61c611a5b3d\") " pod="openstack/placement-5f446cffd4-v2pxw" Oct 06 10:16:30 crc kubenswrapper[4824]: I1006 10:16:30.605020 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1015a252-b07f-4398-9423-41daf44c33b6-public-tls-certs\") pod \"keystone-fbf8cb574-lrj2t\" (UID: \"1015a252-b07f-4398-9423-41daf44c33b6\") " pod="openstack/keystone-fbf8cb574-lrj2t" Oct 06 10:16:30 crc kubenswrapper[4824]: I1006 10:16:30.605042 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1015a252-b07f-4398-9423-41daf44c33b6-internal-tls-certs\") pod \"keystone-fbf8cb574-lrj2t\" (UID: \"1015a252-b07f-4398-9423-41daf44c33b6\") " pod="openstack/keystone-fbf8cb574-lrj2t" Oct 06 10:16:30 crc kubenswrapper[4824]: I1006 10:16:30.605058 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1015a252-b07f-4398-9423-41daf44c33b6-config-data\") pod \"keystone-fbf8cb574-lrj2t\" (UID: \"1015a252-b07f-4398-9423-41daf44c33b6\") " pod="openstack/keystone-fbf8cb574-lrj2t" Oct 06 10:16:30 crc kubenswrapper[4824]: I1006 10:16:30.605072 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1015a252-b07f-4398-9423-41daf44c33b6-fernet-keys\") pod \"keystone-fbf8cb574-lrj2t\" (UID: \"1015a252-b07f-4398-9423-41daf44c33b6\") " pod="openstack/keystone-fbf8cb574-lrj2t" Oct 06 10:16:30 crc kubenswrapper[4824]: I1006 10:16:30.605093 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1d3b99b4-b383-4c33-ae51-b61c611a5b3d-scripts\") pod \"placement-5f446cffd4-v2pxw\" (UID: \"1d3b99b4-b383-4c33-ae51-b61c611a5b3d\") " pod="openstack/placement-5f446cffd4-v2pxw" Oct 06 10:16:30 crc kubenswrapper[4824]: I1006 10:16:30.605118 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1015a252-b07f-4398-9423-41daf44c33b6-scripts\") pod \"keystone-fbf8cb574-lrj2t\" (UID: \"1015a252-b07f-4398-9423-41daf44c33b6\") " pod="openstack/keystone-fbf8cb574-lrj2t" Oct 06 10:16:30 crc kubenswrapper[4824]: I1006 10:16:30.605136 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rr69l\" (UniqueName: \"kubernetes.io/projected/1015a252-b07f-4398-9423-41daf44c33b6-kube-api-access-rr69l\") pod \"keystone-fbf8cb574-lrj2t\" (UID: \"1015a252-b07f-4398-9423-41daf44c33b6\") " pod="openstack/keystone-fbf8cb574-lrj2t" Oct 06 10:16:30 crc kubenswrapper[4824]: I1006 10:16:30.605156 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d3b99b4-b383-4c33-ae51-b61c611a5b3d-config-data\") pod \"placement-5f446cffd4-v2pxw\" (UID: \"1d3b99b4-b383-4c33-ae51-b61c611a5b3d\") " pod="openstack/placement-5f446cffd4-v2pxw" Oct 06 10:16:30 crc kubenswrapper[4824]: I1006 10:16:30.605174 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1d3b99b4-b383-4c33-ae51-b61c611a5b3d-logs\") pod \"placement-5f446cffd4-v2pxw\" (UID: \"1d3b99b4-b383-4c33-ae51-b61c611a5b3d\") " pod="openstack/placement-5f446cffd4-v2pxw" Oct 06 10:16:30 crc kubenswrapper[4824]: I1006 10:16:30.605215 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/1015a252-b07f-4398-9423-41daf44c33b6-credential-keys\") pod \"keystone-fbf8cb574-lrj2t\" (UID: \"1015a252-b07f-4398-9423-41daf44c33b6\") " pod="openstack/keystone-fbf8cb574-lrj2t" Oct 06 10:16:30 crc kubenswrapper[4824]: I1006 10:16:30.606002 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1d3b99b4-b383-4c33-ae51-b61c611a5b3d-logs\") pod \"placement-5f446cffd4-v2pxw\" (UID: \"1d3b99b4-b383-4c33-ae51-b61c611a5b3d\") " pod="openstack/placement-5f446cffd4-v2pxw" Oct 06 10:16:30 crc kubenswrapper[4824]: I1006 10:16:30.610684 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d3b99b4-b383-4c33-ae51-b61c611a5b3d-combined-ca-bundle\") pod \"placement-5f446cffd4-v2pxw\" (UID: \"1d3b99b4-b383-4c33-ae51-b61c611a5b3d\") " pod="openstack/placement-5f446cffd4-v2pxw" Oct 06 10:16:30 crc kubenswrapper[4824]: I1006 10:16:30.611692 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1d3b99b4-b383-4c33-ae51-b61c611a5b3d-scripts\") pod \"placement-5f446cffd4-v2pxw\" (UID: \"1d3b99b4-b383-4c33-ae51-b61c611a5b3d\") " pod="openstack/placement-5f446cffd4-v2pxw" Oct 06 10:16:30 crc kubenswrapper[4824]: I1006 10:16:30.612669 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d3b99b4-b383-4c33-ae51-b61c611a5b3d-config-data\") pod \"placement-5f446cffd4-v2pxw\" (UID: \"1d3b99b4-b383-4c33-ae51-b61c611a5b3d\") " pod="openstack/placement-5f446cffd4-v2pxw" Oct 06 10:16:30 crc kubenswrapper[4824]: I1006 10:16:30.622390 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1d3b99b4-b383-4c33-ae51-b61c611a5b3d-internal-tls-certs\") pod \"placement-5f446cffd4-v2pxw\" (UID: \"1d3b99b4-b383-4c33-ae51-b61c611a5b3d\") " pod="openstack/placement-5f446cffd4-v2pxw" Oct 06 10:16:30 crc kubenswrapper[4824]: I1006 10:16:30.622424 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1d3b99b4-b383-4c33-ae51-b61c611a5b3d-public-tls-certs\") pod \"placement-5f446cffd4-v2pxw\" (UID: \"1d3b99b4-b383-4c33-ae51-b61c611a5b3d\") " pod="openstack/placement-5f446cffd4-v2pxw" Oct 06 10:16:30 crc kubenswrapper[4824]: I1006 10:16:30.625423 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4cv85\" (UniqueName: \"kubernetes.io/projected/1d3b99b4-b383-4c33-ae51-b61c611a5b3d-kube-api-access-4cv85\") pod \"placement-5f446cffd4-v2pxw\" (UID: \"1d3b99b4-b383-4c33-ae51-b61c611a5b3d\") " pod="openstack/placement-5f446cffd4-v2pxw" Oct 06 10:16:30 crc kubenswrapper[4824]: I1006 10:16:30.706200 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/1015a252-b07f-4398-9423-41daf44c33b6-credential-keys\") pod \"keystone-fbf8cb574-lrj2t\" (UID: \"1015a252-b07f-4398-9423-41daf44c33b6\") " pod="openstack/keystone-fbf8cb574-lrj2t" Oct 06 10:16:30 crc kubenswrapper[4824]: I1006 10:16:30.706302 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1015a252-b07f-4398-9423-41daf44c33b6-combined-ca-bundle\") pod \"keystone-fbf8cb574-lrj2t\" (UID: \"1015a252-b07f-4398-9423-41daf44c33b6\") " pod="openstack/keystone-fbf8cb574-lrj2t" Oct 06 10:16:30 crc kubenswrapper[4824]: I1006 10:16:30.706345 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1015a252-b07f-4398-9423-41daf44c33b6-public-tls-certs\") pod \"keystone-fbf8cb574-lrj2t\" (UID: \"1015a252-b07f-4398-9423-41daf44c33b6\") " pod="openstack/keystone-fbf8cb574-lrj2t" Oct 06 10:16:30 crc kubenswrapper[4824]: I1006 10:16:30.706370 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1015a252-b07f-4398-9423-41daf44c33b6-internal-tls-certs\") pod \"keystone-fbf8cb574-lrj2t\" (UID: \"1015a252-b07f-4398-9423-41daf44c33b6\") " pod="openstack/keystone-fbf8cb574-lrj2t" Oct 06 10:16:30 crc kubenswrapper[4824]: I1006 10:16:30.706387 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1015a252-b07f-4398-9423-41daf44c33b6-config-data\") pod \"keystone-fbf8cb574-lrj2t\" (UID: \"1015a252-b07f-4398-9423-41daf44c33b6\") " pod="openstack/keystone-fbf8cb574-lrj2t" Oct 06 10:16:30 crc kubenswrapper[4824]: I1006 10:16:30.706405 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1015a252-b07f-4398-9423-41daf44c33b6-fernet-keys\") pod \"keystone-fbf8cb574-lrj2t\" (UID: \"1015a252-b07f-4398-9423-41daf44c33b6\") " pod="openstack/keystone-fbf8cb574-lrj2t" Oct 06 10:16:30 crc kubenswrapper[4824]: I1006 10:16:30.706434 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1015a252-b07f-4398-9423-41daf44c33b6-scripts\") pod \"keystone-fbf8cb574-lrj2t\" (UID: \"1015a252-b07f-4398-9423-41daf44c33b6\") " pod="openstack/keystone-fbf8cb574-lrj2t" Oct 06 10:16:30 crc kubenswrapper[4824]: I1006 10:16:30.706455 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rr69l\" (UniqueName: \"kubernetes.io/projected/1015a252-b07f-4398-9423-41daf44c33b6-kube-api-access-rr69l\") pod \"keystone-fbf8cb574-lrj2t\" (UID: \"1015a252-b07f-4398-9423-41daf44c33b6\") " pod="openstack/keystone-fbf8cb574-lrj2t" Oct 06 10:16:30 crc kubenswrapper[4824]: I1006 10:16:30.711448 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1015a252-b07f-4398-9423-41daf44c33b6-combined-ca-bundle\") pod \"keystone-fbf8cb574-lrj2t\" (UID: \"1015a252-b07f-4398-9423-41daf44c33b6\") " pod="openstack/keystone-fbf8cb574-lrj2t" Oct 06 10:16:30 crc kubenswrapper[4824]: I1006 10:16:30.711656 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/1015a252-b07f-4398-9423-41daf44c33b6-credential-keys\") pod \"keystone-fbf8cb574-lrj2t\" (UID: \"1015a252-b07f-4398-9423-41daf44c33b6\") " pod="openstack/keystone-fbf8cb574-lrj2t" Oct 06 10:16:30 crc kubenswrapper[4824]: I1006 10:16:30.712126 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1015a252-b07f-4398-9423-41daf44c33b6-public-tls-certs\") pod \"keystone-fbf8cb574-lrj2t\" (UID: \"1015a252-b07f-4398-9423-41daf44c33b6\") " pod="openstack/keystone-fbf8cb574-lrj2t" Oct 06 10:16:30 crc kubenswrapper[4824]: I1006 10:16:30.718520 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1015a252-b07f-4398-9423-41daf44c33b6-internal-tls-certs\") pod \"keystone-fbf8cb574-lrj2t\" (UID: \"1015a252-b07f-4398-9423-41daf44c33b6\") " pod="openstack/keystone-fbf8cb574-lrj2t" Oct 06 10:16:30 crc kubenswrapper[4824]: I1006 10:16:30.718897 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1015a252-b07f-4398-9423-41daf44c33b6-fernet-keys\") pod \"keystone-fbf8cb574-lrj2t\" (UID: \"1015a252-b07f-4398-9423-41daf44c33b6\") " pod="openstack/keystone-fbf8cb574-lrj2t" Oct 06 10:16:30 crc kubenswrapper[4824]: I1006 10:16:30.726856 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1015a252-b07f-4398-9423-41daf44c33b6-config-data\") pod \"keystone-fbf8cb574-lrj2t\" (UID: \"1015a252-b07f-4398-9423-41daf44c33b6\") " pod="openstack/keystone-fbf8cb574-lrj2t" Oct 06 10:16:30 crc kubenswrapper[4824]: I1006 10:16:30.730568 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rr69l\" (UniqueName: \"kubernetes.io/projected/1015a252-b07f-4398-9423-41daf44c33b6-kube-api-access-rr69l\") pod \"keystone-fbf8cb574-lrj2t\" (UID: \"1015a252-b07f-4398-9423-41daf44c33b6\") " pod="openstack/keystone-fbf8cb574-lrj2t" Oct 06 10:16:30 crc kubenswrapper[4824]: I1006 10:16:30.739414 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-5f446cffd4-v2pxw" Oct 06 10:16:30 crc kubenswrapper[4824]: I1006 10:16:30.742327 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1015a252-b07f-4398-9423-41daf44c33b6-scripts\") pod \"keystone-fbf8cb574-lrj2t\" (UID: \"1015a252-b07f-4398-9423-41daf44c33b6\") " pod="openstack/keystone-fbf8cb574-lrj2t" Oct 06 10:16:30 crc kubenswrapper[4824]: I1006 10:16:30.802290 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-fbf8cb574-lrj2t" Oct 06 10:16:31 crc kubenswrapper[4824]: I1006 10:16:31.158054 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 06 10:16:31 crc kubenswrapper[4824]: I1006 10:16:31.161365 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 06 10:16:31 crc kubenswrapper[4824]: I1006 10:16:31.305123 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="629a3cfa-5d3a-416a-9316-8d64594f34f1" path="/var/lib/kubelet/pods/629a3cfa-5d3a-416a-9316-8d64594f34f1/volumes" Oct 06 10:16:31 crc kubenswrapper[4824]: I1006 10:16:31.308359 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-5f446cffd4-v2pxw"] Oct 06 10:16:31 crc kubenswrapper[4824]: I1006 10:16:31.390841 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-fbf8cb574-lrj2t"] Oct 06 10:16:32 crc kubenswrapper[4824]: I1006 10:16:32.284626 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 06 10:16:32 crc kubenswrapper[4824]: I1006 10:16:32.311192 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 06 10:16:32 crc kubenswrapper[4824]: I1006 10:16:32.343456 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-5f446cffd4-v2pxw" event={"ID":"1d3b99b4-b383-4c33-ae51-b61c611a5b3d","Type":"ContainerStarted","Data":"1f3bd650731953135fcd9d7f81365b87516449e4230004f44b1873d448478cf7"} Oct 06 10:16:32 crc kubenswrapper[4824]: I1006 10:16:32.343511 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-5f446cffd4-v2pxw" event={"ID":"1d3b99b4-b383-4c33-ae51-b61c611a5b3d","Type":"ContainerStarted","Data":"2a39a6fd7b550215bf0a31875b906eb3af6ff1fc040df45125c1ff33080af8b0"} Oct 06 10:16:32 crc kubenswrapper[4824]: I1006 10:16:32.343525 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-5f446cffd4-v2pxw" event={"ID":"1d3b99b4-b383-4c33-ae51-b61c611a5b3d","Type":"ContainerStarted","Data":"6774de994ea627ad7737783f9f3910263f63b687f2662f84c111d65ea0151537"} Oct 06 10:16:32 crc kubenswrapper[4824]: I1006 10:16:32.344730 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-5f446cffd4-v2pxw" Oct 06 10:16:32 crc kubenswrapper[4824]: I1006 10:16:32.344766 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-5f446cffd4-v2pxw" Oct 06 10:16:32 crc kubenswrapper[4824]: I1006 10:16:32.379848 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-ljqzg" event={"ID":"dfc3bcd2-9baa-4a3a-8588-295e692d8e3e","Type":"ContainerStarted","Data":"429cd910467ac58cccbd723e79ad017a0046e53a5d5f63976b2ebc8c40185f77"} Oct 06 10:16:32 crc kubenswrapper[4824]: I1006 10:16:32.393818 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-5f446cffd4-v2pxw" podStartSLOduration=2.393790388 podStartE2EDuration="2.393790388s" podCreationTimestamp="2025-10-06 10:16:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:16:32.383816262 +0000 UTC m=+1101.748239123" watchObservedRunningTime="2025-10-06 10:16:32.393790388 +0000 UTC m=+1101.758213249" Oct 06 10:16:32 crc kubenswrapper[4824]: I1006 10:16:32.399177 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-fbf8cb574-lrj2t" event={"ID":"1015a252-b07f-4398-9423-41daf44c33b6","Type":"ContainerStarted","Data":"3d8dca34c44739f95ed4220cb20f3ba783f079e0e9988382482957d575688562"} Oct 06 10:16:32 crc kubenswrapper[4824]: I1006 10:16:32.399227 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-fbf8cb574-lrj2t" event={"ID":"1015a252-b07f-4398-9423-41daf44c33b6","Type":"ContainerStarted","Data":"7850f71655620afd50ad746343c8dc90e20a416b11256dfc5997309753555e7a"} Oct 06 10:16:32 crc kubenswrapper[4824]: I1006 10:16:32.423110 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-ljqzg" podStartSLOduration=2.970577023 podStartE2EDuration="46.423092583s" podCreationTimestamp="2025-10-06 10:15:46 +0000 UTC" firstStartedPulling="2025-10-06 10:15:48.469324205 +0000 UTC m=+1057.833747066" lastFinishedPulling="2025-10-06 10:16:31.921839765 +0000 UTC m=+1101.286262626" observedRunningTime="2025-10-06 10:16:32.412340507 +0000 UTC m=+1101.776763368" watchObservedRunningTime="2025-10-06 10:16:32.423092583 +0000 UTC m=+1101.787515444" Oct 06 10:16:32 crc kubenswrapper[4824]: I1006 10:16:32.436400 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-fbf8cb574-lrj2t" podStartSLOduration=2.436380461 podStartE2EDuration="2.436380461s" podCreationTimestamp="2025-10-06 10:16:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:16:32.432966727 +0000 UTC m=+1101.797389598" watchObservedRunningTime="2025-10-06 10:16:32.436380461 +0000 UTC m=+1101.800803322" Oct 06 10:16:33 crc kubenswrapper[4824]: I1006 10:16:33.409109 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-fbf8cb574-lrj2t" Oct 06 10:16:34 crc kubenswrapper[4824]: I1006 10:16:34.422211 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-ldn9v" event={"ID":"2382b80c-75fc-435f-be5f-0f1eb6a5cd17","Type":"ContainerStarted","Data":"bfe65796f2dadbf11294a2fca859967193ea63bf8d8080fbb171c2311304906e"} Oct 06 10:16:34 crc kubenswrapper[4824]: I1006 10:16:34.441116 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-ldn9v" podStartSLOduration=3.446149703 podStartE2EDuration="48.441100111s" podCreationTimestamp="2025-10-06 10:15:46 +0000 UTC" firstStartedPulling="2025-10-06 10:15:47.802166007 +0000 UTC m=+1057.166588868" lastFinishedPulling="2025-10-06 10:16:32.797116415 +0000 UTC m=+1102.161539276" observedRunningTime="2025-10-06 10:16:34.440359872 +0000 UTC m=+1103.804782743" watchObservedRunningTime="2025-10-06 10:16:34.441100111 +0000 UTC m=+1103.805522972" Oct 06 10:16:35 crc kubenswrapper[4824]: I1006 10:16:35.436661 4824 generic.go:334] "Generic (PLEG): container finished" podID="dfc3bcd2-9baa-4a3a-8588-295e692d8e3e" containerID="429cd910467ac58cccbd723e79ad017a0046e53a5d5f63976b2ebc8c40185f77" exitCode=0 Oct 06 10:16:35 crc kubenswrapper[4824]: I1006 10:16:35.437087 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-ljqzg" event={"ID":"dfc3bcd2-9baa-4a3a-8588-295e692d8e3e","Type":"ContainerDied","Data":"429cd910467ac58cccbd723e79ad017a0046e53a5d5f63976b2ebc8c40185f77"} Oct 06 10:16:36 crc kubenswrapper[4824]: I1006 10:16:36.234609 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-78c847f9f4-tv8zc" podUID="68ef92a4-c4bc-4acc-be58-e14010395b8e" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.149:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.149:8443: connect: connection refused" Oct 06 10:16:36 crc kubenswrapper[4824]: I1006 10:16:36.310888 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-68fdc5dd48-gfq4d" podUID="194bef17-98ac-413f-a91e-0abc9835f2f6" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.150:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.150:8443: connect: connection refused" Oct 06 10:16:37 crc kubenswrapper[4824]: I1006 10:16:37.380376 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-ljqzg" Oct 06 10:16:37 crc kubenswrapper[4824]: I1006 10:16:37.443560 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dfc3bcd2-9baa-4a3a-8588-295e692d8e3e-combined-ca-bundle\") pod \"dfc3bcd2-9baa-4a3a-8588-295e692d8e3e\" (UID: \"dfc3bcd2-9baa-4a3a-8588-295e692d8e3e\") " Oct 06 10:16:37 crc kubenswrapper[4824]: I1006 10:16:37.444261 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vcs8n\" (UniqueName: \"kubernetes.io/projected/dfc3bcd2-9baa-4a3a-8588-295e692d8e3e-kube-api-access-vcs8n\") pod \"dfc3bcd2-9baa-4a3a-8588-295e692d8e3e\" (UID: \"dfc3bcd2-9baa-4a3a-8588-295e692d8e3e\") " Oct 06 10:16:37 crc kubenswrapper[4824]: I1006 10:16:37.444439 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/dfc3bcd2-9baa-4a3a-8588-295e692d8e3e-db-sync-config-data\") pod \"dfc3bcd2-9baa-4a3a-8588-295e692d8e3e\" (UID: \"dfc3bcd2-9baa-4a3a-8588-295e692d8e3e\") " Oct 06 10:16:37 crc kubenswrapper[4824]: I1006 10:16:37.455027 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dfc3bcd2-9baa-4a3a-8588-295e692d8e3e-kube-api-access-vcs8n" (OuterVolumeSpecName: "kube-api-access-vcs8n") pod "dfc3bcd2-9baa-4a3a-8588-295e692d8e3e" (UID: "dfc3bcd2-9baa-4a3a-8588-295e692d8e3e"). InnerVolumeSpecName "kube-api-access-vcs8n". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:16:37 crc kubenswrapper[4824]: I1006 10:16:37.473106 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dfc3bcd2-9baa-4a3a-8588-295e692d8e3e-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "dfc3bcd2-9baa-4a3a-8588-295e692d8e3e" (UID: "dfc3bcd2-9baa-4a3a-8588-295e692d8e3e"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:16:37 crc kubenswrapper[4824]: I1006 10:16:37.498539 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-ljqzg" event={"ID":"dfc3bcd2-9baa-4a3a-8588-295e692d8e3e","Type":"ContainerDied","Data":"f67082f14f4158b88ea1fa1ca65841ad86067e9e8e4793f2a079d532cdac363d"} Oct 06 10:16:37 crc kubenswrapper[4824]: I1006 10:16:37.498624 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f67082f14f4158b88ea1fa1ca65841ad86067e9e8e4793f2a079d532cdac363d" Oct 06 10:16:37 crc kubenswrapper[4824]: I1006 10:16:37.498721 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-ljqzg" Oct 06 10:16:37 crc kubenswrapper[4824]: I1006 10:16:37.532252 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dfc3bcd2-9baa-4a3a-8588-295e692d8e3e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "dfc3bcd2-9baa-4a3a-8588-295e692d8e3e" (UID: "dfc3bcd2-9baa-4a3a-8588-295e692d8e3e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:16:37 crc kubenswrapper[4824]: I1006 10:16:37.547266 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dfc3bcd2-9baa-4a3a-8588-295e692d8e3e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:37 crc kubenswrapper[4824]: I1006 10:16:37.547635 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vcs8n\" (UniqueName: \"kubernetes.io/projected/dfc3bcd2-9baa-4a3a-8588-295e692d8e3e-kube-api-access-vcs8n\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:37 crc kubenswrapper[4824]: I1006 10:16:37.547749 4824 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/dfc3bcd2-9baa-4a3a-8588-295e692d8e3e-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:37 crc kubenswrapper[4824]: I1006 10:16:37.875052 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-659dd56d6c-qpxtd"] Oct 06 10:16:37 crc kubenswrapper[4824]: E1006 10:16:37.876381 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dfc3bcd2-9baa-4a3a-8588-295e692d8e3e" containerName="barbican-db-sync" Oct 06 10:16:37 crc kubenswrapper[4824]: I1006 10:16:37.876399 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="dfc3bcd2-9baa-4a3a-8588-295e692d8e3e" containerName="barbican-db-sync" Oct 06 10:16:37 crc kubenswrapper[4824]: I1006 10:16:37.876808 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="dfc3bcd2-9baa-4a3a-8588-295e692d8e3e" containerName="barbican-db-sync" Oct 06 10:16:37 crc kubenswrapper[4824]: I1006 10:16:37.880319 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-659dd56d6c-qpxtd" Oct 06 10:16:37 crc kubenswrapper[4824]: I1006 10:16:37.886072 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-zw6pr" Oct 06 10:16:37 crc kubenswrapper[4824]: I1006 10:16:37.886522 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Oct 06 10:16:37 crc kubenswrapper[4824]: I1006 10:16:37.886829 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Oct 06 10:16:37 crc kubenswrapper[4824]: I1006 10:16:37.951074 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-659dd56d6c-qpxtd"] Oct 06 10:16:37 crc kubenswrapper[4824]: I1006 10:16:37.968259 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb46f73f-9d47-4189-990a-051fed587851-config-data\") pod \"barbican-worker-659dd56d6c-qpxtd\" (UID: \"fb46f73f-9d47-4189-990a-051fed587851\") " pod="openstack/barbican-worker-659dd56d6c-qpxtd" Oct 06 10:16:37 crc kubenswrapper[4824]: I1006 10:16:37.968316 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fb46f73f-9d47-4189-990a-051fed587851-config-data-custom\") pod \"barbican-worker-659dd56d6c-qpxtd\" (UID: \"fb46f73f-9d47-4189-990a-051fed587851\") " pod="openstack/barbican-worker-659dd56d6c-qpxtd" Oct 06 10:16:37 crc kubenswrapper[4824]: I1006 10:16:37.968376 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fb46f73f-9d47-4189-990a-051fed587851-logs\") pod \"barbican-worker-659dd56d6c-qpxtd\" (UID: \"fb46f73f-9d47-4189-990a-051fed587851\") " pod="openstack/barbican-worker-659dd56d6c-qpxtd" Oct 06 10:16:37 crc kubenswrapper[4824]: I1006 10:16:37.968464 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb46f73f-9d47-4189-990a-051fed587851-combined-ca-bundle\") pod \"barbican-worker-659dd56d6c-qpxtd\" (UID: \"fb46f73f-9d47-4189-990a-051fed587851\") " pod="openstack/barbican-worker-659dd56d6c-qpxtd" Oct 06 10:16:37 crc kubenswrapper[4824]: I1006 10:16:37.968549 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hrvw7\" (UniqueName: \"kubernetes.io/projected/fb46f73f-9d47-4189-990a-051fed587851-kube-api-access-hrvw7\") pod \"barbican-worker-659dd56d6c-qpxtd\" (UID: \"fb46f73f-9d47-4189-990a-051fed587851\") " pod="openstack/barbican-worker-659dd56d6c-qpxtd" Oct 06 10:16:37 crc kubenswrapper[4824]: I1006 10:16:37.992155 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-67979769c4-7fg5z"] Oct 06 10:16:37 crc kubenswrapper[4824]: I1006 10:16:37.994106 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-67979769c4-7fg5z" Oct 06 10:16:38 crc kubenswrapper[4824]: I1006 10:16:38.001480 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Oct 06 10:16:38 crc kubenswrapper[4824]: I1006 10:16:38.011027 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-67979769c4-7fg5z"] Oct 06 10:16:38 crc kubenswrapper[4824]: I1006 10:16:38.053033 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-mgsxh"] Oct 06 10:16:38 crc kubenswrapper[4824]: I1006 10:16:38.055376 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75c8ddd69c-mgsxh" Oct 06 10:16:38 crc kubenswrapper[4824]: I1006 10:16:38.066869 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-mgsxh"] Oct 06 10:16:38 crc kubenswrapper[4824]: I1006 10:16:38.071109 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5e844107-f8f2-48e8-9731-84eff0f8c752-config-data-custom\") pod \"barbican-keystone-listener-67979769c4-7fg5z\" (UID: \"5e844107-f8f2-48e8-9731-84eff0f8c752\") " pod="openstack/barbican-keystone-listener-67979769c4-7fg5z" Oct 06 10:16:38 crc kubenswrapper[4824]: I1006 10:16:38.071178 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rpqsz\" (UniqueName: \"kubernetes.io/projected/932b2879-a6ac-4179-946e-27d4acc68626-kube-api-access-rpqsz\") pod \"dnsmasq-dns-75c8ddd69c-mgsxh\" (UID: \"932b2879-a6ac-4179-946e-27d4acc68626\") " pod="openstack/dnsmasq-dns-75c8ddd69c-mgsxh" Oct 06 10:16:38 crc kubenswrapper[4824]: I1006 10:16:38.071218 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hrvw7\" (UniqueName: \"kubernetes.io/projected/fb46f73f-9d47-4189-990a-051fed587851-kube-api-access-hrvw7\") pod \"barbican-worker-659dd56d6c-qpxtd\" (UID: \"fb46f73f-9d47-4189-990a-051fed587851\") " pod="openstack/barbican-worker-659dd56d6c-qpxtd" Oct 06 10:16:38 crc kubenswrapper[4824]: I1006 10:16:38.071262 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/932b2879-a6ac-4179-946e-27d4acc68626-dns-swift-storage-0\") pod \"dnsmasq-dns-75c8ddd69c-mgsxh\" (UID: \"932b2879-a6ac-4179-946e-27d4acc68626\") " pod="openstack/dnsmasq-dns-75c8ddd69c-mgsxh" Oct 06 10:16:38 crc kubenswrapper[4824]: I1006 10:16:38.071293 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb46f73f-9d47-4189-990a-051fed587851-config-data\") pod \"barbican-worker-659dd56d6c-qpxtd\" (UID: \"fb46f73f-9d47-4189-990a-051fed587851\") " pod="openstack/barbican-worker-659dd56d6c-qpxtd" Oct 06 10:16:38 crc kubenswrapper[4824]: I1006 10:16:38.071718 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e844107-f8f2-48e8-9731-84eff0f8c752-config-data\") pod \"barbican-keystone-listener-67979769c4-7fg5z\" (UID: \"5e844107-f8f2-48e8-9731-84eff0f8c752\") " pod="openstack/barbican-keystone-listener-67979769c4-7fg5z" Oct 06 10:16:38 crc kubenswrapper[4824]: I1006 10:16:38.071767 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fb46f73f-9d47-4189-990a-051fed587851-config-data-custom\") pod \"barbican-worker-659dd56d6c-qpxtd\" (UID: \"fb46f73f-9d47-4189-990a-051fed587851\") " pod="openstack/barbican-worker-659dd56d6c-qpxtd" Oct 06 10:16:38 crc kubenswrapper[4824]: I1006 10:16:38.071825 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/932b2879-a6ac-4179-946e-27d4acc68626-ovsdbserver-nb\") pod \"dnsmasq-dns-75c8ddd69c-mgsxh\" (UID: \"932b2879-a6ac-4179-946e-27d4acc68626\") " pod="openstack/dnsmasq-dns-75c8ddd69c-mgsxh" Oct 06 10:16:38 crc kubenswrapper[4824]: I1006 10:16:38.071848 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j4ghd\" (UniqueName: \"kubernetes.io/projected/5e844107-f8f2-48e8-9731-84eff0f8c752-kube-api-access-j4ghd\") pod \"barbican-keystone-listener-67979769c4-7fg5z\" (UID: \"5e844107-f8f2-48e8-9731-84eff0f8c752\") " pod="openstack/barbican-keystone-listener-67979769c4-7fg5z" Oct 06 10:16:38 crc kubenswrapper[4824]: I1006 10:16:38.071927 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fb46f73f-9d47-4189-990a-051fed587851-logs\") pod \"barbican-worker-659dd56d6c-qpxtd\" (UID: \"fb46f73f-9d47-4189-990a-051fed587851\") " pod="openstack/barbican-worker-659dd56d6c-qpxtd" Oct 06 10:16:38 crc kubenswrapper[4824]: I1006 10:16:38.071954 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/932b2879-a6ac-4179-946e-27d4acc68626-ovsdbserver-sb\") pod \"dnsmasq-dns-75c8ddd69c-mgsxh\" (UID: \"932b2879-a6ac-4179-946e-27d4acc68626\") " pod="openstack/dnsmasq-dns-75c8ddd69c-mgsxh" Oct 06 10:16:38 crc kubenswrapper[4824]: I1006 10:16:38.072007 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/932b2879-a6ac-4179-946e-27d4acc68626-config\") pod \"dnsmasq-dns-75c8ddd69c-mgsxh\" (UID: \"932b2879-a6ac-4179-946e-27d4acc68626\") " pod="openstack/dnsmasq-dns-75c8ddd69c-mgsxh" Oct 06 10:16:38 crc kubenswrapper[4824]: I1006 10:16:38.072117 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5e844107-f8f2-48e8-9731-84eff0f8c752-logs\") pod \"barbican-keystone-listener-67979769c4-7fg5z\" (UID: \"5e844107-f8f2-48e8-9731-84eff0f8c752\") " pod="openstack/barbican-keystone-listener-67979769c4-7fg5z" Oct 06 10:16:38 crc kubenswrapper[4824]: I1006 10:16:38.072200 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/932b2879-a6ac-4179-946e-27d4acc68626-dns-svc\") pod \"dnsmasq-dns-75c8ddd69c-mgsxh\" (UID: \"932b2879-a6ac-4179-946e-27d4acc68626\") " pod="openstack/dnsmasq-dns-75c8ddd69c-mgsxh" Oct 06 10:16:38 crc kubenswrapper[4824]: I1006 10:16:38.072235 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb46f73f-9d47-4189-990a-051fed587851-combined-ca-bundle\") pod \"barbican-worker-659dd56d6c-qpxtd\" (UID: \"fb46f73f-9d47-4189-990a-051fed587851\") " pod="openstack/barbican-worker-659dd56d6c-qpxtd" Oct 06 10:16:38 crc kubenswrapper[4824]: I1006 10:16:38.072283 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e844107-f8f2-48e8-9731-84eff0f8c752-combined-ca-bundle\") pod \"barbican-keystone-listener-67979769c4-7fg5z\" (UID: \"5e844107-f8f2-48e8-9731-84eff0f8c752\") " pod="openstack/barbican-keystone-listener-67979769c4-7fg5z" Oct 06 10:16:38 crc kubenswrapper[4824]: I1006 10:16:38.076020 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fb46f73f-9d47-4189-990a-051fed587851-logs\") pod \"barbican-worker-659dd56d6c-qpxtd\" (UID: \"fb46f73f-9d47-4189-990a-051fed587851\") " pod="openstack/barbican-worker-659dd56d6c-qpxtd" Oct 06 10:16:38 crc kubenswrapper[4824]: I1006 10:16:38.078527 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb46f73f-9d47-4189-990a-051fed587851-combined-ca-bundle\") pod \"barbican-worker-659dd56d6c-qpxtd\" (UID: \"fb46f73f-9d47-4189-990a-051fed587851\") " pod="openstack/barbican-worker-659dd56d6c-qpxtd" Oct 06 10:16:38 crc kubenswrapper[4824]: I1006 10:16:38.100724 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb46f73f-9d47-4189-990a-051fed587851-config-data\") pod \"barbican-worker-659dd56d6c-qpxtd\" (UID: \"fb46f73f-9d47-4189-990a-051fed587851\") " pod="openstack/barbican-worker-659dd56d6c-qpxtd" Oct 06 10:16:38 crc kubenswrapper[4824]: I1006 10:16:38.101270 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fb46f73f-9d47-4189-990a-051fed587851-config-data-custom\") pod \"barbican-worker-659dd56d6c-qpxtd\" (UID: \"fb46f73f-9d47-4189-990a-051fed587851\") " pod="openstack/barbican-worker-659dd56d6c-qpxtd" Oct 06 10:16:38 crc kubenswrapper[4824]: I1006 10:16:38.107527 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hrvw7\" (UniqueName: \"kubernetes.io/projected/fb46f73f-9d47-4189-990a-051fed587851-kube-api-access-hrvw7\") pod \"barbican-worker-659dd56d6c-qpxtd\" (UID: \"fb46f73f-9d47-4189-990a-051fed587851\") " pod="openstack/barbican-worker-659dd56d6c-qpxtd" Oct 06 10:16:38 crc kubenswrapper[4824]: I1006 10:16:38.175650 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j4ghd\" (UniqueName: \"kubernetes.io/projected/5e844107-f8f2-48e8-9731-84eff0f8c752-kube-api-access-j4ghd\") pod \"barbican-keystone-listener-67979769c4-7fg5z\" (UID: \"5e844107-f8f2-48e8-9731-84eff0f8c752\") " pod="openstack/barbican-keystone-listener-67979769c4-7fg5z" Oct 06 10:16:38 crc kubenswrapper[4824]: I1006 10:16:38.175743 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/932b2879-a6ac-4179-946e-27d4acc68626-ovsdbserver-sb\") pod \"dnsmasq-dns-75c8ddd69c-mgsxh\" (UID: \"932b2879-a6ac-4179-946e-27d4acc68626\") " pod="openstack/dnsmasq-dns-75c8ddd69c-mgsxh" Oct 06 10:16:38 crc kubenswrapper[4824]: I1006 10:16:38.175775 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/932b2879-a6ac-4179-946e-27d4acc68626-config\") pod \"dnsmasq-dns-75c8ddd69c-mgsxh\" (UID: \"932b2879-a6ac-4179-946e-27d4acc68626\") " pod="openstack/dnsmasq-dns-75c8ddd69c-mgsxh" Oct 06 10:16:38 crc kubenswrapper[4824]: I1006 10:16:38.175835 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5e844107-f8f2-48e8-9731-84eff0f8c752-logs\") pod \"barbican-keystone-listener-67979769c4-7fg5z\" (UID: \"5e844107-f8f2-48e8-9731-84eff0f8c752\") " pod="openstack/barbican-keystone-listener-67979769c4-7fg5z" Oct 06 10:16:38 crc kubenswrapper[4824]: I1006 10:16:38.175891 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/932b2879-a6ac-4179-946e-27d4acc68626-dns-svc\") pod \"dnsmasq-dns-75c8ddd69c-mgsxh\" (UID: \"932b2879-a6ac-4179-946e-27d4acc68626\") " pod="openstack/dnsmasq-dns-75c8ddd69c-mgsxh" Oct 06 10:16:38 crc kubenswrapper[4824]: I1006 10:16:38.175925 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e844107-f8f2-48e8-9731-84eff0f8c752-combined-ca-bundle\") pod \"barbican-keystone-listener-67979769c4-7fg5z\" (UID: \"5e844107-f8f2-48e8-9731-84eff0f8c752\") " pod="openstack/barbican-keystone-listener-67979769c4-7fg5z" Oct 06 10:16:38 crc kubenswrapper[4824]: I1006 10:16:38.175993 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5e844107-f8f2-48e8-9731-84eff0f8c752-config-data-custom\") pod \"barbican-keystone-listener-67979769c4-7fg5z\" (UID: \"5e844107-f8f2-48e8-9731-84eff0f8c752\") " pod="openstack/barbican-keystone-listener-67979769c4-7fg5z" Oct 06 10:16:38 crc kubenswrapper[4824]: I1006 10:16:38.176020 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rpqsz\" (UniqueName: \"kubernetes.io/projected/932b2879-a6ac-4179-946e-27d4acc68626-kube-api-access-rpqsz\") pod \"dnsmasq-dns-75c8ddd69c-mgsxh\" (UID: \"932b2879-a6ac-4179-946e-27d4acc68626\") " pod="openstack/dnsmasq-dns-75c8ddd69c-mgsxh" Oct 06 10:16:38 crc kubenswrapper[4824]: I1006 10:16:38.176046 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/932b2879-a6ac-4179-946e-27d4acc68626-dns-swift-storage-0\") pod \"dnsmasq-dns-75c8ddd69c-mgsxh\" (UID: \"932b2879-a6ac-4179-946e-27d4acc68626\") " pod="openstack/dnsmasq-dns-75c8ddd69c-mgsxh" Oct 06 10:16:38 crc kubenswrapper[4824]: I1006 10:16:38.176070 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e844107-f8f2-48e8-9731-84eff0f8c752-config-data\") pod \"barbican-keystone-listener-67979769c4-7fg5z\" (UID: \"5e844107-f8f2-48e8-9731-84eff0f8c752\") " pod="openstack/barbican-keystone-listener-67979769c4-7fg5z" Oct 06 10:16:38 crc kubenswrapper[4824]: I1006 10:16:38.176100 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/932b2879-a6ac-4179-946e-27d4acc68626-ovsdbserver-nb\") pod \"dnsmasq-dns-75c8ddd69c-mgsxh\" (UID: \"932b2879-a6ac-4179-946e-27d4acc68626\") " pod="openstack/dnsmasq-dns-75c8ddd69c-mgsxh" Oct 06 10:16:38 crc kubenswrapper[4824]: I1006 10:16:38.176514 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5e844107-f8f2-48e8-9731-84eff0f8c752-logs\") pod \"barbican-keystone-listener-67979769c4-7fg5z\" (UID: \"5e844107-f8f2-48e8-9731-84eff0f8c752\") " pod="openstack/barbican-keystone-listener-67979769c4-7fg5z" Oct 06 10:16:38 crc kubenswrapper[4824]: I1006 10:16:38.177086 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/932b2879-a6ac-4179-946e-27d4acc68626-ovsdbserver-sb\") pod \"dnsmasq-dns-75c8ddd69c-mgsxh\" (UID: \"932b2879-a6ac-4179-946e-27d4acc68626\") " pod="openstack/dnsmasq-dns-75c8ddd69c-mgsxh" Oct 06 10:16:38 crc kubenswrapper[4824]: I1006 10:16:38.177167 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/932b2879-a6ac-4179-946e-27d4acc68626-config\") pod \"dnsmasq-dns-75c8ddd69c-mgsxh\" (UID: \"932b2879-a6ac-4179-946e-27d4acc68626\") " pod="openstack/dnsmasq-dns-75c8ddd69c-mgsxh" Oct 06 10:16:38 crc kubenswrapper[4824]: I1006 10:16:38.186854 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/932b2879-a6ac-4179-946e-27d4acc68626-ovsdbserver-nb\") pod \"dnsmasq-dns-75c8ddd69c-mgsxh\" (UID: \"932b2879-a6ac-4179-946e-27d4acc68626\") " pod="openstack/dnsmasq-dns-75c8ddd69c-mgsxh" Oct 06 10:16:38 crc kubenswrapper[4824]: I1006 10:16:38.188007 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e844107-f8f2-48e8-9731-84eff0f8c752-combined-ca-bundle\") pod \"barbican-keystone-listener-67979769c4-7fg5z\" (UID: \"5e844107-f8f2-48e8-9731-84eff0f8c752\") " pod="openstack/barbican-keystone-listener-67979769c4-7fg5z" Oct 06 10:16:38 crc kubenswrapper[4824]: I1006 10:16:38.189701 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/932b2879-a6ac-4179-946e-27d4acc68626-dns-svc\") pod \"dnsmasq-dns-75c8ddd69c-mgsxh\" (UID: \"932b2879-a6ac-4179-946e-27d4acc68626\") " pod="openstack/dnsmasq-dns-75c8ddd69c-mgsxh" Oct 06 10:16:38 crc kubenswrapper[4824]: I1006 10:16:38.190577 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e844107-f8f2-48e8-9731-84eff0f8c752-config-data\") pod \"barbican-keystone-listener-67979769c4-7fg5z\" (UID: \"5e844107-f8f2-48e8-9731-84eff0f8c752\") " pod="openstack/barbican-keystone-listener-67979769c4-7fg5z" Oct 06 10:16:38 crc kubenswrapper[4824]: I1006 10:16:38.198110 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/932b2879-a6ac-4179-946e-27d4acc68626-dns-swift-storage-0\") pod \"dnsmasq-dns-75c8ddd69c-mgsxh\" (UID: \"932b2879-a6ac-4179-946e-27d4acc68626\") " pod="openstack/dnsmasq-dns-75c8ddd69c-mgsxh" Oct 06 10:16:38 crc kubenswrapper[4824]: I1006 10:16:38.199164 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5e844107-f8f2-48e8-9731-84eff0f8c752-config-data-custom\") pod \"barbican-keystone-listener-67979769c4-7fg5z\" (UID: \"5e844107-f8f2-48e8-9731-84eff0f8c752\") " pod="openstack/barbican-keystone-listener-67979769c4-7fg5z" Oct 06 10:16:38 crc kubenswrapper[4824]: I1006 10:16:38.208680 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rpqsz\" (UniqueName: \"kubernetes.io/projected/932b2879-a6ac-4179-946e-27d4acc68626-kube-api-access-rpqsz\") pod \"dnsmasq-dns-75c8ddd69c-mgsxh\" (UID: \"932b2879-a6ac-4179-946e-27d4acc68626\") " pod="openstack/dnsmasq-dns-75c8ddd69c-mgsxh" Oct 06 10:16:38 crc kubenswrapper[4824]: I1006 10:16:38.210146 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j4ghd\" (UniqueName: \"kubernetes.io/projected/5e844107-f8f2-48e8-9731-84eff0f8c752-kube-api-access-j4ghd\") pod \"barbican-keystone-listener-67979769c4-7fg5z\" (UID: \"5e844107-f8f2-48e8-9731-84eff0f8c752\") " pod="openstack/barbican-keystone-listener-67979769c4-7fg5z" Oct 06 10:16:38 crc kubenswrapper[4824]: I1006 10:16:38.245339 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-5f478db59d-8kntd"] Oct 06 10:16:38 crc kubenswrapper[4824]: I1006 10:16:38.247355 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5f478db59d-8kntd" Oct 06 10:16:38 crc kubenswrapper[4824]: I1006 10:16:38.252348 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Oct 06 10:16:38 crc kubenswrapper[4824]: I1006 10:16:38.261768 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-5f478db59d-8kntd"] Oct 06 10:16:38 crc kubenswrapper[4824]: I1006 10:16:38.379136 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vgb9v\" (UniqueName: \"kubernetes.io/projected/1e792caf-ac02-4e6c-9841-ed58133f1918-kube-api-access-vgb9v\") pod \"barbican-api-5f478db59d-8kntd\" (UID: \"1e792caf-ac02-4e6c-9841-ed58133f1918\") " pod="openstack/barbican-api-5f478db59d-8kntd" Oct 06 10:16:38 crc kubenswrapper[4824]: I1006 10:16:38.379207 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1e792caf-ac02-4e6c-9841-ed58133f1918-config-data-custom\") pod \"barbican-api-5f478db59d-8kntd\" (UID: \"1e792caf-ac02-4e6c-9841-ed58133f1918\") " pod="openstack/barbican-api-5f478db59d-8kntd" Oct 06 10:16:38 crc kubenswrapper[4824]: I1006 10:16:38.379261 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e792caf-ac02-4e6c-9841-ed58133f1918-combined-ca-bundle\") pod \"barbican-api-5f478db59d-8kntd\" (UID: \"1e792caf-ac02-4e6c-9841-ed58133f1918\") " pod="openstack/barbican-api-5f478db59d-8kntd" Oct 06 10:16:38 crc kubenswrapper[4824]: I1006 10:16:38.379336 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1e792caf-ac02-4e6c-9841-ed58133f1918-logs\") pod \"barbican-api-5f478db59d-8kntd\" (UID: \"1e792caf-ac02-4e6c-9841-ed58133f1918\") " pod="openstack/barbican-api-5f478db59d-8kntd" Oct 06 10:16:38 crc kubenswrapper[4824]: I1006 10:16:38.379364 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e792caf-ac02-4e6c-9841-ed58133f1918-config-data\") pod \"barbican-api-5f478db59d-8kntd\" (UID: \"1e792caf-ac02-4e6c-9841-ed58133f1918\") " pod="openstack/barbican-api-5f478db59d-8kntd" Oct 06 10:16:38 crc kubenswrapper[4824]: I1006 10:16:38.408651 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-659dd56d6c-qpxtd" Oct 06 10:16:38 crc kubenswrapper[4824]: I1006 10:16:38.440773 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-67979769c4-7fg5z" Oct 06 10:16:38 crc kubenswrapper[4824]: I1006 10:16:38.455646 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75c8ddd69c-mgsxh" Oct 06 10:16:38 crc kubenswrapper[4824]: I1006 10:16:38.481733 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vgb9v\" (UniqueName: \"kubernetes.io/projected/1e792caf-ac02-4e6c-9841-ed58133f1918-kube-api-access-vgb9v\") pod \"barbican-api-5f478db59d-8kntd\" (UID: \"1e792caf-ac02-4e6c-9841-ed58133f1918\") " pod="openstack/barbican-api-5f478db59d-8kntd" Oct 06 10:16:38 crc kubenswrapper[4824]: I1006 10:16:38.481833 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1e792caf-ac02-4e6c-9841-ed58133f1918-config-data-custom\") pod \"barbican-api-5f478db59d-8kntd\" (UID: \"1e792caf-ac02-4e6c-9841-ed58133f1918\") " pod="openstack/barbican-api-5f478db59d-8kntd" Oct 06 10:16:38 crc kubenswrapper[4824]: I1006 10:16:38.481886 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e792caf-ac02-4e6c-9841-ed58133f1918-combined-ca-bundle\") pod \"barbican-api-5f478db59d-8kntd\" (UID: \"1e792caf-ac02-4e6c-9841-ed58133f1918\") " pod="openstack/barbican-api-5f478db59d-8kntd" Oct 06 10:16:38 crc kubenswrapper[4824]: I1006 10:16:38.481917 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1e792caf-ac02-4e6c-9841-ed58133f1918-logs\") pod \"barbican-api-5f478db59d-8kntd\" (UID: \"1e792caf-ac02-4e6c-9841-ed58133f1918\") " pod="openstack/barbican-api-5f478db59d-8kntd" Oct 06 10:16:38 crc kubenswrapper[4824]: I1006 10:16:38.481941 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e792caf-ac02-4e6c-9841-ed58133f1918-config-data\") pod \"barbican-api-5f478db59d-8kntd\" (UID: \"1e792caf-ac02-4e6c-9841-ed58133f1918\") " pod="openstack/barbican-api-5f478db59d-8kntd" Oct 06 10:16:38 crc kubenswrapper[4824]: I1006 10:16:38.483010 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1e792caf-ac02-4e6c-9841-ed58133f1918-logs\") pod \"barbican-api-5f478db59d-8kntd\" (UID: \"1e792caf-ac02-4e6c-9841-ed58133f1918\") " pod="openstack/barbican-api-5f478db59d-8kntd" Oct 06 10:16:38 crc kubenswrapper[4824]: I1006 10:16:38.495998 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e792caf-ac02-4e6c-9841-ed58133f1918-combined-ca-bundle\") pod \"barbican-api-5f478db59d-8kntd\" (UID: \"1e792caf-ac02-4e6c-9841-ed58133f1918\") " pod="openstack/barbican-api-5f478db59d-8kntd" Oct 06 10:16:38 crc kubenswrapper[4824]: I1006 10:16:38.508559 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e792caf-ac02-4e6c-9841-ed58133f1918-config-data\") pod \"barbican-api-5f478db59d-8kntd\" (UID: \"1e792caf-ac02-4e6c-9841-ed58133f1918\") " pod="openstack/barbican-api-5f478db59d-8kntd" Oct 06 10:16:38 crc kubenswrapper[4824]: I1006 10:16:38.514341 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1e792caf-ac02-4e6c-9841-ed58133f1918-config-data-custom\") pod \"barbican-api-5f478db59d-8kntd\" (UID: \"1e792caf-ac02-4e6c-9841-ed58133f1918\") " pod="openstack/barbican-api-5f478db59d-8kntd" Oct 06 10:16:38 crc kubenswrapper[4824]: I1006 10:16:38.518308 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vgb9v\" (UniqueName: \"kubernetes.io/projected/1e792caf-ac02-4e6c-9841-ed58133f1918-kube-api-access-vgb9v\") pod \"barbican-api-5f478db59d-8kntd\" (UID: \"1e792caf-ac02-4e6c-9841-ed58133f1918\") " pod="openstack/barbican-api-5f478db59d-8kntd" Oct 06 10:16:38 crc kubenswrapper[4824]: I1006 10:16:38.568843 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5f478db59d-8kntd" Oct 06 10:16:40 crc kubenswrapper[4824]: E1006 10:16:40.222253 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ceilometer-0" podUID="13781991-323c-49ae-9c5d-625e2ffd019b" Oct 06 10:16:40 crc kubenswrapper[4824]: I1006 10:16:40.455693 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-659dd56d6c-qpxtd"] Oct 06 10:16:40 crc kubenswrapper[4824]: I1006 10:16:40.480728 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-5f478db59d-8kntd"] Oct 06 10:16:40 crc kubenswrapper[4824]: I1006 10:16:40.493161 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-mgsxh"] Oct 06 10:16:40 crc kubenswrapper[4824]: I1006 10:16:40.586512 4824 generic.go:334] "Generic (PLEG): container finished" podID="2382b80c-75fc-435f-be5f-0f1eb6a5cd17" containerID="bfe65796f2dadbf11294a2fca859967193ea63bf8d8080fbb171c2311304906e" exitCode=0 Oct 06 10:16:40 crc kubenswrapper[4824]: I1006 10:16:40.586731 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-ldn9v" event={"ID":"2382b80c-75fc-435f-be5f-0f1eb6a5cd17","Type":"ContainerDied","Data":"bfe65796f2dadbf11294a2fca859967193ea63bf8d8080fbb171c2311304906e"} Oct 06 10:16:40 crc kubenswrapper[4824]: I1006 10:16:40.589736 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75c8ddd69c-mgsxh" event={"ID":"932b2879-a6ac-4179-946e-27d4acc68626","Type":"ContainerStarted","Data":"c3154eb6ae50eceddaa43515eba04c71c4a3fcf033cd2b00d8567d37a8b43034"} Oct 06 10:16:40 crc kubenswrapper[4824]: I1006 10:16:40.591133 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5f478db59d-8kntd" event={"ID":"1e792caf-ac02-4e6c-9841-ed58133f1918","Type":"ContainerStarted","Data":"48a22e6159cff9549064e391a6121b0e4b88133f150e41f49a193cc664a59156"} Oct 06 10:16:40 crc kubenswrapper[4824]: I1006 10:16:40.593319 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-659dd56d6c-qpxtd" event={"ID":"fb46f73f-9d47-4189-990a-051fed587851","Type":"ContainerStarted","Data":"698f0dcffa5e1432a06aca88a9510f7bc52c3ea0933f22acf395be6b59016bef"} Oct 06 10:16:40 crc kubenswrapper[4824]: I1006 10:16:40.596261 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-67979769c4-7fg5z"] Oct 06 10:16:40 crc kubenswrapper[4824]: I1006 10:16:40.599407 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"13781991-323c-49ae-9c5d-625e2ffd019b","Type":"ContainerStarted","Data":"e5bda72f4bbab5d1f8e75631b7c6e3d944ce6bce9021bc30b8827ebfa9880e75"} Oct 06 10:16:40 crc kubenswrapper[4824]: I1006 10:16:40.599784 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="13781991-323c-49ae-9c5d-625e2ffd019b" containerName="ceilometer-notification-agent" containerID="cri-o://c75d8022f4b4f06e5bd8a64705b9bfbcd9ee5e6a7edb3e34e026c46d0f72d761" gracePeriod=30 Oct 06 10:16:40 crc kubenswrapper[4824]: I1006 10:16:40.599831 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 06 10:16:40 crc kubenswrapper[4824]: I1006 10:16:40.599887 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="13781991-323c-49ae-9c5d-625e2ffd019b" containerName="sg-core" containerID="cri-o://68acaf0ce4e83b374f4f0fabe242c7a9dd535c1809d4f2f3f0cd8f1a64d60fe5" gracePeriod=30 Oct 06 10:16:40 crc kubenswrapper[4824]: I1006 10:16:40.599794 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="13781991-323c-49ae-9c5d-625e2ffd019b" containerName="proxy-httpd" containerID="cri-o://e5bda72f4bbab5d1f8e75631b7c6e3d944ce6bce9021bc30b8827ebfa9880e75" gracePeriod=30 Oct 06 10:16:40 crc kubenswrapper[4824]: W1006 10:16:40.607205 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5e844107_f8f2_48e8_9731_84eff0f8c752.slice/crio-d4f3103d76d0434861b41b4c1dc08ef465e236466a5466ea1249aa2ac1540a8d WatchSource:0}: Error finding container d4f3103d76d0434861b41b4c1dc08ef465e236466a5466ea1249aa2ac1540a8d: Status 404 returned error can't find the container with id d4f3103d76d0434861b41b4c1dc08ef465e236466a5466ea1249aa2ac1540a8d Oct 06 10:16:41 crc kubenswrapper[4824]: I1006 10:16:41.083365 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-78c986c6fb-bzcht"] Oct 06 10:16:41 crc kubenswrapper[4824]: I1006 10:16:41.085484 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-78c986c6fb-bzcht" Oct 06 10:16:41 crc kubenswrapper[4824]: I1006 10:16:41.087149 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Oct 06 10:16:41 crc kubenswrapper[4824]: I1006 10:16:41.090487 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Oct 06 10:16:41 crc kubenswrapper[4824]: I1006 10:16:41.095595 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-78c986c6fb-bzcht"] Oct 06 10:16:41 crc kubenswrapper[4824]: I1006 10:16:41.141097 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b8a8eafe-1a7a-4eca-b61c-2d744512a54b-config-data\") pod \"barbican-api-78c986c6fb-bzcht\" (UID: \"b8a8eafe-1a7a-4eca-b61c-2d744512a54b\") " pod="openstack/barbican-api-78c986c6fb-bzcht" Oct 06 10:16:41 crc kubenswrapper[4824]: I1006 10:16:41.141152 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b8a8eafe-1a7a-4eca-b61c-2d744512a54b-config-data-custom\") pod \"barbican-api-78c986c6fb-bzcht\" (UID: \"b8a8eafe-1a7a-4eca-b61c-2d744512a54b\") " pod="openstack/barbican-api-78c986c6fb-bzcht" Oct 06 10:16:41 crc kubenswrapper[4824]: I1006 10:16:41.141187 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b8a8eafe-1a7a-4eca-b61c-2d744512a54b-logs\") pod \"barbican-api-78c986c6fb-bzcht\" (UID: \"b8a8eafe-1a7a-4eca-b61c-2d744512a54b\") " pod="openstack/barbican-api-78c986c6fb-bzcht" Oct 06 10:16:41 crc kubenswrapper[4824]: I1006 10:16:41.141242 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8a8eafe-1a7a-4eca-b61c-2d744512a54b-combined-ca-bundle\") pod \"barbican-api-78c986c6fb-bzcht\" (UID: \"b8a8eafe-1a7a-4eca-b61c-2d744512a54b\") " pod="openstack/barbican-api-78c986c6fb-bzcht" Oct 06 10:16:41 crc kubenswrapper[4824]: I1006 10:16:41.141276 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pjc8h\" (UniqueName: \"kubernetes.io/projected/b8a8eafe-1a7a-4eca-b61c-2d744512a54b-kube-api-access-pjc8h\") pod \"barbican-api-78c986c6fb-bzcht\" (UID: \"b8a8eafe-1a7a-4eca-b61c-2d744512a54b\") " pod="openstack/barbican-api-78c986c6fb-bzcht" Oct 06 10:16:41 crc kubenswrapper[4824]: I1006 10:16:41.141301 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b8a8eafe-1a7a-4eca-b61c-2d744512a54b-public-tls-certs\") pod \"barbican-api-78c986c6fb-bzcht\" (UID: \"b8a8eafe-1a7a-4eca-b61c-2d744512a54b\") " pod="openstack/barbican-api-78c986c6fb-bzcht" Oct 06 10:16:41 crc kubenswrapper[4824]: I1006 10:16:41.141671 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b8a8eafe-1a7a-4eca-b61c-2d744512a54b-internal-tls-certs\") pod \"barbican-api-78c986c6fb-bzcht\" (UID: \"b8a8eafe-1a7a-4eca-b61c-2d744512a54b\") " pod="openstack/barbican-api-78c986c6fb-bzcht" Oct 06 10:16:41 crc kubenswrapper[4824]: I1006 10:16:41.243101 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pjc8h\" (UniqueName: \"kubernetes.io/projected/b8a8eafe-1a7a-4eca-b61c-2d744512a54b-kube-api-access-pjc8h\") pod \"barbican-api-78c986c6fb-bzcht\" (UID: \"b8a8eafe-1a7a-4eca-b61c-2d744512a54b\") " pod="openstack/barbican-api-78c986c6fb-bzcht" Oct 06 10:16:41 crc kubenswrapper[4824]: I1006 10:16:41.243183 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b8a8eafe-1a7a-4eca-b61c-2d744512a54b-public-tls-certs\") pod \"barbican-api-78c986c6fb-bzcht\" (UID: \"b8a8eafe-1a7a-4eca-b61c-2d744512a54b\") " pod="openstack/barbican-api-78c986c6fb-bzcht" Oct 06 10:16:41 crc kubenswrapper[4824]: I1006 10:16:41.243285 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b8a8eafe-1a7a-4eca-b61c-2d744512a54b-internal-tls-certs\") pod \"barbican-api-78c986c6fb-bzcht\" (UID: \"b8a8eafe-1a7a-4eca-b61c-2d744512a54b\") " pod="openstack/barbican-api-78c986c6fb-bzcht" Oct 06 10:16:41 crc kubenswrapper[4824]: I1006 10:16:41.243366 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b8a8eafe-1a7a-4eca-b61c-2d744512a54b-config-data\") pod \"barbican-api-78c986c6fb-bzcht\" (UID: \"b8a8eafe-1a7a-4eca-b61c-2d744512a54b\") " pod="openstack/barbican-api-78c986c6fb-bzcht" Oct 06 10:16:41 crc kubenswrapper[4824]: I1006 10:16:41.243396 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b8a8eafe-1a7a-4eca-b61c-2d744512a54b-config-data-custom\") pod \"barbican-api-78c986c6fb-bzcht\" (UID: \"b8a8eafe-1a7a-4eca-b61c-2d744512a54b\") " pod="openstack/barbican-api-78c986c6fb-bzcht" Oct 06 10:16:41 crc kubenswrapper[4824]: I1006 10:16:41.243442 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b8a8eafe-1a7a-4eca-b61c-2d744512a54b-logs\") pod \"barbican-api-78c986c6fb-bzcht\" (UID: \"b8a8eafe-1a7a-4eca-b61c-2d744512a54b\") " pod="openstack/barbican-api-78c986c6fb-bzcht" Oct 06 10:16:41 crc kubenswrapper[4824]: I1006 10:16:41.243476 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8a8eafe-1a7a-4eca-b61c-2d744512a54b-combined-ca-bundle\") pod \"barbican-api-78c986c6fb-bzcht\" (UID: \"b8a8eafe-1a7a-4eca-b61c-2d744512a54b\") " pod="openstack/barbican-api-78c986c6fb-bzcht" Oct 06 10:16:41 crc kubenswrapper[4824]: I1006 10:16:41.244710 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b8a8eafe-1a7a-4eca-b61c-2d744512a54b-logs\") pod \"barbican-api-78c986c6fb-bzcht\" (UID: \"b8a8eafe-1a7a-4eca-b61c-2d744512a54b\") " pod="openstack/barbican-api-78c986c6fb-bzcht" Oct 06 10:16:41 crc kubenswrapper[4824]: I1006 10:16:41.248512 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b8a8eafe-1a7a-4eca-b61c-2d744512a54b-public-tls-certs\") pod \"barbican-api-78c986c6fb-bzcht\" (UID: \"b8a8eafe-1a7a-4eca-b61c-2d744512a54b\") " pod="openstack/barbican-api-78c986c6fb-bzcht" Oct 06 10:16:41 crc kubenswrapper[4824]: I1006 10:16:41.254273 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b8a8eafe-1a7a-4eca-b61c-2d744512a54b-config-data\") pod \"barbican-api-78c986c6fb-bzcht\" (UID: \"b8a8eafe-1a7a-4eca-b61c-2d744512a54b\") " pod="openstack/barbican-api-78c986c6fb-bzcht" Oct 06 10:16:41 crc kubenswrapper[4824]: I1006 10:16:41.260910 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8a8eafe-1a7a-4eca-b61c-2d744512a54b-combined-ca-bundle\") pod \"barbican-api-78c986c6fb-bzcht\" (UID: \"b8a8eafe-1a7a-4eca-b61c-2d744512a54b\") " pod="openstack/barbican-api-78c986c6fb-bzcht" Oct 06 10:16:41 crc kubenswrapper[4824]: I1006 10:16:41.268216 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b8a8eafe-1a7a-4eca-b61c-2d744512a54b-config-data-custom\") pod \"barbican-api-78c986c6fb-bzcht\" (UID: \"b8a8eafe-1a7a-4eca-b61c-2d744512a54b\") " pod="openstack/barbican-api-78c986c6fb-bzcht" Oct 06 10:16:41 crc kubenswrapper[4824]: I1006 10:16:41.268245 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b8a8eafe-1a7a-4eca-b61c-2d744512a54b-internal-tls-certs\") pod \"barbican-api-78c986c6fb-bzcht\" (UID: \"b8a8eafe-1a7a-4eca-b61c-2d744512a54b\") " pod="openstack/barbican-api-78c986c6fb-bzcht" Oct 06 10:16:41 crc kubenswrapper[4824]: I1006 10:16:41.272289 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pjc8h\" (UniqueName: \"kubernetes.io/projected/b8a8eafe-1a7a-4eca-b61c-2d744512a54b-kube-api-access-pjc8h\") pod \"barbican-api-78c986c6fb-bzcht\" (UID: \"b8a8eafe-1a7a-4eca-b61c-2d744512a54b\") " pod="openstack/barbican-api-78c986c6fb-bzcht" Oct 06 10:16:41 crc kubenswrapper[4824]: I1006 10:16:41.416564 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-78c986c6fb-bzcht" Oct 06 10:16:41 crc kubenswrapper[4824]: I1006 10:16:41.647684 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5f478db59d-8kntd" event={"ID":"1e792caf-ac02-4e6c-9841-ed58133f1918","Type":"ContainerStarted","Data":"717654dbd581395068e7062c4a749a857288b21cb683c2bb83c1b3355e2232ca"} Oct 06 10:16:41 crc kubenswrapper[4824]: I1006 10:16:41.648137 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5f478db59d-8kntd" event={"ID":"1e792caf-ac02-4e6c-9841-ed58133f1918","Type":"ContainerStarted","Data":"38b1e0c9677bb96ef4d1872779ef930ca0e79ec2b31786d8860e86c3613db462"} Oct 06 10:16:41 crc kubenswrapper[4824]: I1006 10:16:41.648163 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-5f478db59d-8kntd" Oct 06 10:16:41 crc kubenswrapper[4824]: I1006 10:16:41.648178 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-5f478db59d-8kntd" Oct 06 10:16:41 crc kubenswrapper[4824]: I1006 10:16:41.653418 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-67979769c4-7fg5z" event={"ID":"5e844107-f8f2-48e8-9731-84eff0f8c752","Type":"ContainerStarted","Data":"d4f3103d76d0434861b41b4c1dc08ef465e236466a5466ea1249aa2ac1540a8d"} Oct 06 10:16:41 crc kubenswrapper[4824]: I1006 10:16:41.679251 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-5f478db59d-8kntd" podStartSLOduration=3.679217258 podStartE2EDuration="3.679217258s" podCreationTimestamp="2025-10-06 10:16:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:16:41.673050145 +0000 UTC m=+1111.037473026" watchObservedRunningTime="2025-10-06 10:16:41.679217258 +0000 UTC m=+1111.043640119" Oct 06 10:16:41 crc kubenswrapper[4824]: I1006 10:16:41.705126 4824 generic.go:334] "Generic (PLEG): container finished" podID="13781991-323c-49ae-9c5d-625e2ffd019b" containerID="e5bda72f4bbab5d1f8e75631b7c6e3d944ce6bce9021bc30b8827ebfa9880e75" exitCode=0 Oct 06 10:16:41 crc kubenswrapper[4824]: I1006 10:16:41.705159 4824 generic.go:334] "Generic (PLEG): container finished" podID="13781991-323c-49ae-9c5d-625e2ffd019b" containerID="68acaf0ce4e83b374f4f0fabe242c7a9dd535c1809d4f2f3f0cd8f1a64d60fe5" exitCode=2 Oct 06 10:16:41 crc kubenswrapper[4824]: I1006 10:16:41.705213 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"13781991-323c-49ae-9c5d-625e2ffd019b","Type":"ContainerDied","Data":"e5bda72f4bbab5d1f8e75631b7c6e3d944ce6bce9021bc30b8827ebfa9880e75"} Oct 06 10:16:41 crc kubenswrapper[4824]: I1006 10:16:41.705241 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"13781991-323c-49ae-9c5d-625e2ffd019b","Type":"ContainerDied","Data":"68acaf0ce4e83b374f4f0fabe242c7a9dd535c1809d4f2f3f0cd8f1a64d60fe5"} Oct 06 10:16:41 crc kubenswrapper[4824]: I1006 10:16:41.729227 4824 generic.go:334] "Generic (PLEG): container finished" podID="932b2879-a6ac-4179-946e-27d4acc68626" containerID="f64a3cb7ead4a6db4fbf94a4a22a831eb389091f67782891d71e95b15dcd1f56" exitCode=0 Oct 06 10:16:41 crc kubenswrapper[4824]: I1006 10:16:41.730348 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75c8ddd69c-mgsxh" event={"ID":"932b2879-a6ac-4179-946e-27d4acc68626","Type":"ContainerDied","Data":"f64a3cb7ead4a6db4fbf94a4a22a831eb389091f67782891d71e95b15dcd1f56"} Oct 06 10:16:41 crc kubenswrapper[4824]: I1006 10:16:41.998219 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-78c986c6fb-bzcht"] Oct 06 10:16:42 crc kubenswrapper[4824]: I1006 10:16:42.426860 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-ldn9v" Oct 06 10:16:42 crc kubenswrapper[4824]: I1006 10:16:42.488900 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqk7s\" (UniqueName: \"kubernetes.io/projected/2382b80c-75fc-435f-be5f-0f1eb6a5cd17-kube-api-access-fqk7s\") pod \"2382b80c-75fc-435f-be5f-0f1eb6a5cd17\" (UID: \"2382b80c-75fc-435f-be5f-0f1eb6a5cd17\") " Oct 06 10:16:42 crc kubenswrapper[4824]: I1006 10:16:42.489051 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2382b80c-75fc-435f-be5f-0f1eb6a5cd17-config-data\") pod \"2382b80c-75fc-435f-be5f-0f1eb6a5cd17\" (UID: \"2382b80c-75fc-435f-be5f-0f1eb6a5cd17\") " Oct 06 10:16:42 crc kubenswrapper[4824]: I1006 10:16:42.489134 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2382b80c-75fc-435f-be5f-0f1eb6a5cd17-combined-ca-bundle\") pod \"2382b80c-75fc-435f-be5f-0f1eb6a5cd17\" (UID: \"2382b80c-75fc-435f-be5f-0f1eb6a5cd17\") " Oct 06 10:16:42 crc kubenswrapper[4824]: I1006 10:16:42.489210 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2382b80c-75fc-435f-be5f-0f1eb6a5cd17-scripts\") pod \"2382b80c-75fc-435f-be5f-0f1eb6a5cd17\" (UID: \"2382b80c-75fc-435f-be5f-0f1eb6a5cd17\") " Oct 06 10:16:42 crc kubenswrapper[4824]: I1006 10:16:42.489264 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/2382b80c-75fc-435f-be5f-0f1eb6a5cd17-db-sync-config-data\") pod \"2382b80c-75fc-435f-be5f-0f1eb6a5cd17\" (UID: \"2382b80c-75fc-435f-be5f-0f1eb6a5cd17\") " Oct 06 10:16:42 crc kubenswrapper[4824]: I1006 10:16:42.489299 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2382b80c-75fc-435f-be5f-0f1eb6a5cd17-etc-machine-id\") pod \"2382b80c-75fc-435f-be5f-0f1eb6a5cd17\" (UID: \"2382b80c-75fc-435f-be5f-0f1eb6a5cd17\") " Oct 06 10:16:42 crc kubenswrapper[4824]: I1006 10:16:42.490289 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2382b80c-75fc-435f-be5f-0f1eb6a5cd17-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "2382b80c-75fc-435f-be5f-0f1eb6a5cd17" (UID: "2382b80c-75fc-435f-be5f-0f1eb6a5cd17"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 06 10:16:42 crc kubenswrapper[4824]: I1006 10:16:42.508418 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2382b80c-75fc-435f-be5f-0f1eb6a5cd17-scripts" (OuterVolumeSpecName: "scripts") pod "2382b80c-75fc-435f-be5f-0f1eb6a5cd17" (UID: "2382b80c-75fc-435f-be5f-0f1eb6a5cd17"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:16:42 crc kubenswrapper[4824]: I1006 10:16:42.511874 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2382b80c-75fc-435f-be5f-0f1eb6a5cd17-kube-api-access-fqk7s" (OuterVolumeSpecName: "kube-api-access-fqk7s") pod "2382b80c-75fc-435f-be5f-0f1eb6a5cd17" (UID: "2382b80c-75fc-435f-be5f-0f1eb6a5cd17"). InnerVolumeSpecName "kube-api-access-fqk7s". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:16:42 crc kubenswrapper[4824]: I1006 10:16:42.523311 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2382b80c-75fc-435f-be5f-0f1eb6a5cd17-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "2382b80c-75fc-435f-be5f-0f1eb6a5cd17" (UID: "2382b80c-75fc-435f-be5f-0f1eb6a5cd17"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:16:42 crc kubenswrapper[4824]: I1006 10:16:42.542717 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2382b80c-75fc-435f-be5f-0f1eb6a5cd17-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2382b80c-75fc-435f-be5f-0f1eb6a5cd17" (UID: "2382b80c-75fc-435f-be5f-0f1eb6a5cd17"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:16:42 crc kubenswrapper[4824]: I1006 10:16:42.591432 4824 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/2382b80c-75fc-435f-be5f-0f1eb6a5cd17-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:42 crc kubenswrapper[4824]: I1006 10:16:42.591840 4824 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2382b80c-75fc-435f-be5f-0f1eb6a5cd17-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:42 crc kubenswrapper[4824]: I1006 10:16:42.591855 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqk7s\" (UniqueName: \"kubernetes.io/projected/2382b80c-75fc-435f-be5f-0f1eb6a5cd17-kube-api-access-fqk7s\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:42 crc kubenswrapper[4824]: I1006 10:16:42.591871 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2382b80c-75fc-435f-be5f-0f1eb6a5cd17-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:42 crc kubenswrapper[4824]: I1006 10:16:42.591882 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2382b80c-75fc-435f-be5f-0f1eb6a5cd17-scripts\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:42 crc kubenswrapper[4824]: I1006 10:16:42.593090 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2382b80c-75fc-435f-be5f-0f1eb6a5cd17-config-data" (OuterVolumeSpecName: "config-data") pod "2382b80c-75fc-435f-be5f-0f1eb6a5cd17" (UID: "2382b80c-75fc-435f-be5f-0f1eb6a5cd17"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:16:42 crc kubenswrapper[4824]: I1006 10:16:42.693410 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2382b80c-75fc-435f-be5f-0f1eb6a5cd17-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:42 crc kubenswrapper[4824]: I1006 10:16:42.741146 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-ldn9v" event={"ID":"2382b80c-75fc-435f-be5f-0f1eb6a5cd17","Type":"ContainerDied","Data":"dac702736fb8c27487a0f0433392419f3af8a038eb66fe6f9aa60c761cd56c89"} Oct 06 10:16:42 crc kubenswrapper[4824]: I1006 10:16:42.741187 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dac702736fb8c27487a0f0433392419f3af8a038eb66fe6f9aa60c761cd56c89" Oct 06 10:16:42 crc kubenswrapper[4824]: I1006 10:16:42.741209 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-ldn9v" Oct 06 10:16:42 crc kubenswrapper[4824]: I1006 10:16:42.952402 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Oct 06 10:16:42 crc kubenswrapper[4824]: E1006 10:16:42.953542 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2382b80c-75fc-435f-be5f-0f1eb6a5cd17" containerName="cinder-db-sync" Oct 06 10:16:42 crc kubenswrapper[4824]: I1006 10:16:42.953569 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="2382b80c-75fc-435f-be5f-0f1eb6a5cd17" containerName="cinder-db-sync" Oct 06 10:16:42 crc kubenswrapper[4824]: I1006 10:16:42.953833 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="2382b80c-75fc-435f-be5f-0f1eb6a5cd17" containerName="cinder-db-sync" Oct 06 10:16:42 crc kubenswrapper[4824]: I1006 10:16:42.956010 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 06 10:16:42 crc kubenswrapper[4824]: I1006 10:16:42.962242 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 06 10:16:42 crc kubenswrapper[4824]: I1006 10:16:42.964368 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Oct 06 10:16:42 crc kubenswrapper[4824]: I1006 10:16:42.964652 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-v75vf" Oct 06 10:16:42 crc kubenswrapper[4824]: I1006 10:16:42.964776 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Oct 06 10:16:42 crc kubenswrapper[4824]: I1006 10:16:42.967823 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Oct 06 10:16:43 crc kubenswrapper[4824]: I1006 10:16:43.108573 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-mgsxh"] Oct 06 10:16:43 crc kubenswrapper[4824]: I1006 10:16:43.127224 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a56ad79-83f8-4fe1-be09-fa2be02d7644-config-data\") pod \"cinder-scheduler-0\" (UID: \"4a56ad79-83f8-4fe1-be09-fa2be02d7644\") " pod="openstack/cinder-scheduler-0" Oct 06 10:16:43 crc kubenswrapper[4824]: I1006 10:16:43.127289 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4a56ad79-83f8-4fe1-be09-fa2be02d7644-scripts\") pod \"cinder-scheduler-0\" (UID: \"4a56ad79-83f8-4fe1-be09-fa2be02d7644\") " pod="openstack/cinder-scheduler-0" Oct 06 10:16:43 crc kubenswrapper[4824]: I1006 10:16:43.127321 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a56ad79-83f8-4fe1-be09-fa2be02d7644-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"4a56ad79-83f8-4fe1-be09-fa2be02d7644\") " pod="openstack/cinder-scheduler-0" Oct 06 10:16:43 crc kubenswrapper[4824]: I1006 10:16:43.127346 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j49q4\" (UniqueName: \"kubernetes.io/projected/4a56ad79-83f8-4fe1-be09-fa2be02d7644-kube-api-access-j49q4\") pod \"cinder-scheduler-0\" (UID: \"4a56ad79-83f8-4fe1-be09-fa2be02d7644\") " pod="openstack/cinder-scheduler-0" Oct 06 10:16:43 crc kubenswrapper[4824]: I1006 10:16:43.127485 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4a56ad79-83f8-4fe1-be09-fa2be02d7644-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"4a56ad79-83f8-4fe1-be09-fa2be02d7644\") " pod="openstack/cinder-scheduler-0" Oct 06 10:16:43 crc kubenswrapper[4824]: I1006 10:16:43.127551 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4a56ad79-83f8-4fe1-be09-fa2be02d7644-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"4a56ad79-83f8-4fe1-be09-fa2be02d7644\") " pod="openstack/cinder-scheduler-0" Oct 06 10:16:43 crc kubenswrapper[4824]: I1006 10:16:43.153234 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-k2mqf"] Oct 06 10:16:43 crc kubenswrapper[4824]: I1006 10:16:43.155623 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5784cf869f-k2mqf" Oct 06 10:16:43 crc kubenswrapper[4824]: I1006 10:16:43.165168 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-k2mqf"] Oct 06 10:16:43 crc kubenswrapper[4824]: I1006 10:16:43.230420 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8de9ec83-5bc3-47a8-9a63-84acb7d6b774-dns-swift-storage-0\") pod \"dnsmasq-dns-5784cf869f-k2mqf\" (UID: \"8de9ec83-5bc3-47a8-9a63-84acb7d6b774\") " pod="openstack/dnsmasq-dns-5784cf869f-k2mqf" Oct 06 10:16:43 crc kubenswrapper[4824]: I1006 10:16:43.230478 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wbdxh\" (UniqueName: \"kubernetes.io/projected/8de9ec83-5bc3-47a8-9a63-84acb7d6b774-kube-api-access-wbdxh\") pod \"dnsmasq-dns-5784cf869f-k2mqf\" (UID: \"8de9ec83-5bc3-47a8-9a63-84acb7d6b774\") " pod="openstack/dnsmasq-dns-5784cf869f-k2mqf" Oct 06 10:16:43 crc kubenswrapper[4824]: I1006 10:16:43.230553 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a56ad79-83f8-4fe1-be09-fa2be02d7644-config-data\") pod \"cinder-scheduler-0\" (UID: \"4a56ad79-83f8-4fe1-be09-fa2be02d7644\") " pod="openstack/cinder-scheduler-0" Oct 06 10:16:43 crc kubenswrapper[4824]: I1006 10:16:43.230581 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8de9ec83-5bc3-47a8-9a63-84acb7d6b774-dns-svc\") pod \"dnsmasq-dns-5784cf869f-k2mqf\" (UID: \"8de9ec83-5bc3-47a8-9a63-84acb7d6b774\") " pod="openstack/dnsmasq-dns-5784cf869f-k2mqf" Oct 06 10:16:43 crc kubenswrapper[4824]: I1006 10:16:43.230609 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4a56ad79-83f8-4fe1-be09-fa2be02d7644-scripts\") pod \"cinder-scheduler-0\" (UID: \"4a56ad79-83f8-4fe1-be09-fa2be02d7644\") " pod="openstack/cinder-scheduler-0" Oct 06 10:16:43 crc kubenswrapper[4824]: I1006 10:16:43.230632 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8de9ec83-5bc3-47a8-9a63-84acb7d6b774-ovsdbserver-nb\") pod \"dnsmasq-dns-5784cf869f-k2mqf\" (UID: \"8de9ec83-5bc3-47a8-9a63-84acb7d6b774\") " pod="openstack/dnsmasq-dns-5784cf869f-k2mqf" Oct 06 10:16:43 crc kubenswrapper[4824]: I1006 10:16:43.230653 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8de9ec83-5bc3-47a8-9a63-84acb7d6b774-config\") pod \"dnsmasq-dns-5784cf869f-k2mqf\" (UID: \"8de9ec83-5bc3-47a8-9a63-84acb7d6b774\") " pod="openstack/dnsmasq-dns-5784cf869f-k2mqf" Oct 06 10:16:43 crc kubenswrapper[4824]: I1006 10:16:43.230672 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a56ad79-83f8-4fe1-be09-fa2be02d7644-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"4a56ad79-83f8-4fe1-be09-fa2be02d7644\") " pod="openstack/cinder-scheduler-0" Oct 06 10:16:43 crc kubenswrapper[4824]: I1006 10:16:43.230697 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j49q4\" (UniqueName: \"kubernetes.io/projected/4a56ad79-83f8-4fe1-be09-fa2be02d7644-kube-api-access-j49q4\") pod \"cinder-scheduler-0\" (UID: \"4a56ad79-83f8-4fe1-be09-fa2be02d7644\") " pod="openstack/cinder-scheduler-0" Oct 06 10:16:43 crc kubenswrapper[4824]: I1006 10:16:43.230721 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4a56ad79-83f8-4fe1-be09-fa2be02d7644-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"4a56ad79-83f8-4fe1-be09-fa2be02d7644\") " pod="openstack/cinder-scheduler-0" Oct 06 10:16:43 crc kubenswrapper[4824]: I1006 10:16:43.230769 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8de9ec83-5bc3-47a8-9a63-84acb7d6b774-ovsdbserver-sb\") pod \"dnsmasq-dns-5784cf869f-k2mqf\" (UID: \"8de9ec83-5bc3-47a8-9a63-84acb7d6b774\") " pod="openstack/dnsmasq-dns-5784cf869f-k2mqf" Oct 06 10:16:43 crc kubenswrapper[4824]: I1006 10:16:43.230789 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4a56ad79-83f8-4fe1-be09-fa2be02d7644-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"4a56ad79-83f8-4fe1-be09-fa2be02d7644\") " pod="openstack/cinder-scheduler-0" Oct 06 10:16:43 crc kubenswrapper[4824]: I1006 10:16:43.236118 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4a56ad79-83f8-4fe1-be09-fa2be02d7644-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"4a56ad79-83f8-4fe1-be09-fa2be02d7644\") " pod="openstack/cinder-scheduler-0" Oct 06 10:16:43 crc kubenswrapper[4824]: I1006 10:16:43.249280 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4a56ad79-83f8-4fe1-be09-fa2be02d7644-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"4a56ad79-83f8-4fe1-be09-fa2be02d7644\") " pod="openstack/cinder-scheduler-0" Oct 06 10:16:43 crc kubenswrapper[4824]: I1006 10:16:43.251692 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4a56ad79-83f8-4fe1-be09-fa2be02d7644-scripts\") pod \"cinder-scheduler-0\" (UID: \"4a56ad79-83f8-4fe1-be09-fa2be02d7644\") " pod="openstack/cinder-scheduler-0" Oct 06 10:16:43 crc kubenswrapper[4824]: I1006 10:16:43.252037 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a56ad79-83f8-4fe1-be09-fa2be02d7644-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"4a56ad79-83f8-4fe1-be09-fa2be02d7644\") " pod="openstack/cinder-scheduler-0" Oct 06 10:16:43 crc kubenswrapper[4824]: I1006 10:16:43.254148 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Oct 06 10:16:43 crc kubenswrapper[4824]: I1006 10:16:43.255741 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a56ad79-83f8-4fe1-be09-fa2be02d7644-config-data\") pod \"cinder-scheduler-0\" (UID: \"4a56ad79-83f8-4fe1-be09-fa2be02d7644\") " pod="openstack/cinder-scheduler-0" Oct 06 10:16:43 crc kubenswrapper[4824]: I1006 10:16:43.256764 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 06 10:16:43 crc kubenswrapper[4824]: I1006 10:16:43.257245 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j49q4\" (UniqueName: \"kubernetes.io/projected/4a56ad79-83f8-4fe1-be09-fa2be02d7644-kube-api-access-j49q4\") pod \"cinder-scheduler-0\" (UID: \"4a56ad79-83f8-4fe1-be09-fa2be02d7644\") " pod="openstack/cinder-scheduler-0" Oct 06 10:16:43 crc kubenswrapper[4824]: I1006 10:16:43.262568 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Oct 06 10:16:43 crc kubenswrapper[4824]: I1006 10:16:43.308438 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 06 10:16:43 crc kubenswrapper[4824]: I1006 10:16:43.331339 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 06 10:16:43 crc kubenswrapper[4824]: I1006 10:16:43.335776 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/820fdaac-421d-473e-bbbc-9ac03f95ab75-logs\") pod \"cinder-api-0\" (UID: \"820fdaac-421d-473e-bbbc-9ac03f95ab75\") " pod="openstack/cinder-api-0" Oct 06 10:16:43 crc kubenswrapper[4824]: I1006 10:16:43.335838 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8de9ec83-5bc3-47a8-9a63-84acb7d6b774-dns-swift-storage-0\") pod \"dnsmasq-dns-5784cf869f-k2mqf\" (UID: \"8de9ec83-5bc3-47a8-9a63-84acb7d6b774\") " pod="openstack/dnsmasq-dns-5784cf869f-k2mqf" Oct 06 10:16:43 crc kubenswrapper[4824]: I1006 10:16:43.335868 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wbdxh\" (UniqueName: \"kubernetes.io/projected/8de9ec83-5bc3-47a8-9a63-84acb7d6b774-kube-api-access-wbdxh\") pod \"dnsmasq-dns-5784cf869f-k2mqf\" (UID: \"8de9ec83-5bc3-47a8-9a63-84acb7d6b774\") " pod="openstack/dnsmasq-dns-5784cf869f-k2mqf" Oct 06 10:16:43 crc kubenswrapper[4824]: I1006 10:16:43.335899 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/820fdaac-421d-473e-bbbc-9ac03f95ab75-scripts\") pod \"cinder-api-0\" (UID: \"820fdaac-421d-473e-bbbc-9ac03f95ab75\") " pod="openstack/cinder-api-0" Oct 06 10:16:43 crc kubenswrapper[4824]: I1006 10:16:43.335920 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/820fdaac-421d-473e-bbbc-9ac03f95ab75-config-data\") pod \"cinder-api-0\" (UID: \"820fdaac-421d-473e-bbbc-9ac03f95ab75\") " pod="openstack/cinder-api-0" Oct 06 10:16:43 crc kubenswrapper[4824]: I1006 10:16:43.336013 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8de9ec83-5bc3-47a8-9a63-84acb7d6b774-dns-svc\") pod \"dnsmasq-dns-5784cf869f-k2mqf\" (UID: \"8de9ec83-5bc3-47a8-9a63-84acb7d6b774\") " pod="openstack/dnsmasq-dns-5784cf869f-k2mqf" Oct 06 10:16:43 crc kubenswrapper[4824]: I1006 10:16:43.336040 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/820fdaac-421d-473e-bbbc-9ac03f95ab75-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"820fdaac-421d-473e-bbbc-9ac03f95ab75\") " pod="openstack/cinder-api-0" Oct 06 10:16:43 crc kubenswrapper[4824]: I1006 10:16:43.336066 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/820fdaac-421d-473e-bbbc-9ac03f95ab75-config-data-custom\") pod \"cinder-api-0\" (UID: \"820fdaac-421d-473e-bbbc-9ac03f95ab75\") " pod="openstack/cinder-api-0" Oct 06 10:16:43 crc kubenswrapper[4824]: I1006 10:16:43.336100 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8de9ec83-5bc3-47a8-9a63-84acb7d6b774-ovsdbserver-nb\") pod \"dnsmasq-dns-5784cf869f-k2mqf\" (UID: \"8de9ec83-5bc3-47a8-9a63-84acb7d6b774\") " pod="openstack/dnsmasq-dns-5784cf869f-k2mqf" Oct 06 10:16:43 crc kubenswrapper[4824]: I1006 10:16:43.336140 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8de9ec83-5bc3-47a8-9a63-84acb7d6b774-config\") pod \"dnsmasq-dns-5784cf869f-k2mqf\" (UID: \"8de9ec83-5bc3-47a8-9a63-84acb7d6b774\") " pod="openstack/dnsmasq-dns-5784cf869f-k2mqf" Oct 06 10:16:43 crc kubenswrapper[4824]: I1006 10:16:43.336227 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-scvpn\" (UniqueName: \"kubernetes.io/projected/820fdaac-421d-473e-bbbc-9ac03f95ab75-kube-api-access-scvpn\") pod \"cinder-api-0\" (UID: \"820fdaac-421d-473e-bbbc-9ac03f95ab75\") " pod="openstack/cinder-api-0" Oct 06 10:16:43 crc kubenswrapper[4824]: I1006 10:16:43.336286 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8de9ec83-5bc3-47a8-9a63-84acb7d6b774-ovsdbserver-sb\") pod \"dnsmasq-dns-5784cf869f-k2mqf\" (UID: \"8de9ec83-5bc3-47a8-9a63-84acb7d6b774\") " pod="openstack/dnsmasq-dns-5784cf869f-k2mqf" Oct 06 10:16:43 crc kubenswrapper[4824]: I1006 10:16:43.336317 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/820fdaac-421d-473e-bbbc-9ac03f95ab75-etc-machine-id\") pod \"cinder-api-0\" (UID: \"820fdaac-421d-473e-bbbc-9ac03f95ab75\") " pod="openstack/cinder-api-0" Oct 06 10:16:43 crc kubenswrapper[4824]: I1006 10:16:43.337040 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8de9ec83-5bc3-47a8-9a63-84acb7d6b774-dns-swift-storage-0\") pod \"dnsmasq-dns-5784cf869f-k2mqf\" (UID: \"8de9ec83-5bc3-47a8-9a63-84acb7d6b774\") " pod="openstack/dnsmasq-dns-5784cf869f-k2mqf" Oct 06 10:16:43 crc kubenswrapper[4824]: I1006 10:16:43.338027 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8de9ec83-5bc3-47a8-9a63-84acb7d6b774-ovsdbserver-nb\") pod \"dnsmasq-dns-5784cf869f-k2mqf\" (UID: \"8de9ec83-5bc3-47a8-9a63-84acb7d6b774\") " pod="openstack/dnsmasq-dns-5784cf869f-k2mqf" Oct 06 10:16:43 crc kubenswrapper[4824]: I1006 10:16:43.338351 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8de9ec83-5bc3-47a8-9a63-84acb7d6b774-config\") pod \"dnsmasq-dns-5784cf869f-k2mqf\" (UID: \"8de9ec83-5bc3-47a8-9a63-84acb7d6b774\") " pod="openstack/dnsmasq-dns-5784cf869f-k2mqf" Oct 06 10:16:43 crc kubenswrapper[4824]: I1006 10:16:43.338656 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8de9ec83-5bc3-47a8-9a63-84acb7d6b774-ovsdbserver-sb\") pod \"dnsmasq-dns-5784cf869f-k2mqf\" (UID: \"8de9ec83-5bc3-47a8-9a63-84acb7d6b774\") " pod="openstack/dnsmasq-dns-5784cf869f-k2mqf" Oct 06 10:16:43 crc kubenswrapper[4824]: I1006 10:16:43.341017 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8de9ec83-5bc3-47a8-9a63-84acb7d6b774-dns-svc\") pod \"dnsmasq-dns-5784cf869f-k2mqf\" (UID: \"8de9ec83-5bc3-47a8-9a63-84acb7d6b774\") " pod="openstack/dnsmasq-dns-5784cf869f-k2mqf" Oct 06 10:16:43 crc kubenswrapper[4824]: I1006 10:16:43.358237 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wbdxh\" (UniqueName: \"kubernetes.io/projected/8de9ec83-5bc3-47a8-9a63-84acb7d6b774-kube-api-access-wbdxh\") pod \"dnsmasq-dns-5784cf869f-k2mqf\" (UID: \"8de9ec83-5bc3-47a8-9a63-84acb7d6b774\") " pod="openstack/dnsmasq-dns-5784cf869f-k2mqf" Oct 06 10:16:43 crc kubenswrapper[4824]: I1006 10:16:43.438110 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-scvpn\" (UniqueName: \"kubernetes.io/projected/820fdaac-421d-473e-bbbc-9ac03f95ab75-kube-api-access-scvpn\") pod \"cinder-api-0\" (UID: \"820fdaac-421d-473e-bbbc-9ac03f95ab75\") " pod="openstack/cinder-api-0" Oct 06 10:16:43 crc kubenswrapper[4824]: I1006 10:16:43.438212 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/820fdaac-421d-473e-bbbc-9ac03f95ab75-etc-machine-id\") pod \"cinder-api-0\" (UID: \"820fdaac-421d-473e-bbbc-9ac03f95ab75\") " pod="openstack/cinder-api-0" Oct 06 10:16:43 crc kubenswrapper[4824]: I1006 10:16:43.438268 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/820fdaac-421d-473e-bbbc-9ac03f95ab75-logs\") pod \"cinder-api-0\" (UID: \"820fdaac-421d-473e-bbbc-9ac03f95ab75\") " pod="openstack/cinder-api-0" Oct 06 10:16:43 crc kubenswrapper[4824]: I1006 10:16:43.438318 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/820fdaac-421d-473e-bbbc-9ac03f95ab75-scripts\") pod \"cinder-api-0\" (UID: \"820fdaac-421d-473e-bbbc-9ac03f95ab75\") " pod="openstack/cinder-api-0" Oct 06 10:16:43 crc kubenswrapper[4824]: I1006 10:16:43.438340 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/820fdaac-421d-473e-bbbc-9ac03f95ab75-config-data\") pod \"cinder-api-0\" (UID: \"820fdaac-421d-473e-bbbc-9ac03f95ab75\") " pod="openstack/cinder-api-0" Oct 06 10:16:43 crc kubenswrapper[4824]: I1006 10:16:43.438423 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/820fdaac-421d-473e-bbbc-9ac03f95ab75-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"820fdaac-421d-473e-bbbc-9ac03f95ab75\") " pod="openstack/cinder-api-0" Oct 06 10:16:43 crc kubenswrapper[4824]: I1006 10:16:43.438452 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/820fdaac-421d-473e-bbbc-9ac03f95ab75-config-data-custom\") pod \"cinder-api-0\" (UID: \"820fdaac-421d-473e-bbbc-9ac03f95ab75\") " pod="openstack/cinder-api-0" Oct 06 10:16:43 crc kubenswrapper[4824]: I1006 10:16:43.439711 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/820fdaac-421d-473e-bbbc-9ac03f95ab75-logs\") pod \"cinder-api-0\" (UID: \"820fdaac-421d-473e-bbbc-9ac03f95ab75\") " pod="openstack/cinder-api-0" Oct 06 10:16:43 crc kubenswrapper[4824]: I1006 10:16:43.440218 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/820fdaac-421d-473e-bbbc-9ac03f95ab75-etc-machine-id\") pod \"cinder-api-0\" (UID: \"820fdaac-421d-473e-bbbc-9ac03f95ab75\") " pod="openstack/cinder-api-0" Oct 06 10:16:43 crc kubenswrapper[4824]: I1006 10:16:43.444207 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/820fdaac-421d-473e-bbbc-9ac03f95ab75-config-data-custom\") pod \"cinder-api-0\" (UID: \"820fdaac-421d-473e-bbbc-9ac03f95ab75\") " pod="openstack/cinder-api-0" Oct 06 10:16:43 crc kubenswrapper[4824]: I1006 10:16:43.446930 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/820fdaac-421d-473e-bbbc-9ac03f95ab75-config-data\") pod \"cinder-api-0\" (UID: \"820fdaac-421d-473e-bbbc-9ac03f95ab75\") " pod="openstack/cinder-api-0" Oct 06 10:16:43 crc kubenswrapper[4824]: I1006 10:16:43.447711 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/820fdaac-421d-473e-bbbc-9ac03f95ab75-scripts\") pod \"cinder-api-0\" (UID: \"820fdaac-421d-473e-bbbc-9ac03f95ab75\") " pod="openstack/cinder-api-0" Oct 06 10:16:43 crc kubenswrapper[4824]: I1006 10:16:43.448227 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/820fdaac-421d-473e-bbbc-9ac03f95ab75-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"820fdaac-421d-473e-bbbc-9ac03f95ab75\") " pod="openstack/cinder-api-0" Oct 06 10:16:43 crc kubenswrapper[4824]: I1006 10:16:43.467627 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-scvpn\" (UniqueName: \"kubernetes.io/projected/820fdaac-421d-473e-bbbc-9ac03f95ab75-kube-api-access-scvpn\") pod \"cinder-api-0\" (UID: \"820fdaac-421d-473e-bbbc-9ac03f95ab75\") " pod="openstack/cinder-api-0" Oct 06 10:16:43 crc kubenswrapper[4824]: I1006 10:16:43.477140 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 06 10:16:43 crc kubenswrapper[4824]: I1006 10:16:43.529570 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5784cf869f-k2mqf" Oct 06 10:16:43 crc kubenswrapper[4824]: I1006 10:16:43.803710 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-75c8ddd69c-mgsxh" podUID="932b2879-a6ac-4179-946e-27d4acc68626" containerName="dnsmasq-dns" containerID="cri-o://5c30e360bb6f109a0ef23a8ca46fac513948882f257357f29acaba2deee64f77" gracePeriod=10 Oct 06 10:16:43 crc kubenswrapper[4824]: I1006 10:16:43.804315 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75c8ddd69c-mgsxh" event={"ID":"932b2879-a6ac-4179-946e-27d4acc68626","Type":"ContainerStarted","Data":"5c30e360bb6f109a0ef23a8ca46fac513948882f257357f29acaba2deee64f77"} Oct 06 10:16:43 crc kubenswrapper[4824]: I1006 10:16:43.804374 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-75c8ddd69c-mgsxh" Oct 06 10:16:43 crc kubenswrapper[4824]: I1006 10:16:43.829287 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-78c986c6fb-bzcht" event={"ID":"b8a8eafe-1a7a-4eca-b61c-2d744512a54b","Type":"ContainerStarted","Data":"070086a2cafd4cadd532ea223109610258d84c96f68c7e411fd5f94379ca9b5e"} Oct 06 10:16:43 crc kubenswrapper[4824]: I1006 10:16:43.829352 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-78c986c6fb-bzcht" event={"ID":"b8a8eafe-1a7a-4eca-b61c-2d744512a54b","Type":"ContainerStarted","Data":"fb4079f11b9a49365cbcceaaa236413f8cb6913c9c6b6596fd25e2e7dca1e643"} Oct 06 10:16:43 crc kubenswrapper[4824]: I1006 10:16:43.834263 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-75c8ddd69c-mgsxh" podStartSLOduration=6.834245126 podStartE2EDuration="6.834245126s" podCreationTimestamp="2025-10-06 10:16:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:16:43.833393755 +0000 UTC m=+1113.197816616" watchObservedRunningTime="2025-10-06 10:16:43.834245126 +0000 UTC m=+1113.198667977" Oct 06 10:16:43 crc kubenswrapper[4824]: I1006 10:16:43.851148 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-67979769c4-7fg5z" event={"ID":"5e844107-f8f2-48e8-9731-84eff0f8c752","Type":"ContainerStarted","Data":"8eb0b58050530f290454cbc229376d72bc3943f3457a9c0571e3089eae1b6ad5"} Oct 06 10:16:43 crc kubenswrapper[4824]: I1006 10:16:43.859334 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 06 10:16:43 crc kubenswrapper[4824]: I1006 10:16:43.874579 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-659dd56d6c-qpxtd" event={"ID":"fb46f73f-9d47-4189-990a-051fed587851","Type":"ContainerStarted","Data":"def4b02d3da981a23c52c92d0ca7f730f8dc5d3a15689ea8311b55801810dd4d"} Oct 06 10:16:44 crc kubenswrapper[4824]: I1006 10:16:44.105871 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 06 10:16:44 crc kubenswrapper[4824]: W1006 10:16:44.119450 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod820fdaac_421d_473e_bbbc_9ac03f95ab75.slice/crio-5e4d2f85c22e5145e8d005d76e21461e6d6f3f95e5e9d6ce93403f8e2dc0dd6d WatchSource:0}: Error finding container 5e4d2f85c22e5145e8d005d76e21461e6d6f3f95e5e9d6ce93403f8e2dc0dd6d: Status 404 returned error can't find the container with id 5e4d2f85c22e5145e8d005d76e21461e6d6f3f95e5e9d6ce93403f8e2dc0dd6d Oct 06 10:16:44 crc kubenswrapper[4824]: I1006 10:16:44.269556 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-k2mqf"] Oct 06 10:16:45 crc kubenswrapper[4824]: I1006 10:16:44.465206 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75c8ddd69c-mgsxh" Oct 06 10:16:45 crc kubenswrapper[4824]: I1006 10:16:44.511848 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 06 10:16:45 crc kubenswrapper[4824]: I1006 10:16:44.580007 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/932b2879-a6ac-4179-946e-27d4acc68626-config\") pod \"932b2879-a6ac-4179-946e-27d4acc68626\" (UID: \"932b2879-a6ac-4179-946e-27d4acc68626\") " Oct 06 10:16:45 crc kubenswrapper[4824]: I1006 10:16:44.580079 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/932b2879-a6ac-4179-946e-27d4acc68626-dns-swift-storage-0\") pod \"932b2879-a6ac-4179-946e-27d4acc68626\" (UID: \"932b2879-a6ac-4179-946e-27d4acc68626\") " Oct 06 10:16:45 crc kubenswrapper[4824]: I1006 10:16:44.580147 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/932b2879-a6ac-4179-946e-27d4acc68626-dns-svc\") pod \"932b2879-a6ac-4179-946e-27d4acc68626\" (UID: \"932b2879-a6ac-4179-946e-27d4acc68626\") " Oct 06 10:16:45 crc kubenswrapper[4824]: I1006 10:16:44.580194 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/932b2879-a6ac-4179-946e-27d4acc68626-ovsdbserver-nb\") pod \"932b2879-a6ac-4179-946e-27d4acc68626\" (UID: \"932b2879-a6ac-4179-946e-27d4acc68626\") " Oct 06 10:16:45 crc kubenswrapper[4824]: I1006 10:16:44.580303 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/932b2879-a6ac-4179-946e-27d4acc68626-ovsdbserver-sb\") pod \"932b2879-a6ac-4179-946e-27d4acc68626\" (UID: \"932b2879-a6ac-4179-946e-27d4acc68626\") " Oct 06 10:16:45 crc kubenswrapper[4824]: I1006 10:16:44.580390 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rpqsz\" (UniqueName: \"kubernetes.io/projected/932b2879-a6ac-4179-946e-27d4acc68626-kube-api-access-rpqsz\") pod \"932b2879-a6ac-4179-946e-27d4acc68626\" (UID: \"932b2879-a6ac-4179-946e-27d4acc68626\") " Oct 06 10:16:45 crc kubenswrapper[4824]: I1006 10:16:44.639318 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/932b2879-a6ac-4179-946e-27d4acc68626-kube-api-access-rpqsz" (OuterVolumeSpecName: "kube-api-access-rpqsz") pod "932b2879-a6ac-4179-946e-27d4acc68626" (UID: "932b2879-a6ac-4179-946e-27d4acc68626"). InnerVolumeSpecName "kube-api-access-rpqsz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:16:45 crc kubenswrapper[4824]: I1006 10:16:44.685381 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w6rqx\" (UniqueName: \"kubernetes.io/projected/13781991-323c-49ae-9c5d-625e2ffd019b-kube-api-access-w6rqx\") pod \"13781991-323c-49ae-9c5d-625e2ffd019b\" (UID: \"13781991-323c-49ae-9c5d-625e2ffd019b\") " Oct 06 10:16:45 crc kubenswrapper[4824]: I1006 10:16:44.685544 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/13781991-323c-49ae-9c5d-625e2ffd019b-config-data\") pod \"13781991-323c-49ae-9c5d-625e2ffd019b\" (UID: \"13781991-323c-49ae-9c5d-625e2ffd019b\") " Oct 06 10:16:45 crc kubenswrapper[4824]: I1006 10:16:44.685698 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13781991-323c-49ae-9c5d-625e2ffd019b-combined-ca-bundle\") pod \"13781991-323c-49ae-9c5d-625e2ffd019b\" (UID: \"13781991-323c-49ae-9c5d-625e2ffd019b\") " Oct 06 10:16:45 crc kubenswrapper[4824]: I1006 10:16:44.685778 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/13781991-323c-49ae-9c5d-625e2ffd019b-run-httpd\") pod \"13781991-323c-49ae-9c5d-625e2ffd019b\" (UID: \"13781991-323c-49ae-9c5d-625e2ffd019b\") " Oct 06 10:16:45 crc kubenswrapper[4824]: I1006 10:16:44.685876 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/13781991-323c-49ae-9c5d-625e2ffd019b-log-httpd\") pod \"13781991-323c-49ae-9c5d-625e2ffd019b\" (UID: \"13781991-323c-49ae-9c5d-625e2ffd019b\") " Oct 06 10:16:45 crc kubenswrapper[4824]: I1006 10:16:44.685917 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/13781991-323c-49ae-9c5d-625e2ffd019b-scripts\") pod \"13781991-323c-49ae-9c5d-625e2ffd019b\" (UID: \"13781991-323c-49ae-9c5d-625e2ffd019b\") " Oct 06 10:16:45 crc kubenswrapper[4824]: I1006 10:16:44.686045 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/13781991-323c-49ae-9c5d-625e2ffd019b-sg-core-conf-yaml\") pod \"13781991-323c-49ae-9c5d-625e2ffd019b\" (UID: \"13781991-323c-49ae-9c5d-625e2ffd019b\") " Oct 06 10:16:45 crc kubenswrapper[4824]: I1006 10:16:44.686760 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rpqsz\" (UniqueName: \"kubernetes.io/projected/932b2879-a6ac-4179-946e-27d4acc68626-kube-api-access-rpqsz\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:45 crc kubenswrapper[4824]: I1006 10:16:44.703398 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/13781991-323c-49ae-9c5d-625e2ffd019b-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "13781991-323c-49ae-9c5d-625e2ffd019b" (UID: "13781991-323c-49ae-9c5d-625e2ffd019b"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:16:45 crc kubenswrapper[4824]: I1006 10:16:44.703756 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/13781991-323c-49ae-9c5d-625e2ffd019b-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "13781991-323c-49ae-9c5d-625e2ffd019b" (UID: "13781991-323c-49ae-9c5d-625e2ffd019b"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:16:45 crc kubenswrapper[4824]: I1006 10:16:44.801598 4824 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/13781991-323c-49ae-9c5d-625e2ffd019b-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:45 crc kubenswrapper[4824]: I1006 10:16:44.801627 4824 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/13781991-323c-49ae-9c5d-625e2ffd019b-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:45 crc kubenswrapper[4824]: I1006 10:16:44.892376 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/13781991-323c-49ae-9c5d-625e2ffd019b-kube-api-access-w6rqx" (OuterVolumeSpecName: "kube-api-access-w6rqx") pod "13781991-323c-49ae-9c5d-625e2ffd019b" (UID: "13781991-323c-49ae-9c5d-625e2ffd019b"). InnerVolumeSpecName "kube-api-access-w6rqx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:16:45 crc kubenswrapper[4824]: I1006 10:16:44.905636 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w6rqx\" (UniqueName: \"kubernetes.io/projected/13781991-323c-49ae-9c5d-625e2ffd019b-kube-api-access-w6rqx\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:45 crc kubenswrapper[4824]: I1006 10:16:44.923931 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/13781991-323c-49ae-9c5d-625e2ffd019b-scripts" (OuterVolumeSpecName: "scripts") pod "13781991-323c-49ae-9c5d-625e2ffd019b" (UID: "13781991-323c-49ae-9c5d-625e2ffd019b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:16:45 crc kubenswrapper[4824]: I1006 10:16:44.968650 4824 generic.go:334] "Generic (PLEG): container finished" podID="932b2879-a6ac-4179-946e-27d4acc68626" containerID="5c30e360bb6f109a0ef23a8ca46fac513948882f257357f29acaba2deee64f77" exitCode=0 Oct 06 10:16:45 crc kubenswrapper[4824]: I1006 10:16:44.968810 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75c8ddd69c-mgsxh" event={"ID":"932b2879-a6ac-4179-946e-27d4acc68626","Type":"ContainerDied","Data":"5c30e360bb6f109a0ef23a8ca46fac513948882f257357f29acaba2deee64f77"} Oct 06 10:16:45 crc kubenswrapper[4824]: I1006 10:16:44.968863 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75c8ddd69c-mgsxh" event={"ID":"932b2879-a6ac-4179-946e-27d4acc68626","Type":"ContainerDied","Data":"c3154eb6ae50eceddaa43515eba04c71c4a3fcf033cd2b00d8567d37a8b43034"} Oct 06 10:16:45 crc kubenswrapper[4824]: I1006 10:16:44.968892 4824 scope.go:117] "RemoveContainer" containerID="5c30e360bb6f109a0ef23a8ca46fac513948882f257357f29acaba2deee64f77" Oct 06 10:16:45 crc kubenswrapper[4824]: I1006 10:16:44.969166 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75c8ddd69c-mgsxh" Oct 06 10:16:45 crc kubenswrapper[4824]: I1006 10:16:45.008887 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/13781991-323c-49ae-9c5d-625e2ffd019b-scripts\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:45 crc kubenswrapper[4824]: I1006 10:16:45.028186 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/13781991-323c-49ae-9c5d-625e2ffd019b-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "13781991-323c-49ae-9c5d-625e2ffd019b" (UID: "13781991-323c-49ae-9c5d-625e2ffd019b"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:16:45 crc kubenswrapper[4824]: I1006 10:16:45.048505 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"820fdaac-421d-473e-bbbc-9ac03f95ab75","Type":"ContainerStarted","Data":"5e4d2f85c22e5145e8d005d76e21461e6d6f3f95e5e9d6ce93403f8e2dc0dd6d"} Oct 06 10:16:45 crc kubenswrapper[4824]: I1006 10:16:45.059401 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/932b2879-a6ac-4179-946e-27d4acc68626-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "932b2879-a6ac-4179-946e-27d4acc68626" (UID: "932b2879-a6ac-4179-946e-27d4acc68626"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:16:45 crc kubenswrapper[4824]: I1006 10:16:45.065881 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-78c986c6fb-bzcht" event={"ID":"b8a8eafe-1a7a-4eca-b61c-2d744512a54b","Type":"ContainerStarted","Data":"a13648464fa7733d3d3b4dac1836f13ab9d972bb95cee4e05feae884a78ef5ec"} Oct 06 10:16:45 crc kubenswrapper[4824]: I1006 10:16:45.065960 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-78c986c6fb-bzcht" Oct 06 10:16:45 crc kubenswrapper[4824]: I1006 10:16:45.066004 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-78c986c6fb-bzcht" Oct 06 10:16:45 crc kubenswrapper[4824]: I1006 10:16:45.082360 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/932b2879-a6ac-4179-946e-27d4acc68626-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "932b2879-a6ac-4179-946e-27d4acc68626" (UID: "932b2879-a6ac-4179-946e-27d4acc68626"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:16:45 crc kubenswrapper[4824]: I1006 10:16:45.084294 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-67979769c4-7fg5z" event={"ID":"5e844107-f8f2-48e8-9731-84eff0f8c752","Type":"ContainerStarted","Data":"1ea064fa848d6f5e2c614742555af35dbe892bdb19afd30d3623b9abc31b23d6"} Oct 06 10:16:45 crc kubenswrapper[4824]: I1006 10:16:45.085857 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/932b2879-a6ac-4179-946e-27d4acc68626-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "932b2879-a6ac-4179-946e-27d4acc68626" (UID: "932b2879-a6ac-4179-946e-27d4acc68626"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:16:45 crc kubenswrapper[4824]: I1006 10:16:45.091119 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"4a56ad79-83f8-4fe1-be09-fa2be02d7644","Type":"ContainerStarted","Data":"9824f2cc9d5eb11f41080dc542eaaf7b5d2b7099d462c38b6c1cd386c4819c5f"} Oct 06 10:16:45 crc kubenswrapper[4824]: I1006 10:16:45.112175 4824 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/13781991-323c-49ae-9c5d-625e2ffd019b-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:45 crc kubenswrapper[4824]: I1006 10:16:45.112206 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/932b2879-a6ac-4179-946e-27d4acc68626-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:45 crc kubenswrapper[4824]: I1006 10:16:45.112217 4824 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/932b2879-a6ac-4179-946e-27d4acc68626-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:45 crc kubenswrapper[4824]: I1006 10:16:45.112228 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/932b2879-a6ac-4179-946e-27d4acc68626-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:45 crc kubenswrapper[4824]: I1006 10:16:45.114718 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-78c986c6fb-bzcht" podStartSLOduration=4.11469422 podStartE2EDuration="4.11469422s" podCreationTimestamp="2025-10-06 10:16:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:16:45.095391902 +0000 UTC m=+1114.459814763" watchObservedRunningTime="2025-10-06 10:16:45.11469422 +0000 UTC m=+1114.479117081" Oct 06 10:16:45 crc kubenswrapper[4824]: I1006 10:16:45.129179 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-67979769c4-7fg5z" podStartSLOduration=5.521125414 podStartE2EDuration="8.128794198s" podCreationTimestamp="2025-10-06 10:16:37 +0000 UTC" firstStartedPulling="2025-10-06 10:16:40.616527761 +0000 UTC m=+1109.980950622" lastFinishedPulling="2025-10-06 10:16:43.224196555 +0000 UTC m=+1112.588619406" observedRunningTime="2025-10-06 10:16:45.12481725 +0000 UTC m=+1114.489240121" watchObservedRunningTime="2025-10-06 10:16:45.128794198 +0000 UTC m=+1114.493217059" Oct 06 10:16:45 crc kubenswrapper[4824]: I1006 10:16:45.131601 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/932b2879-a6ac-4179-946e-27d4acc68626-config" (OuterVolumeSpecName: "config") pod "932b2879-a6ac-4179-946e-27d4acc68626" (UID: "932b2879-a6ac-4179-946e-27d4acc68626"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:16:45 crc kubenswrapper[4824]: I1006 10:16:45.138786 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-659dd56d6c-qpxtd" event={"ID":"fb46f73f-9d47-4189-990a-051fed587851","Type":"ContainerStarted","Data":"d31104578444df0a98ec19563ac29b608a3b45071c481af06ab80d699656b2d8"} Oct 06 10:16:45 crc kubenswrapper[4824]: I1006 10:16:45.148942 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/932b2879-a6ac-4179-946e-27d4acc68626-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "932b2879-a6ac-4179-946e-27d4acc68626" (UID: "932b2879-a6ac-4179-946e-27d4acc68626"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:16:45 crc kubenswrapper[4824]: I1006 10:16:45.181782 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/13781991-323c-49ae-9c5d-625e2ffd019b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "13781991-323c-49ae-9c5d-625e2ffd019b" (UID: "13781991-323c-49ae-9c5d-625e2ffd019b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:16:45 crc kubenswrapper[4824]: I1006 10:16:45.194446 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-k2mqf" event={"ID":"8de9ec83-5bc3-47a8-9a63-84acb7d6b774","Type":"ContainerStarted","Data":"eeb1879cd0e57c70a9013aa1b3b8d54c760330a1e66cc57dd26dd82535dfb51d"} Oct 06 10:16:45 crc kubenswrapper[4824]: I1006 10:16:45.204368 4824 generic.go:334] "Generic (PLEG): container finished" podID="13781991-323c-49ae-9c5d-625e2ffd019b" containerID="c75d8022f4b4f06e5bd8a64705b9bfbcd9ee5e6a7edb3e34e026c46d0f72d761" exitCode=0 Oct 06 10:16:45 crc kubenswrapper[4824]: I1006 10:16:45.204422 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"13781991-323c-49ae-9c5d-625e2ffd019b","Type":"ContainerDied","Data":"c75d8022f4b4f06e5bd8a64705b9bfbcd9ee5e6a7edb3e34e026c46d0f72d761"} Oct 06 10:16:45 crc kubenswrapper[4824]: I1006 10:16:45.204457 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"13781991-323c-49ae-9c5d-625e2ffd019b","Type":"ContainerDied","Data":"bcefb09d69c2ac87b5481618d163857107777ec48e52728499f611ab6ba9ea44"} Oct 06 10:16:45 crc kubenswrapper[4824]: I1006 10:16:45.204547 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 06 10:16:45 crc kubenswrapper[4824]: I1006 10:16:45.210383 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/13781991-323c-49ae-9c5d-625e2ffd019b-config-data" (OuterVolumeSpecName: "config-data") pod "13781991-323c-49ae-9c5d-625e2ffd019b" (UID: "13781991-323c-49ae-9c5d-625e2ffd019b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:16:45 crc kubenswrapper[4824]: I1006 10:16:45.220145 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/13781991-323c-49ae-9c5d-625e2ffd019b-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:45 crc kubenswrapper[4824]: I1006 10:16:45.220167 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/932b2879-a6ac-4179-946e-27d4acc68626-config\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:45 crc kubenswrapper[4824]: I1006 10:16:45.220178 4824 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/932b2879-a6ac-4179-946e-27d4acc68626-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:45 crc kubenswrapper[4824]: I1006 10:16:45.220191 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13781991-323c-49ae-9c5d-625e2ffd019b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:45 crc kubenswrapper[4824]: I1006 10:16:45.256646 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-659dd56d6c-qpxtd" podStartSLOduration=5.572544786 podStartE2EDuration="8.256625661s" podCreationTimestamp="2025-10-06 10:16:37 +0000 UTC" firstStartedPulling="2025-10-06 10:16:40.478557478 +0000 UTC m=+1109.842980339" lastFinishedPulling="2025-10-06 10:16:43.162638353 +0000 UTC m=+1112.527061214" observedRunningTime="2025-10-06 10:16:45.184190789 +0000 UTC m=+1114.548613660" watchObservedRunningTime="2025-10-06 10:16:45.256625661 +0000 UTC m=+1114.621048522" Oct 06 10:16:45 crc kubenswrapper[4824]: I1006 10:16:45.450137 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-mgsxh"] Oct 06 10:16:45 crc kubenswrapper[4824]: I1006 10:16:45.464567 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-mgsxh"] Oct 06 10:16:45 crc kubenswrapper[4824]: I1006 10:16:45.485883 4824 scope.go:117] "RemoveContainer" containerID="f64a3cb7ead4a6db4fbf94a4a22a831eb389091f67782891d71e95b15dcd1f56" Oct 06 10:16:45 crc kubenswrapper[4824]: I1006 10:16:45.567649 4824 scope.go:117] "RemoveContainer" containerID="5c30e360bb6f109a0ef23a8ca46fac513948882f257357f29acaba2deee64f77" Oct 06 10:16:45 crc kubenswrapper[4824]: I1006 10:16:45.586419 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 06 10:16:45 crc kubenswrapper[4824]: E1006 10:16:45.586689 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5c30e360bb6f109a0ef23a8ca46fac513948882f257357f29acaba2deee64f77\": container with ID starting with 5c30e360bb6f109a0ef23a8ca46fac513948882f257357f29acaba2deee64f77 not found: ID does not exist" containerID="5c30e360bb6f109a0ef23a8ca46fac513948882f257357f29acaba2deee64f77" Oct 06 10:16:45 crc kubenswrapper[4824]: I1006 10:16:45.587049 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5c30e360bb6f109a0ef23a8ca46fac513948882f257357f29acaba2deee64f77"} err="failed to get container status \"5c30e360bb6f109a0ef23a8ca46fac513948882f257357f29acaba2deee64f77\": rpc error: code = NotFound desc = could not find container \"5c30e360bb6f109a0ef23a8ca46fac513948882f257357f29acaba2deee64f77\": container with ID starting with 5c30e360bb6f109a0ef23a8ca46fac513948882f257357f29acaba2deee64f77 not found: ID does not exist" Oct 06 10:16:45 crc kubenswrapper[4824]: I1006 10:16:45.587172 4824 scope.go:117] "RemoveContainer" containerID="f64a3cb7ead4a6db4fbf94a4a22a831eb389091f67782891d71e95b15dcd1f56" Oct 06 10:16:45 crc kubenswrapper[4824]: E1006 10:16:45.587968 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f64a3cb7ead4a6db4fbf94a4a22a831eb389091f67782891d71e95b15dcd1f56\": container with ID starting with f64a3cb7ead4a6db4fbf94a4a22a831eb389091f67782891d71e95b15dcd1f56 not found: ID does not exist" containerID="f64a3cb7ead4a6db4fbf94a4a22a831eb389091f67782891d71e95b15dcd1f56" Oct 06 10:16:45 crc kubenswrapper[4824]: I1006 10:16:45.588190 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f64a3cb7ead4a6db4fbf94a4a22a831eb389091f67782891d71e95b15dcd1f56"} err="failed to get container status \"f64a3cb7ead4a6db4fbf94a4a22a831eb389091f67782891d71e95b15dcd1f56\": rpc error: code = NotFound desc = could not find container \"f64a3cb7ead4a6db4fbf94a4a22a831eb389091f67782891d71e95b15dcd1f56\": container with ID starting with f64a3cb7ead4a6db4fbf94a4a22a831eb389091f67782891d71e95b15dcd1f56 not found: ID does not exist" Oct 06 10:16:45 crc kubenswrapper[4824]: I1006 10:16:45.588261 4824 scope.go:117] "RemoveContainer" containerID="e5bda72f4bbab5d1f8e75631b7c6e3d944ce6bce9021bc30b8827ebfa9880e75" Oct 06 10:16:45 crc kubenswrapper[4824]: I1006 10:16:45.602393 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 06 10:16:45 crc kubenswrapper[4824]: I1006 10:16:45.621119 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 06 10:16:45 crc kubenswrapper[4824]: E1006 10:16:45.621731 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="932b2879-a6ac-4179-946e-27d4acc68626" containerName="dnsmasq-dns" Oct 06 10:16:45 crc kubenswrapper[4824]: I1006 10:16:45.621794 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="932b2879-a6ac-4179-946e-27d4acc68626" containerName="dnsmasq-dns" Oct 06 10:16:45 crc kubenswrapper[4824]: E1006 10:16:45.621852 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13781991-323c-49ae-9c5d-625e2ffd019b" containerName="sg-core" Oct 06 10:16:45 crc kubenswrapper[4824]: I1006 10:16:45.621899 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="13781991-323c-49ae-9c5d-625e2ffd019b" containerName="sg-core" Oct 06 10:16:45 crc kubenswrapper[4824]: E1006 10:16:45.621956 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13781991-323c-49ae-9c5d-625e2ffd019b" containerName="ceilometer-notification-agent" Oct 06 10:16:45 crc kubenswrapper[4824]: I1006 10:16:45.622037 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="13781991-323c-49ae-9c5d-625e2ffd019b" containerName="ceilometer-notification-agent" Oct 06 10:16:45 crc kubenswrapper[4824]: E1006 10:16:45.622104 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13781991-323c-49ae-9c5d-625e2ffd019b" containerName="proxy-httpd" Oct 06 10:16:45 crc kubenswrapper[4824]: I1006 10:16:45.622151 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="13781991-323c-49ae-9c5d-625e2ffd019b" containerName="proxy-httpd" Oct 06 10:16:45 crc kubenswrapper[4824]: E1006 10:16:45.622210 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="932b2879-a6ac-4179-946e-27d4acc68626" containerName="init" Oct 06 10:16:45 crc kubenswrapper[4824]: I1006 10:16:45.622253 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="932b2879-a6ac-4179-946e-27d4acc68626" containerName="init" Oct 06 10:16:45 crc kubenswrapper[4824]: I1006 10:16:45.622502 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="932b2879-a6ac-4179-946e-27d4acc68626" containerName="dnsmasq-dns" Oct 06 10:16:45 crc kubenswrapper[4824]: I1006 10:16:45.622560 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="13781991-323c-49ae-9c5d-625e2ffd019b" containerName="ceilometer-notification-agent" Oct 06 10:16:45 crc kubenswrapper[4824]: I1006 10:16:45.622630 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="13781991-323c-49ae-9c5d-625e2ffd019b" containerName="proxy-httpd" Oct 06 10:16:45 crc kubenswrapper[4824]: I1006 10:16:45.622679 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="13781991-323c-49ae-9c5d-625e2ffd019b" containerName="sg-core" Oct 06 10:16:45 crc kubenswrapper[4824]: I1006 10:16:45.624763 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 06 10:16:45 crc kubenswrapper[4824]: I1006 10:16:45.628802 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 06 10:16:45 crc kubenswrapper[4824]: I1006 10:16:45.632703 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 06 10:16:45 crc kubenswrapper[4824]: I1006 10:16:45.633680 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 06 10:16:45 crc kubenswrapper[4824]: I1006 10:16:45.672082 4824 scope.go:117] "RemoveContainer" containerID="68acaf0ce4e83b374f4f0fabe242c7a9dd535c1809d4f2f3f0cd8f1a64d60fe5" Oct 06 10:16:45 crc kubenswrapper[4824]: I1006 10:16:45.698346 4824 scope.go:117] "RemoveContainer" containerID="c75d8022f4b4f06e5bd8a64705b9bfbcd9ee5e6a7edb3e34e026c46d0f72d761" Oct 06 10:16:45 crc kubenswrapper[4824]: I1006 10:16:45.760090 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/54d572d5-25f5-4531-bfc0-5a75e1a90cc6-config-data\") pod \"ceilometer-0\" (UID: \"54d572d5-25f5-4531-bfc0-5a75e1a90cc6\") " pod="openstack/ceilometer-0" Oct 06 10:16:45 crc kubenswrapper[4824]: I1006 10:16:45.760135 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/54d572d5-25f5-4531-bfc0-5a75e1a90cc6-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"54d572d5-25f5-4531-bfc0-5a75e1a90cc6\") " pod="openstack/ceilometer-0" Oct 06 10:16:45 crc kubenswrapper[4824]: I1006 10:16:45.760163 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-29xlz\" (UniqueName: \"kubernetes.io/projected/54d572d5-25f5-4531-bfc0-5a75e1a90cc6-kube-api-access-29xlz\") pod \"ceilometer-0\" (UID: \"54d572d5-25f5-4531-bfc0-5a75e1a90cc6\") " pod="openstack/ceilometer-0" Oct 06 10:16:45 crc kubenswrapper[4824]: I1006 10:16:45.760184 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/54d572d5-25f5-4531-bfc0-5a75e1a90cc6-scripts\") pod \"ceilometer-0\" (UID: \"54d572d5-25f5-4531-bfc0-5a75e1a90cc6\") " pod="openstack/ceilometer-0" Oct 06 10:16:45 crc kubenswrapper[4824]: I1006 10:16:45.760214 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/54d572d5-25f5-4531-bfc0-5a75e1a90cc6-run-httpd\") pod \"ceilometer-0\" (UID: \"54d572d5-25f5-4531-bfc0-5a75e1a90cc6\") " pod="openstack/ceilometer-0" Oct 06 10:16:45 crc kubenswrapper[4824]: I1006 10:16:45.760372 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54d572d5-25f5-4531-bfc0-5a75e1a90cc6-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"54d572d5-25f5-4531-bfc0-5a75e1a90cc6\") " pod="openstack/ceilometer-0" Oct 06 10:16:45 crc kubenswrapper[4824]: I1006 10:16:45.760602 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/54d572d5-25f5-4531-bfc0-5a75e1a90cc6-log-httpd\") pod \"ceilometer-0\" (UID: \"54d572d5-25f5-4531-bfc0-5a75e1a90cc6\") " pod="openstack/ceilometer-0" Oct 06 10:16:45 crc kubenswrapper[4824]: I1006 10:16:45.818584 4824 scope.go:117] "RemoveContainer" containerID="e5bda72f4bbab5d1f8e75631b7c6e3d944ce6bce9021bc30b8827ebfa9880e75" Oct 06 10:16:45 crc kubenswrapper[4824]: E1006 10:16:45.822031 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e5bda72f4bbab5d1f8e75631b7c6e3d944ce6bce9021bc30b8827ebfa9880e75\": container with ID starting with e5bda72f4bbab5d1f8e75631b7c6e3d944ce6bce9021bc30b8827ebfa9880e75 not found: ID does not exist" containerID="e5bda72f4bbab5d1f8e75631b7c6e3d944ce6bce9021bc30b8827ebfa9880e75" Oct 06 10:16:45 crc kubenswrapper[4824]: I1006 10:16:45.822377 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e5bda72f4bbab5d1f8e75631b7c6e3d944ce6bce9021bc30b8827ebfa9880e75"} err="failed to get container status \"e5bda72f4bbab5d1f8e75631b7c6e3d944ce6bce9021bc30b8827ebfa9880e75\": rpc error: code = NotFound desc = could not find container \"e5bda72f4bbab5d1f8e75631b7c6e3d944ce6bce9021bc30b8827ebfa9880e75\": container with ID starting with e5bda72f4bbab5d1f8e75631b7c6e3d944ce6bce9021bc30b8827ebfa9880e75 not found: ID does not exist" Oct 06 10:16:45 crc kubenswrapper[4824]: I1006 10:16:45.822419 4824 scope.go:117] "RemoveContainer" containerID="68acaf0ce4e83b374f4f0fabe242c7a9dd535c1809d4f2f3f0cd8f1a64d60fe5" Oct 06 10:16:45 crc kubenswrapper[4824]: E1006 10:16:45.827269 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"68acaf0ce4e83b374f4f0fabe242c7a9dd535c1809d4f2f3f0cd8f1a64d60fe5\": container with ID starting with 68acaf0ce4e83b374f4f0fabe242c7a9dd535c1809d4f2f3f0cd8f1a64d60fe5 not found: ID does not exist" containerID="68acaf0ce4e83b374f4f0fabe242c7a9dd535c1809d4f2f3f0cd8f1a64d60fe5" Oct 06 10:16:45 crc kubenswrapper[4824]: I1006 10:16:45.827332 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"68acaf0ce4e83b374f4f0fabe242c7a9dd535c1809d4f2f3f0cd8f1a64d60fe5"} err="failed to get container status \"68acaf0ce4e83b374f4f0fabe242c7a9dd535c1809d4f2f3f0cd8f1a64d60fe5\": rpc error: code = NotFound desc = could not find container \"68acaf0ce4e83b374f4f0fabe242c7a9dd535c1809d4f2f3f0cd8f1a64d60fe5\": container with ID starting with 68acaf0ce4e83b374f4f0fabe242c7a9dd535c1809d4f2f3f0cd8f1a64d60fe5 not found: ID does not exist" Oct 06 10:16:45 crc kubenswrapper[4824]: I1006 10:16:45.827378 4824 scope.go:117] "RemoveContainer" containerID="c75d8022f4b4f06e5bd8a64705b9bfbcd9ee5e6a7edb3e34e026c46d0f72d761" Oct 06 10:16:45 crc kubenswrapper[4824]: E1006 10:16:45.832549 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c75d8022f4b4f06e5bd8a64705b9bfbcd9ee5e6a7edb3e34e026c46d0f72d761\": container with ID starting with c75d8022f4b4f06e5bd8a64705b9bfbcd9ee5e6a7edb3e34e026c46d0f72d761 not found: ID does not exist" containerID="c75d8022f4b4f06e5bd8a64705b9bfbcd9ee5e6a7edb3e34e026c46d0f72d761" Oct 06 10:16:45 crc kubenswrapper[4824]: I1006 10:16:45.832596 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c75d8022f4b4f06e5bd8a64705b9bfbcd9ee5e6a7edb3e34e026c46d0f72d761"} err="failed to get container status \"c75d8022f4b4f06e5bd8a64705b9bfbcd9ee5e6a7edb3e34e026c46d0f72d761\": rpc error: code = NotFound desc = could not find container \"c75d8022f4b4f06e5bd8a64705b9bfbcd9ee5e6a7edb3e34e026c46d0f72d761\": container with ID starting with c75d8022f4b4f06e5bd8a64705b9bfbcd9ee5e6a7edb3e34e026c46d0f72d761 not found: ID does not exist" Oct 06 10:16:45 crc kubenswrapper[4824]: I1006 10:16:45.862829 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/54d572d5-25f5-4531-bfc0-5a75e1a90cc6-config-data\") pod \"ceilometer-0\" (UID: \"54d572d5-25f5-4531-bfc0-5a75e1a90cc6\") " pod="openstack/ceilometer-0" Oct 06 10:16:45 crc kubenswrapper[4824]: I1006 10:16:45.862887 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/54d572d5-25f5-4531-bfc0-5a75e1a90cc6-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"54d572d5-25f5-4531-bfc0-5a75e1a90cc6\") " pod="openstack/ceilometer-0" Oct 06 10:16:45 crc kubenswrapper[4824]: I1006 10:16:45.862919 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-29xlz\" (UniqueName: \"kubernetes.io/projected/54d572d5-25f5-4531-bfc0-5a75e1a90cc6-kube-api-access-29xlz\") pod \"ceilometer-0\" (UID: \"54d572d5-25f5-4531-bfc0-5a75e1a90cc6\") " pod="openstack/ceilometer-0" Oct 06 10:16:45 crc kubenswrapper[4824]: I1006 10:16:45.862957 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/54d572d5-25f5-4531-bfc0-5a75e1a90cc6-scripts\") pod \"ceilometer-0\" (UID: \"54d572d5-25f5-4531-bfc0-5a75e1a90cc6\") " pod="openstack/ceilometer-0" Oct 06 10:16:45 crc kubenswrapper[4824]: I1006 10:16:45.863020 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/54d572d5-25f5-4531-bfc0-5a75e1a90cc6-run-httpd\") pod \"ceilometer-0\" (UID: \"54d572d5-25f5-4531-bfc0-5a75e1a90cc6\") " pod="openstack/ceilometer-0" Oct 06 10:16:45 crc kubenswrapper[4824]: I1006 10:16:45.863063 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54d572d5-25f5-4531-bfc0-5a75e1a90cc6-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"54d572d5-25f5-4531-bfc0-5a75e1a90cc6\") " pod="openstack/ceilometer-0" Oct 06 10:16:45 crc kubenswrapper[4824]: I1006 10:16:45.863159 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/54d572d5-25f5-4531-bfc0-5a75e1a90cc6-log-httpd\") pod \"ceilometer-0\" (UID: \"54d572d5-25f5-4531-bfc0-5a75e1a90cc6\") " pod="openstack/ceilometer-0" Oct 06 10:16:45 crc kubenswrapper[4824]: I1006 10:16:45.864352 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/54d572d5-25f5-4531-bfc0-5a75e1a90cc6-log-httpd\") pod \"ceilometer-0\" (UID: \"54d572d5-25f5-4531-bfc0-5a75e1a90cc6\") " pod="openstack/ceilometer-0" Oct 06 10:16:45 crc kubenswrapper[4824]: I1006 10:16:45.864872 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/54d572d5-25f5-4531-bfc0-5a75e1a90cc6-run-httpd\") pod \"ceilometer-0\" (UID: \"54d572d5-25f5-4531-bfc0-5a75e1a90cc6\") " pod="openstack/ceilometer-0" Oct 06 10:16:45 crc kubenswrapper[4824]: I1006 10:16:45.870110 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/54d572d5-25f5-4531-bfc0-5a75e1a90cc6-scripts\") pod \"ceilometer-0\" (UID: \"54d572d5-25f5-4531-bfc0-5a75e1a90cc6\") " pod="openstack/ceilometer-0" Oct 06 10:16:45 crc kubenswrapper[4824]: I1006 10:16:45.890542 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/54d572d5-25f5-4531-bfc0-5a75e1a90cc6-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"54d572d5-25f5-4531-bfc0-5a75e1a90cc6\") " pod="openstack/ceilometer-0" Oct 06 10:16:45 crc kubenswrapper[4824]: I1006 10:16:45.890589 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/54d572d5-25f5-4531-bfc0-5a75e1a90cc6-config-data\") pod \"ceilometer-0\" (UID: \"54d572d5-25f5-4531-bfc0-5a75e1a90cc6\") " pod="openstack/ceilometer-0" Oct 06 10:16:45 crc kubenswrapper[4824]: I1006 10:16:45.891141 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-29xlz\" (UniqueName: \"kubernetes.io/projected/54d572d5-25f5-4531-bfc0-5a75e1a90cc6-kube-api-access-29xlz\") pod \"ceilometer-0\" (UID: \"54d572d5-25f5-4531-bfc0-5a75e1a90cc6\") " pod="openstack/ceilometer-0" Oct 06 10:16:45 crc kubenswrapper[4824]: I1006 10:16:45.895447 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54d572d5-25f5-4531-bfc0-5a75e1a90cc6-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"54d572d5-25f5-4531-bfc0-5a75e1a90cc6\") " pod="openstack/ceilometer-0" Oct 06 10:16:45 crc kubenswrapper[4824]: I1006 10:16:45.969885 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 06 10:16:46 crc kubenswrapper[4824]: I1006 10:16:46.263198 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"820fdaac-421d-473e-bbbc-9ac03f95ab75","Type":"ContainerStarted","Data":"1b545a990607f469d55bc679e8ccc9c2f7d8df27eb8306eb02ea7535ac403378"} Oct 06 10:16:46 crc kubenswrapper[4824]: I1006 10:16:46.284393 4824 generic.go:334] "Generic (PLEG): container finished" podID="8de9ec83-5bc3-47a8-9a63-84acb7d6b774" containerID="63dd1ad8bb50bab14c438eb74051e3b85ef276dcae7b29ca87493769fd16c862" exitCode=0 Oct 06 10:16:46 crc kubenswrapper[4824]: I1006 10:16:46.285076 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-k2mqf" event={"ID":"8de9ec83-5bc3-47a8-9a63-84acb7d6b774","Type":"ContainerStarted","Data":"dbc55ac9aa7e345cc7e9f4cc96d3aeba3693c9c1e166bdf53a0fa3d1eb241ab1"} Oct 06 10:16:46 crc kubenswrapper[4824]: I1006 10:16:46.285114 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-k2mqf" event={"ID":"8de9ec83-5bc3-47a8-9a63-84acb7d6b774","Type":"ContainerDied","Data":"63dd1ad8bb50bab14c438eb74051e3b85ef276dcae7b29ca87493769fd16c862"} Oct 06 10:16:46 crc kubenswrapper[4824]: I1006 10:16:46.285152 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5784cf869f-k2mqf" Oct 06 10:16:46 crc kubenswrapper[4824]: I1006 10:16:46.309573 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5784cf869f-k2mqf" podStartSLOduration=3.309555977 podStartE2EDuration="3.309555977s" podCreationTimestamp="2025-10-06 10:16:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:16:46.307012464 +0000 UTC m=+1115.671435335" watchObservedRunningTime="2025-10-06 10:16:46.309555977 +0000 UTC m=+1115.673978838" Oct 06 10:16:46 crc kubenswrapper[4824]: I1006 10:16:46.540083 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 06 10:16:46 crc kubenswrapper[4824]: I1006 10:16:46.677524 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 06 10:16:47 crc kubenswrapper[4824]: I1006 10:16:47.289104 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="13781991-323c-49ae-9c5d-625e2ffd019b" path="/var/lib/kubelet/pods/13781991-323c-49ae-9c5d-625e2ffd019b/volumes" Oct 06 10:16:47 crc kubenswrapper[4824]: I1006 10:16:47.290948 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="932b2879-a6ac-4179-946e-27d4acc68626" path="/var/lib/kubelet/pods/932b2879-a6ac-4179-946e-27d4acc68626/volumes" Oct 06 10:16:47 crc kubenswrapper[4824]: I1006 10:16:47.302805 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"820fdaac-421d-473e-bbbc-9ac03f95ab75","Type":"ContainerStarted","Data":"4eab99fe2b3dad93331f5ba25ee0acd535d47ca8ca01342b50bf2425750798f1"} Oct 06 10:16:47 crc kubenswrapper[4824]: I1006 10:16:47.303047 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="820fdaac-421d-473e-bbbc-9ac03f95ab75" containerName="cinder-api-log" containerID="cri-o://1b545a990607f469d55bc679e8ccc9c2f7d8df27eb8306eb02ea7535ac403378" gracePeriod=30 Oct 06 10:16:47 crc kubenswrapper[4824]: I1006 10:16:47.304148 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Oct 06 10:16:47 crc kubenswrapper[4824]: I1006 10:16:47.304753 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="820fdaac-421d-473e-bbbc-9ac03f95ab75" containerName="cinder-api" containerID="cri-o://4eab99fe2b3dad93331f5ba25ee0acd535d47ca8ca01342b50bf2425750798f1" gracePeriod=30 Oct 06 10:16:47 crc kubenswrapper[4824]: I1006 10:16:47.318292 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"4a56ad79-83f8-4fe1-be09-fa2be02d7644","Type":"ContainerStarted","Data":"3e0c8ceb85bea720a29c6e4e40be493316944e7dbb5ef9fb850e8f56b782e7c2"} Oct 06 10:16:47 crc kubenswrapper[4824]: I1006 10:16:47.322586 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"54d572d5-25f5-4531-bfc0-5a75e1a90cc6","Type":"ContainerStarted","Data":"ab72b288fa642b607e8a99fd684161f7670f804ca0215c1aa53da21e02ef7d9b"} Oct 06 10:16:47 crc kubenswrapper[4824]: I1006 10:16:47.346724 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=4.346697902 podStartE2EDuration="4.346697902s" podCreationTimestamp="2025-10-06 10:16:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:16:47.338649423 +0000 UTC m=+1116.703072284" watchObservedRunningTime="2025-10-06 10:16:47.346697902 +0000 UTC m=+1116.711120763" Oct 06 10:16:48 crc kubenswrapper[4824]: I1006 10:16:48.224493 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-65448d5ddd-mfzdl" Oct 06 10:16:48 crc kubenswrapper[4824]: I1006 10:16:48.319798 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 06 10:16:48 crc kubenswrapper[4824]: I1006 10:16:48.333563 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"54d572d5-25f5-4531-bfc0-5a75e1a90cc6","Type":"ContainerStarted","Data":"1bff648dcf7b79ca89921e2e518b5cb358f03fa3f398af9219f6fb069693118b"} Oct 06 10:16:48 crc kubenswrapper[4824]: I1006 10:16:48.337780 4824 generic.go:334] "Generic (PLEG): container finished" podID="820fdaac-421d-473e-bbbc-9ac03f95ab75" containerID="4eab99fe2b3dad93331f5ba25ee0acd535d47ca8ca01342b50bf2425750798f1" exitCode=0 Oct 06 10:16:48 crc kubenswrapper[4824]: I1006 10:16:48.337801 4824 generic.go:334] "Generic (PLEG): container finished" podID="820fdaac-421d-473e-bbbc-9ac03f95ab75" containerID="1b545a990607f469d55bc679e8ccc9c2f7d8df27eb8306eb02ea7535ac403378" exitCode=143 Oct 06 10:16:48 crc kubenswrapper[4824]: I1006 10:16:48.337828 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"820fdaac-421d-473e-bbbc-9ac03f95ab75","Type":"ContainerDied","Data":"4eab99fe2b3dad93331f5ba25ee0acd535d47ca8ca01342b50bf2425750798f1"} Oct 06 10:16:48 crc kubenswrapper[4824]: I1006 10:16:48.337845 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"820fdaac-421d-473e-bbbc-9ac03f95ab75","Type":"ContainerDied","Data":"1b545a990607f469d55bc679e8ccc9c2f7d8df27eb8306eb02ea7535ac403378"} Oct 06 10:16:48 crc kubenswrapper[4824]: I1006 10:16:48.337854 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"820fdaac-421d-473e-bbbc-9ac03f95ab75","Type":"ContainerDied","Data":"5e4d2f85c22e5145e8d005d76e21461e6d6f3f95e5e9d6ce93403f8e2dc0dd6d"} Oct 06 10:16:48 crc kubenswrapper[4824]: I1006 10:16:48.337868 4824 scope.go:117] "RemoveContainer" containerID="4eab99fe2b3dad93331f5ba25ee0acd535d47ca8ca01342b50bf2425750798f1" Oct 06 10:16:48 crc kubenswrapper[4824]: I1006 10:16:48.337959 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 06 10:16:48 crc kubenswrapper[4824]: I1006 10:16:48.368350 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"4a56ad79-83f8-4fe1-be09-fa2be02d7644","Type":"ContainerStarted","Data":"a4bb0261a913d9416e0e86f150ab44a4aa20058631428603f67dbce53ccaee6b"} Oct 06 10:16:48 crc kubenswrapper[4824]: I1006 10:16:48.413462 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=4.732410797 podStartE2EDuration="6.413431569s" podCreationTimestamp="2025-10-06 10:16:42 +0000 UTC" firstStartedPulling="2025-10-06 10:16:43.933457481 +0000 UTC m=+1113.297880342" lastFinishedPulling="2025-10-06 10:16:45.614478253 +0000 UTC m=+1114.978901114" observedRunningTime="2025-10-06 10:16:48.398366796 +0000 UTC m=+1117.762789667" watchObservedRunningTime="2025-10-06 10:16:48.413431569 +0000 UTC m=+1117.777854430" Oct 06 10:16:48 crc kubenswrapper[4824]: I1006 10:16:48.431261 4824 scope.go:117] "RemoveContainer" containerID="1b545a990607f469d55bc679e8ccc9c2f7d8df27eb8306eb02ea7535ac403378" Oct 06 10:16:48 crc kubenswrapper[4824]: I1006 10:16:48.463606 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/820fdaac-421d-473e-bbbc-9ac03f95ab75-logs\") pod \"820fdaac-421d-473e-bbbc-9ac03f95ab75\" (UID: \"820fdaac-421d-473e-bbbc-9ac03f95ab75\") " Oct 06 10:16:48 crc kubenswrapper[4824]: I1006 10:16:48.463787 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/820fdaac-421d-473e-bbbc-9ac03f95ab75-config-data-custom\") pod \"820fdaac-421d-473e-bbbc-9ac03f95ab75\" (UID: \"820fdaac-421d-473e-bbbc-9ac03f95ab75\") " Oct 06 10:16:48 crc kubenswrapper[4824]: I1006 10:16:48.464083 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/820fdaac-421d-473e-bbbc-9ac03f95ab75-scripts\") pod \"820fdaac-421d-473e-bbbc-9ac03f95ab75\" (UID: \"820fdaac-421d-473e-bbbc-9ac03f95ab75\") " Oct 06 10:16:48 crc kubenswrapper[4824]: I1006 10:16:48.464179 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/820fdaac-421d-473e-bbbc-9ac03f95ab75-config-data\") pod \"820fdaac-421d-473e-bbbc-9ac03f95ab75\" (UID: \"820fdaac-421d-473e-bbbc-9ac03f95ab75\") " Oct 06 10:16:48 crc kubenswrapper[4824]: I1006 10:16:48.464254 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-scvpn\" (UniqueName: \"kubernetes.io/projected/820fdaac-421d-473e-bbbc-9ac03f95ab75-kube-api-access-scvpn\") pod \"820fdaac-421d-473e-bbbc-9ac03f95ab75\" (UID: \"820fdaac-421d-473e-bbbc-9ac03f95ab75\") " Oct 06 10:16:48 crc kubenswrapper[4824]: I1006 10:16:48.464336 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/820fdaac-421d-473e-bbbc-9ac03f95ab75-etc-machine-id\") pod \"820fdaac-421d-473e-bbbc-9ac03f95ab75\" (UID: \"820fdaac-421d-473e-bbbc-9ac03f95ab75\") " Oct 06 10:16:48 crc kubenswrapper[4824]: I1006 10:16:48.464568 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/820fdaac-421d-473e-bbbc-9ac03f95ab75-combined-ca-bundle\") pod \"820fdaac-421d-473e-bbbc-9ac03f95ab75\" (UID: \"820fdaac-421d-473e-bbbc-9ac03f95ab75\") " Oct 06 10:16:48 crc kubenswrapper[4824]: I1006 10:16:48.466425 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/820fdaac-421d-473e-bbbc-9ac03f95ab75-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "820fdaac-421d-473e-bbbc-9ac03f95ab75" (UID: "820fdaac-421d-473e-bbbc-9ac03f95ab75"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 06 10:16:48 crc kubenswrapper[4824]: I1006 10:16:48.466611 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/820fdaac-421d-473e-bbbc-9ac03f95ab75-logs" (OuterVolumeSpecName: "logs") pod "820fdaac-421d-473e-bbbc-9ac03f95ab75" (UID: "820fdaac-421d-473e-bbbc-9ac03f95ab75"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:16:48 crc kubenswrapper[4824]: I1006 10:16:48.484467 4824 scope.go:117] "RemoveContainer" containerID="4eab99fe2b3dad93331f5ba25ee0acd535d47ca8ca01342b50bf2425750798f1" Oct 06 10:16:48 crc kubenswrapper[4824]: I1006 10:16:48.507608 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/820fdaac-421d-473e-bbbc-9ac03f95ab75-scripts" (OuterVolumeSpecName: "scripts") pod "820fdaac-421d-473e-bbbc-9ac03f95ab75" (UID: "820fdaac-421d-473e-bbbc-9ac03f95ab75"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:16:48 crc kubenswrapper[4824]: I1006 10:16:48.507650 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/820fdaac-421d-473e-bbbc-9ac03f95ab75-kube-api-access-scvpn" (OuterVolumeSpecName: "kube-api-access-scvpn") pod "820fdaac-421d-473e-bbbc-9ac03f95ab75" (UID: "820fdaac-421d-473e-bbbc-9ac03f95ab75"). InnerVolumeSpecName "kube-api-access-scvpn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:16:48 crc kubenswrapper[4824]: I1006 10:16:48.507747 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/820fdaac-421d-473e-bbbc-9ac03f95ab75-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "820fdaac-421d-473e-bbbc-9ac03f95ab75" (UID: "820fdaac-421d-473e-bbbc-9ac03f95ab75"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:16:48 crc kubenswrapper[4824]: E1006 10:16:48.507854 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4eab99fe2b3dad93331f5ba25ee0acd535d47ca8ca01342b50bf2425750798f1\": container with ID starting with 4eab99fe2b3dad93331f5ba25ee0acd535d47ca8ca01342b50bf2425750798f1 not found: ID does not exist" containerID="4eab99fe2b3dad93331f5ba25ee0acd535d47ca8ca01342b50bf2425750798f1" Oct 06 10:16:48 crc kubenswrapper[4824]: I1006 10:16:48.508038 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4eab99fe2b3dad93331f5ba25ee0acd535d47ca8ca01342b50bf2425750798f1"} err="failed to get container status \"4eab99fe2b3dad93331f5ba25ee0acd535d47ca8ca01342b50bf2425750798f1\": rpc error: code = NotFound desc = could not find container \"4eab99fe2b3dad93331f5ba25ee0acd535d47ca8ca01342b50bf2425750798f1\": container with ID starting with 4eab99fe2b3dad93331f5ba25ee0acd535d47ca8ca01342b50bf2425750798f1 not found: ID does not exist" Oct 06 10:16:48 crc kubenswrapper[4824]: I1006 10:16:48.508163 4824 scope.go:117] "RemoveContainer" containerID="1b545a990607f469d55bc679e8ccc9c2f7d8df27eb8306eb02ea7535ac403378" Oct 06 10:16:48 crc kubenswrapper[4824]: I1006 10:16:48.517199 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/820fdaac-421d-473e-bbbc-9ac03f95ab75-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "820fdaac-421d-473e-bbbc-9ac03f95ab75" (UID: "820fdaac-421d-473e-bbbc-9ac03f95ab75"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:16:48 crc kubenswrapper[4824]: E1006 10:16:48.526085 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1b545a990607f469d55bc679e8ccc9c2f7d8df27eb8306eb02ea7535ac403378\": container with ID starting with 1b545a990607f469d55bc679e8ccc9c2f7d8df27eb8306eb02ea7535ac403378 not found: ID does not exist" containerID="1b545a990607f469d55bc679e8ccc9c2f7d8df27eb8306eb02ea7535ac403378" Oct 06 10:16:48 crc kubenswrapper[4824]: I1006 10:16:48.526142 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1b545a990607f469d55bc679e8ccc9c2f7d8df27eb8306eb02ea7535ac403378"} err="failed to get container status \"1b545a990607f469d55bc679e8ccc9c2f7d8df27eb8306eb02ea7535ac403378\": rpc error: code = NotFound desc = could not find container \"1b545a990607f469d55bc679e8ccc9c2f7d8df27eb8306eb02ea7535ac403378\": container with ID starting with 1b545a990607f469d55bc679e8ccc9c2f7d8df27eb8306eb02ea7535ac403378 not found: ID does not exist" Oct 06 10:16:48 crc kubenswrapper[4824]: I1006 10:16:48.526169 4824 scope.go:117] "RemoveContainer" containerID="4eab99fe2b3dad93331f5ba25ee0acd535d47ca8ca01342b50bf2425750798f1" Oct 06 10:16:48 crc kubenswrapper[4824]: I1006 10:16:48.530193 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4eab99fe2b3dad93331f5ba25ee0acd535d47ca8ca01342b50bf2425750798f1"} err="failed to get container status \"4eab99fe2b3dad93331f5ba25ee0acd535d47ca8ca01342b50bf2425750798f1\": rpc error: code = NotFound desc = could not find container \"4eab99fe2b3dad93331f5ba25ee0acd535d47ca8ca01342b50bf2425750798f1\": container with ID starting with 4eab99fe2b3dad93331f5ba25ee0acd535d47ca8ca01342b50bf2425750798f1 not found: ID does not exist" Oct 06 10:16:48 crc kubenswrapper[4824]: I1006 10:16:48.530397 4824 scope.go:117] "RemoveContainer" containerID="1b545a990607f469d55bc679e8ccc9c2f7d8df27eb8306eb02ea7535ac403378" Oct 06 10:16:48 crc kubenswrapper[4824]: I1006 10:16:48.533757 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1b545a990607f469d55bc679e8ccc9c2f7d8df27eb8306eb02ea7535ac403378"} err="failed to get container status \"1b545a990607f469d55bc679e8ccc9c2f7d8df27eb8306eb02ea7535ac403378\": rpc error: code = NotFound desc = could not find container \"1b545a990607f469d55bc679e8ccc9c2f7d8df27eb8306eb02ea7535ac403378\": container with ID starting with 1b545a990607f469d55bc679e8ccc9c2f7d8df27eb8306eb02ea7535ac403378 not found: ID does not exist" Oct 06 10:16:48 crc kubenswrapper[4824]: I1006 10:16:48.568049 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/820fdaac-421d-473e-bbbc-9ac03f95ab75-scripts\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:48 crc kubenswrapper[4824]: I1006 10:16:48.568366 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-scvpn\" (UniqueName: \"kubernetes.io/projected/820fdaac-421d-473e-bbbc-9ac03f95ab75-kube-api-access-scvpn\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:48 crc kubenswrapper[4824]: I1006 10:16:48.570662 4824 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/820fdaac-421d-473e-bbbc-9ac03f95ab75-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:48 crc kubenswrapper[4824]: I1006 10:16:48.570756 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/820fdaac-421d-473e-bbbc-9ac03f95ab75-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:48 crc kubenswrapper[4824]: I1006 10:16:48.570834 4824 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/820fdaac-421d-473e-bbbc-9ac03f95ab75-logs\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:48 crc kubenswrapper[4824]: I1006 10:16:48.570915 4824 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/820fdaac-421d-473e-bbbc-9ac03f95ab75-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:48 crc kubenswrapper[4824]: I1006 10:16:48.593802 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/820fdaac-421d-473e-bbbc-9ac03f95ab75-config-data" (OuterVolumeSpecName: "config-data") pod "820fdaac-421d-473e-bbbc-9ac03f95ab75" (UID: "820fdaac-421d-473e-bbbc-9ac03f95ab75"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:16:48 crc kubenswrapper[4824]: I1006 10:16:48.676510 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/820fdaac-421d-473e-bbbc-9ac03f95ab75-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:48 crc kubenswrapper[4824]: I1006 10:16:48.712437 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 06 10:16:48 crc kubenswrapper[4824]: I1006 10:16:48.721310 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Oct 06 10:16:48 crc kubenswrapper[4824]: I1006 10:16:48.747230 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Oct 06 10:16:48 crc kubenswrapper[4824]: E1006 10:16:48.753854 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="820fdaac-421d-473e-bbbc-9ac03f95ab75" containerName="cinder-api-log" Oct 06 10:16:48 crc kubenswrapper[4824]: I1006 10:16:48.754440 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="820fdaac-421d-473e-bbbc-9ac03f95ab75" containerName="cinder-api-log" Oct 06 10:16:48 crc kubenswrapper[4824]: E1006 10:16:48.754510 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="820fdaac-421d-473e-bbbc-9ac03f95ab75" containerName="cinder-api" Oct 06 10:16:48 crc kubenswrapper[4824]: I1006 10:16:48.754562 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="820fdaac-421d-473e-bbbc-9ac03f95ab75" containerName="cinder-api" Oct 06 10:16:48 crc kubenswrapper[4824]: I1006 10:16:48.754860 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="820fdaac-421d-473e-bbbc-9ac03f95ab75" containerName="cinder-api-log" Oct 06 10:16:48 crc kubenswrapper[4824]: I1006 10:16:48.754936 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="820fdaac-421d-473e-bbbc-9ac03f95ab75" containerName="cinder-api" Oct 06 10:16:48 crc kubenswrapper[4824]: I1006 10:16:48.756263 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 06 10:16:48 crc kubenswrapper[4824]: I1006 10:16:48.759784 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Oct 06 10:16:48 crc kubenswrapper[4824]: I1006 10:16:48.761278 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Oct 06 10:16:48 crc kubenswrapper[4824]: I1006 10:16:48.765527 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Oct 06 10:16:48 crc kubenswrapper[4824]: I1006 10:16:48.777523 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 06 10:16:48 crc kubenswrapper[4824]: I1006 10:16:48.880994 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/13a74e55-383d-4105-9c9c-aafb63b75c39-etc-machine-id\") pod \"cinder-api-0\" (UID: \"13a74e55-383d-4105-9c9c-aafb63b75c39\") " pod="openstack/cinder-api-0" Oct 06 10:16:48 crc kubenswrapper[4824]: I1006 10:16:48.881071 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kn7b7\" (UniqueName: \"kubernetes.io/projected/13a74e55-383d-4105-9c9c-aafb63b75c39-kube-api-access-kn7b7\") pod \"cinder-api-0\" (UID: \"13a74e55-383d-4105-9c9c-aafb63b75c39\") " pod="openstack/cinder-api-0" Oct 06 10:16:48 crc kubenswrapper[4824]: I1006 10:16:48.881133 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/13a74e55-383d-4105-9c9c-aafb63b75c39-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"13a74e55-383d-4105-9c9c-aafb63b75c39\") " pod="openstack/cinder-api-0" Oct 06 10:16:48 crc kubenswrapper[4824]: I1006 10:16:48.881169 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/13a74e55-383d-4105-9c9c-aafb63b75c39-logs\") pod \"cinder-api-0\" (UID: \"13a74e55-383d-4105-9c9c-aafb63b75c39\") " pod="openstack/cinder-api-0" Oct 06 10:16:48 crc kubenswrapper[4824]: I1006 10:16:48.881198 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/13a74e55-383d-4105-9c9c-aafb63b75c39-scripts\") pod \"cinder-api-0\" (UID: \"13a74e55-383d-4105-9c9c-aafb63b75c39\") " pod="openstack/cinder-api-0" Oct 06 10:16:48 crc kubenswrapper[4824]: I1006 10:16:48.881222 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/13a74e55-383d-4105-9c9c-aafb63b75c39-config-data-custom\") pod \"cinder-api-0\" (UID: \"13a74e55-383d-4105-9c9c-aafb63b75c39\") " pod="openstack/cinder-api-0" Oct 06 10:16:48 crc kubenswrapper[4824]: I1006 10:16:48.881250 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/13a74e55-383d-4105-9c9c-aafb63b75c39-public-tls-certs\") pod \"cinder-api-0\" (UID: \"13a74e55-383d-4105-9c9c-aafb63b75c39\") " pod="openstack/cinder-api-0" Oct 06 10:16:48 crc kubenswrapper[4824]: I1006 10:16:48.881273 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13a74e55-383d-4105-9c9c-aafb63b75c39-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"13a74e55-383d-4105-9c9c-aafb63b75c39\") " pod="openstack/cinder-api-0" Oct 06 10:16:48 crc kubenswrapper[4824]: I1006 10:16:48.881299 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/13a74e55-383d-4105-9c9c-aafb63b75c39-config-data\") pod \"cinder-api-0\" (UID: \"13a74e55-383d-4105-9c9c-aafb63b75c39\") " pod="openstack/cinder-api-0" Oct 06 10:16:48 crc kubenswrapper[4824]: I1006 10:16:48.982695 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/13a74e55-383d-4105-9c9c-aafb63b75c39-scripts\") pod \"cinder-api-0\" (UID: \"13a74e55-383d-4105-9c9c-aafb63b75c39\") " pod="openstack/cinder-api-0" Oct 06 10:16:48 crc kubenswrapper[4824]: I1006 10:16:48.982756 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/13a74e55-383d-4105-9c9c-aafb63b75c39-config-data-custom\") pod \"cinder-api-0\" (UID: \"13a74e55-383d-4105-9c9c-aafb63b75c39\") " pod="openstack/cinder-api-0" Oct 06 10:16:48 crc kubenswrapper[4824]: I1006 10:16:48.982789 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/13a74e55-383d-4105-9c9c-aafb63b75c39-public-tls-certs\") pod \"cinder-api-0\" (UID: \"13a74e55-383d-4105-9c9c-aafb63b75c39\") " pod="openstack/cinder-api-0" Oct 06 10:16:48 crc kubenswrapper[4824]: I1006 10:16:48.982816 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13a74e55-383d-4105-9c9c-aafb63b75c39-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"13a74e55-383d-4105-9c9c-aafb63b75c39\") " pod="openstack/cinder-api-0" Oct 06 10:16:48 crc kubenswrapper[4824]: I1006 10:16:48.982843 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/13a74e55-383d-4105-9c9c-aafb63b75c39-config-data\") pod \"cinder-api-0\" (UID: \"13a74e55-383d-4105-9c9c-aafb63b75c39\") " pod="openstack/cinder-api-0" Oct 06 10:16:48 crc kubenswrapper[4824]: I1006 10:16:48.982873 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/13a74e55-383d-4105-9c9c-aafb63b75c39-etc-machine-id\") pod \"cinder-api-0\" (UID: \"13a74e55-383d-4105-9c9c-aafb63b75c39\") " pod="openstack/cinder-api-0" Oct 06 10:16:48 crc kubenswrapper[4824]: I1006 10:16:48.982903 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kn7b7\" (UniqueName: \"kubernetes.io/projected/13a74e55-383d-4105-9c9c-aafb63b75c39-kube-api-access-kn7b7\") pod \"cinder-api-0\" (UID: \"13a74e55-383d-4105-9c9c-aafb63b75c39\") " pod="openstack/cinder-api-0" Oct 06 10:16:48 crc kubenswrapper[4824]: I1006 10:16:48.982952 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/13a74e55-383d-4105-9c9c-aafb63b75c39-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"13a74e55-383d-4105-9c9c-aafb63b75c39\") " pod="openstack/cinder-api-0" Oct 06 10:16:48 crc kubenswrapper[4824]: I1006 10:16:48.983001 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/13a74e55-383d-4105-9c9c-aafb63b75c39-logs\") pod \"cinder-api-0\" (UID: \"13a74e55-383d-4105-9c9c-aafb63b75c39\") " pod="openstack/cinder-api-0" Oct 06 10:16:48 crc kubenswrapper[4824]: I1006 10:16:48.983389 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/13a74e55-383d-4105-9c9c-aafb63b75c39-logs\") pod \"cinder-api-0\" (UID: \"13a74e55-383d-4105-9c9c-aafb63b75c39\") " pod="openstack/cinder-api-0" Oct 06 10:16:48 crc kubenswrapper[4824]: I1006 10:16:48.984668 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/13a74e55-383d-4105-9c9c-aafb63b75c39-etc-machine-id\") pod \"cinder-api-0\" (UID: \"13a74e55-383d-4105-9c9c-aafb63b75c39\") " pod="openstack/cinder-api-0" Oct 06 10:16:48 crc kubenswrapper[4824]: I1006 10:16:48.990719 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/13a74e55-383d-4105-9c9c-aafb63b75c39-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"13a74e55-383d-4105-9c9c-aafb63b75c39\") " pod="openstack/cinder-api-0" Oct 06 10:16:48 crc kubenswrapper[4824]: I1006 10:16:48.991912 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13a74e55-383d-4105-9c9c-aafb63b75c39-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"13a74e55-383d-4105-9c9c-aafb63b75c39\") " pod="openstack/cinder-api-0" Oct 06 10:16:48 crc kubenswrapper[4824]: I1006 10:16:48.997558 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/13a74e55-383d-4105-9c9c-aafb63b75c39-public-tls-certs\") pod \"cinder-api-0\" (UID: \"13a74e55-383d-4105-9c9c-aafb63b75c39\") " pod="openstack/cinder-api-0" Oct 06 10:16:48 crc kubenswrapper[4824]: I1006 10:16:48.997937 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/13a74e55-383d-4105-9c9c-aafb63b75c39-scripts\") pod \"cinder-api-0\" (UID: \"13a74e55-383d-4105-9c9c-aafb63b75c39\") " pod="openstack/cinder-api-0" Oct 06 10:16:49 crc kubenswrapper[4824]: I1006 10:16:49.001264 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/13a74e55-383d-4105-9c9c-aafb63b75c39-config-data\") pod \"cinder-api-0\" (UID: \"13a74e55-383d-4105-9c9c-aafb63b75c39\") " pod="openstack/cinder-api-0" Oct 06 10:16:49 crc kubenswrapper[4824]: I1006 10:16:49.013056 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kn7b7\" (UniqueName: \"kubernetes.io/projected/13a74e55-383d-4105-9c9c-aafb63b75c39-kube-api-access-kn7b7\") pod \"cinder-api-0\" (UID: \"13a74e55-383d-4105-9c9c-aafb63b75c39\") " pod="openstack/cinder-api-0" Oct 06 10:16:49 crc kubenswrapper[4824]: I1006 10:16:49.014866 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/13a74e55-383d-4105-9c9c-aafb63b75c39-config-data-custom\") pod \"cinder-api-0\" (UID: \"13a74e55-383d-4105-9c9c-aafb63b75c39\") " pod="openstack/cinder-api-0" Oct 06 10:16:49 crc kubenswrapper[4824]: I1006 10:16:49.156922 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 06 10:16:49 crc kubenswrapper[4824]: I1006 10:16:49.315041 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="820fdaac-421d-473e-bbbc-9ac03f95ab75" path="/var/lib/kubelet/pods/820fdaac-421d-473e-bbbc-9ac03f95ab75/volumes" Oct 06 10:16:49 crc kubenswrapper[4824]: I1006 10:16:49.401674 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"54d572d5-25f5-4531-bfc0-5a75e1a90cc6","Type":"ContainerStarted","Data":"6a35262fed9eb98e214f87a986dd0b40d5221f00dfeef9089eabbd903577021d"} Oct 06 10:16:49 crc kubenswrapper[4824]: I1006 10:16:49.562109 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 06 10:16:50 crc kubenswrapper[4824]: I1006 10:16:50.118040 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-c7c794649-hcxzr" Oct 06 10:16:50 crc kubenswrapper[4824]: I1006 10:16:50.211626 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-65448d5ddd-mfzdl"] Oct 06 10:16:50 crc kubenswrapper[4824]: I1006 10:16:50.211959 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-65448d5ddd-mfzdl" podUID="299892ea-e7b0-455b-9277-865866f47b63" containerName="neutron-api" containerID="cri-o://2cb51964aca5d0c9ee841b955b8b7366929eb144c4e2e4aca4376e930b9de5cf" gracePeriod=30 Oct 06 10:16:50 crc kubenswrapper[4824]: I1006 10:16:50.212167 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-65448d5ddd-mfzdl" podUID="299892ea-e7b0-455b-9277-865866f47b63" containerName="neutron-httpd" containerID="cri-o://d3552bb35c20a93f32e2830bccd410137d4b0c1ca566b67532d960b5ca86d33a" gracePeriod=30 Oct 06 10:16:50 crc kubenswrapper[4824]: I1006 10:16:50.458080 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"13a74e55-383d-4105-9c9c-aafb63b75c39","Type":"ContainerStarted","Data":"8361c0aaf2e50cf18439b5335a4584edad291a18686fcc5160292be4f761c4f5"} Oct 06 10:16:50 crc kubenswrapper[4824]: I1006 10:16:50.486012 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-78c847f9f4-tv8zc" Oct 06 10:16:50 crc kubenswrapper[4824]: I1006 10:16:50.498297 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"54d572d5-25f5-4531-bfc0-5a75e1a90cc6","Type":"ContainerStarted","Data":"465d1fb4fc5bd81e1a6167185933ebfaf1a5fa5bfe19dccc85f947a72337e87f"} Oct 06 10:16:50 crc kubenswrapper[4824]: I1006 10:16:50.900668 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-68fdc5dd48-gfq4d" Oct 06 10:16:51 crc kubenswrapper[4824]: I1006 10:16:51.321052 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-5f478db59d-8kntd" Oct 06 10:16:51 crc kubenswrapper[4824]: I1006 10:16:51.543898 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"13a74e55-383d-4105-9c9c-aafb63b75c39","Type":"ContainerStarted","Data":"8c41abac85797d139e77e774bbad5713ab239d35cd8f2775c360c51b75d03489"} Oct 06 10:16:51 crc kubenswrapper[4824]: I1006 10:16:51.559579 4824 generic.go:334] "Generic (PLEG): container finished" podID="299892ea-e7b0-455b-9277-865866f47b63" containerID="d3552bb35c20a93f32e2830bccd410137d4b0c1ca566b67532d960b5ca86d33a" exitCode=0 Oct 06 10:16:51 crc kubenswrapper[4824]: I1006 10:16:51.560242 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-65448d5ddd-mfzdl" event={"ID":"299892ea-e7b0-455b-9277-865866f47b63","Type":"ContainerDied","Data":"d3552bb35c20a93f32e2830bccd410137d4b0c1ca566b67532d960b5ca86d33a"} Oct 06 10:16:51 crc kubenswrapper[4824]: I1006 10:16:51.647218 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-5f478db59d-8kntd" Oct 06 10:16:52 crc kubenswrapper[4824]: I1006 10:16:52.570968 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"13a74e55-383d-4105-9c9c-aafb63b75c39","Type":"ContainerStarted","Data":"858118436b39f140e599b49dd5fb80f9cc157d0c6bff55a6350025ac50b45dbd"} Oct 06 10:16:52 crc kubenswrapper[4824]: I1006 10:16:52.571693 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Oct 06 10:16:52 crc kubenswrapper[4824]: I1006 10:16:52.575026 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"54d572d5-25f5-4531-bfc0-5a75e1a90cc6","Type":"ContainerStarted","Data":"0300131a20e09d3243c9587d3d719b3b8e6fdbb7eb8b5fd65050788326cf67ff"} Oct 06 10:16:52 crc kubenswrapper[4824]: I1006 10:16:52.575157 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 06 10:16:52 crc kubenswrapper[4824]: I1006 10:16:52.605220 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=4.60520407 podStartE2EDuration="4.60520407s" podCreationTimestamp="2025-10-06 10:16:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:16:52.596173427 +0000 UTC m=+1121.960596278" watchObservedRunningTime="2025-10-06 10:16:52.60520407 +0000 UTC m=+1121.969626931" Oct 06 10:16:52 crc kubenswrapper[4824]: I1006 10:16:52.634618 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.94931045 podStartE2EDuration="7.634598237s" podCreationTimestamp="2025-10-06 10:16:45 +0000 UTC" firstStartedPulling="2025-10-06 10:16:46.686971593 +0000 UTC m=+1116.051394454" lastFinishedPulling="2025-10-06 10:16:51.37225938 +0000 UTC m=+1120.736682241" observedRunningTime="2025-10-06 10:16:52.622874407 +0000 UTC m=+1121.987297268" watchObservedRunningTime="2025-10-06 10:16:52.634598237 +0000 UTC m=+1121.999021098" Oct 06 10:16:53 crc kubenswrapper[4824]: I1006 10:16:53.313054 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Oct 06 10:16:53 crc kubenswrapper[4824]: I1006 10:16:53.498860 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Oct 06 10:16:53 crc kubenswrapper[4824]: I1006 10:16:53.533282 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5784cf869f-k2mqf" Oct 06 10:16:53 crc kubenswrapper[4824]: I1006 10:16:53.608842 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-84b966f6c9-qwzmc"] Oct 06 10:16:53 crc kubenswrapper[4824]: I1006 10:16:53.609098 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-84b966f6c9-qwzmc" podUID="edde940b-8e99-477c-bdd3-9e767af9a4bc" containerName="dnsmasq-dns" containerID="cri-o://91658e4810109b74e59028cee75dd39cbad5b51837f934261d99c5a6fdd8496c" gracePeriod=10 Oct 06 10:16:53 crc kubenswrapper[4824]: I1006 10:16:53.653106 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 06 10:16:53 crc kubenswrapper[4824]: I1006 10:16:53.751418 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-78c986c6fb-bzcht" Oct 06 10:16:53 crc kubenswrapper[4824]: I1006 10:16:53.764915 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-68fdc5dd48-gfq4d" Oct 06 10:16:53 crc kubenswrapper[4824]: I1006 10:16:53.780839 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-78c847f9f4-tv8zc" Oct 06 10:16:53 crc kubenswrapper[4824]: I1006 10:16:53.907776 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-78c847f9f4-tv8zc"] Oct 06 10:16:54 crc kubenswrapper[4824]: I1006 10:16:54.194197 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84b966f6c9-qwzmc" Oct 06 10:16:54 crc kubenswrapper[4824]: I1006 10:16:54.279842 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/edde940b-8e99-477c-bdd3-9e767af9a4bc-ovsdbserver-nb\") pod \"edde940b-8e99-477c-bdd3-9e767af9a4bc\" (UID: \"edde940b-8e99-477c-bdd3-9e767af9a4bc\") " Oct 06 10:16:54 crc kubenswrapper[4824]: I1006 10:16:54.279911 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/edde940b-8e99-477c-bdd3-9e767af9a4bc-ovsdbserver-sb\") pod \"edde940b-8e99-477c-bdd3-9e767af9a4bc\" (UID: \"edde940b-8e99-477c-bdd3-9e767af9a4bc\") " Oct 06 10:16:54 crc kubenswrapper[4824]: I1006 10:16:54.279959 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/edde940b-8e99-477c-bdd3-9e767af9a4bc-dns-swift-storage-0\") pod \"edde940b-8e99-477c-bdd3-9e767af9a4bc\" (UID: \"edde940b-8e99-477c-bdd3-9e767af9a4bc\") " Oct 06 10:16:54 crc kubenswrapper[4824]: I1006 10:16:54.280110 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/edde940b-8e99-477c-bdd3-9e767af9a4bc-dns-svc\") pod \"edde940b-8e99-477c-bdd3-9e767af9a4bc\" (UID: \"edde940b-8e99-477c-bdd3-9e767af9a4bc\") " Oct 06 10:16:54 crc kubenswrapper[4824]: I1006 10:16:54.280271 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8qhq9\" (UniqueName: \"kubernetes.io/projected/edde940b-8e99-477c-bdd3-9e767af9a4bc-kube-api-access-8qhq9\") pod \"edde940b-8e99-477c-bdd3-9e767af9a4bc\" (UID: \"edde940b-8e99-477c-bdd3-9e767af9a4bc\") " Oct 06 10:16:54 crc kubenswrapper[4824]: I1006 10:16:54.280346 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/edde940b-8e99-477c-bdd3-9e767af9a4bc-config\") pod \"edde940b-8e99-477c-bdd3-9e767af9a4bc\" (UID: \"edde940b-8e99-477c-bdd3-9e767af9a4bc\") " Oct 06 10:16:54 crc kubenswrapper[4824]: I1006 10:16:54.307268 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/edde940b-8e99-477c-bdd3-9e767af9a4bc-kube-api-access-8qhq9" (OuterVolumeSpecName: "kube-api-access-8qhq9") pod "edde940b-8e99-477c-bdd3-9e767af9a4bc" (UID: "edde940b-8e99-477c-bdd3-9e767af9a4bc"). InnerVolumeSpecName "kube-api-access-8qhq9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:16:54 crc kubenswrapper[4824]: I1006 10:16:54.363010 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/edde940b-8e99-477c-bdd3-9e767af9a4bc-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "edde940b-8e99-477c-bdd3-9e767af9a4bc" (UID: "edde940b-8e99-477c-bdd3-9e767af9a4bc"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:16:54 crc kubenswrapper[4824]: I1006 10:16:54.383107 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8qhq9\" (UniqueName: \"kubernetes.io/projected/edde940b-8e99-477c-bdd3-9e767af9a4bc-kube-api-access-8qhq9\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:54 crc kubenswrapper[4824]: I1006 10:16:54.383376 4824 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/edde940b-8e99-477c-bdd3-9e767af9a4bc-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:54 crc kubenswrapper[4824]: I1006 10:16:54.385707 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/edde940b-8e99-477c-bdd3-9e767af9a4bc-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "edde940b-8e99-477c-bdd3-9e767af9a4bc" (UID: "edde940b-8e99-477c-bdd3-9e767af9a4bc"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:16:54 crc kubenswrapper[4824]: I1006 10:16:54.389119 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/edde940b-8e99-477c-bdd3-9e767af9a4bc-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "edde940b-8e99-477c-bdd3-9e767af9a4bc" (UID: "edde940b-8e99-477c-bdd3-9e767af9a4bc"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:16:54 crc kubenswrapper[4824]: I1006 10:16:54.394211 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/edde940b-8e99-477c-bdd3-9e767af9a4bc-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "edde940b-8e99-477c-bdd3-9e767af9a4bc" (UID: "edde940b-8e99-477c-bdd3-9e767af9a4bc"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:16:54 crc kubenswrapper[4824]: I1006 10:16:54.396516 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/edde940b-8e99-477c-bdd3-9e767af9a4bc-config" (OuterVolumeSpecName: "config") pod "edde940b-8e99-477c-bdd3-9e767af9a4bc" (UID: "edde940b-8e99-477c-bdd3-9e767af9a4bc"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:16:54 crc kubenswrapper[4824]: I1006 10:16:54.485227 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/edde940b-8e99-477c-bdd3-9e767af9a4bc-config\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:54 crc kubenswrapper[4824]: I1006 10:16:54.485261 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/edde940b-8e99-477c-bdd3-9e767af9a4bc-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:54 crc kubenswrapper[4824]: I1006 10:16:54.485272 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/edde940b-8e99-477c-bdd3-9e767af9a4bc-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:54 crc kubenswrapper[4824]: I1006 10:16:54.485281 4824 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/edde940b-8e99-477c-bdd3-9e767af9a4bc-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:54 crc kubenswrapper[4824]: I1006 10:16:54.506808 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-78c986c6fb-bzcht" Oct 06 10:16:54 crc kubenswrapper[4824]: I1006 10:16:54.581507 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-5f478db59d-8kntd"] Oct 06 10:16:54 crc kubenswrapper[4824]: I1006 10:16:54.581774 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-5f478db59d-8kntd" podUID="1e792caf-ac02-4e6c-9841-ed58133f1918" containerName="barbican-api-log" containerID="cri-o://38b1e0c9677bb96ef4d1872779ef930ca0e79ec2b31786d8860e86c3613db462" gracePeriod=30 Oct 06 10:16:54 crc kubenswrapper[4824]: I1006 10:16:54.581920 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-5f478db59d-8kntd" podUID="1e792caf-ac02-4e6c-9841-ed58133f1918" containerName="barbican-api" containerID="cri-o://717654dbd581395068e7062c4a749a857288b21cb683c2bb83c1b3355e2232ca" gracePeriod=30 Oct 06 10:16:54 crc kubenswrapper[4824]: I1006 10:16:54.593599 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-5f478db59d-8kntd" podUID="1e792caf-ac02-4e6c-9841-ed58133f1918" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.162:9311/healthcheck\": EOF" Oct 06 10:16:54 crc kubenswrapper[4824]: I1006 10:16:54.600573 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-5f478db59d-8kntd" podUID="1e792caf-ac02-4e6c-9841-ed58133f1918" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.162:9311/healthcheck\": EOF" Oct 06 10:16:54 crc kubenswrapper[4824]: I1006 10:16:54.698545 4824 generic.go:334] "Generic (PLEG): container finished" podID="edde940b-8e99-477c-bdd3-9e767af9a4bc" containerID="91658e4810109b74e59028cee75dd39cbad5b51837f934261d99c5a6fdd8496c" exitCode=0 Oct 06 10:16:54 crc kubenswrapper[4824]: I1006 10:16:54.699819 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84b966f6c9-qwzmc" Oct 06 10:16:54 crc kubenswrapper[4824]: I1006 10:16:54.700242 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84b966f6c9-qwzmc" event={"ID":"edde940b-8e99-477c-bdd3-9e767af9a4bc","Type":"ContainerDied","Data":"91658e4810109b74e59028cee75dd39cbad5b51837f934261d99c5a6fdd8496c"} Oct 06 10:16:54 crc kubenswrapper[4824]: I1006 10:16:54.700354 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84b966f6c9-qwzmc" event={"ID":"edde940b-8e99-477c-bdd3-9e767af9a4bc","Type":"ContainerDied","Data":"2de991af05c79688d84d822398a984767c382aabfb3e0b7684f1fe348a6926bc"} Oct 06 10:16:54 crc kubenswrapper[4824]: I1006 10:16:54.700455 4824 scope.go:117] "RemoveContainer" containerID="91658e4810109b74e59028cee75dd39cbad5b51837f934261d99c5a6fdd8496c" Oct 06 10:16:54 crc kubenswrapper[4824]: I1006 10:16:54.701034 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="4a56ad79-83f8-4fe1-be09-fa2be02d7644" containerName="cinder-scheduler" containerID="cri-o://3e0c8ceb85bea720a29c6e4e40be493316944e7dbb5ef9fb850e8f56b782e7c2" gracePeriod=30 Oct 06 10:16:54 crc kubenswrapper[4824]: I1006 10:16:54.701511 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="4a56ad79-83f8-4fe1-be09-fa2be02d7644" containerName="probe" containerID="cri-o://a4bb0261a913d9416e0e86f150ab44a4aa20058631428603f67dbce53ccaee6b" gracePeriod=30 Oct 06 10:16:54 crc kubenswrapper[4824]: I1006 10:16:54.701869 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-78c847f9f4-tv8zc" podUID="68ef92a4-c4bc-4acc-be58-e14010395b8e" containerName="horizon" containerID="cri-o://62bfee4f7781b10bfd4f3b2388238d938ea44501a5d20380123b779025d75b32" gracePeriod=30 Oct 06 10:16:54 crc kubenswrapper[4824]: I1006 10:16:54.701861 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-78c847f9f4-tv8zc" podUID="68ef92a4-c4bc-4acc-be58-e14010395b8e" containerName="horizon-log" containerID="cri-o://e347cab97c01ec84b5735cd302c782ea65d188c9d2c13baf5983c3a2843b3758" gracePeriod=30 Oct 06 10:16:54 crc kubenswrapper[4824]: I1006 10:16:54.813512 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-84b966f6c9-qwzmc"] Oct 06 10:16:54 crc kubenswrapper[4824]: I1006 10:16:54.815619 4824 scope.go:117] "RemoveContainer" containerID="8d3c656a9a067207dfb41d0c1771e02a3f4d418075ce19eecfff8a23b5e7f9e3" Oct 06 10:16:54 crc kubenswrapper[4824]: I1006 10:16:54.841132 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-84b966f6c9-qwzmc"] Oct 06 10:16:54 crc kubenswrapper[4824]: I1006 10:16:54.859101 4824 scope.go:117] "RemoveContainer" containerID="91658e4810109b74e59028cee75dd39cbad5b51837f934261d99c5a6fdd8496c" Oct 06 10:16:54 crc kubenswrapper[4824]: E1006 10:16:54.860377 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"91658e4810109b74e59028cee75dd39cbad5b51837f934261d99c5a6fdd8496c\": container with ID starting with 91658e4810109b74e59028cee75dd39cbad5b51837f934261d99c5a6fdd8496c not found: ID does not exist" containerID="91658e4810109b74e59028cee75dd39cbad5b51837f934261d99c5a6fdd8496c" Oct 06 10:16:54 crc kubenswrapper[4824]: I1006 10:16:54.860462 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"91658e4810109b74e59028cee75dd39cbad5b51837f934261d99c5a6fdd8496c"} err="failed to get container status \"91658e4810109b74e59028cee75dd39cbad5b51837f934261d99c5a6fdd8496c\": rpc error: code = NotFound desc = could not find container \"91658e4810109b74e59028cee75dd39cbad5b51837f934261d99c5a6fdd8496c\": container with ID starting with 91658e4810109b74e59028cee75dd39cbad5b51837f934261d99c5a6fdd8496c not found: ID does not exist" Oct 06 10:16:54 crc kubenswrapper[4824]: I1006 10:16:54.860547 4824 scope.go:117] "RemoveContainer" containerID="8d3c656a9a067207dfb41d0c1771e02a3f4d418075ce19eecfff8a23b5e7f9e3" Oct 06 10:16:54 crc kubenswrapper[4824]: E1006 10:16:54.860899 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8d3c656a9a067207dfb41d0c1771e02a3f4d418075ce19eecfff8a23b5e7f9e3\": container with ID starting with 8d3c656a9a067207dfb41d0c1771e02a3f4d418075ce19eecfff8a23b5e7f9e3 not found: ID does not exist" containerID="8d3c656a9a067207dfb41d0c1771e02a3f4d418075ce19eecfff8a23b5e7f9e3" Oct 06 10:16:54 crc kubenswrapper[4824]: I1006 10:16:54.860942 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8d3c656a9a067207dfb41d0c1771e02a3f4d418075ce19eecfff8a23b5e7f9e3"} err="failed to get container status \"8d3c656a9a067207dfb41d0c1771e02a3f4d418075ce19eecfff8a23b5e7f9e3\": rpc error: code = NotFound desc = could not find container \"8d3c656a9a067207dfb41d0c1771e02a3f4d418075ce19eecfff8a23b5e7f9e3\": container with ID starting with 8d3c656a9a067207dfb41d0c1771e02a3f4d418075ce19eecfff8a23b5e7f9e3 not found: ID does not exist" Oct 06 10:16:55 crc kubenswrapper[4824]: I1006 10:16:55.289354 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="edde940b-8e99-477c-bdd3-9e767af9a4bc" path="/var/lib/kubelet/pods/edde940b-8e99-477c-bdd3-9e767af9a4bc/volumes" Oct 06 10:16:55 crc kubenswrapper[4824]: I1006 10:16:55.716519 4824 generic.go:334] "Generic (PLEG): container finished" podID="1e792caf-ac02-4e6c-9841-ed58133f1918" containerID="38b1e0c9677bb96ef4d1872779ef930ca0e79ec2b31786d8860e86c3613db462" exitCode=143 Oct 06 10:16:55 crc kubenswrapper[4824]: I1006 10:16:55.716616 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5f478db59d-8kntd" event={"ID":"1e792caf-ac02-4e6c-9841-ed58133f1918","Type":"ContainerDied","Data":"38b1e0c9677bb96ef4d1872779ef930ca0e79ec2b31786d8860e86c3613db462"} Oct 06 10:16:55 crc kubenswrapper[4824]: I1006 10:16:55.719799 4824 generic.go:334] "Generic (PLEG): container finished" podID="4a56ad79-83f8-4fe1-be09-fa2be02d7644" containerID="a4bb0261a913d9416e0e86f150ab44a4aa20058631428603f67dbce53ccaee6b" exitCode=0 Oct 06 10:16:55 crc kubenswrapper[4824]: I1006 10:16:55.719901 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"4a56ad79-83f8-4fe1-be09-fa2be02d7644","Type":"ContainerDied","Data":"a4bb0261a913d9416e0e86f150ab44a4aa20058631428603f67dbce53ccaee6b"} Oct 06 10:16:55 crc kubenswrapper[4824]: I1006 10:16:55.721835 4824 generic.go:334] "Generic (PLEG): container finished" podID="299892ea-e7b0-455b-9277-865866f47b63" containerID="2cb51964aca5d0c9ee841b955b8b7366929eb144c4e2e4aca4376e930b9de5cf" exitCode=0 Oct 06 10:16:55 crc kubenswrapper[4824]: I1006 10:16:55.721859 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-65448d5ddd-mfzdl" event={"ID":"299892ea-e7b0-455b-9277-865866f47b63","Type":"ContainerDied","Data":"2cb51964aca5d0c9ee841b955b8b7366929eb144c4e2e4aca4376e930b9de5cf"} Oct 06 10:16:55 crc kubenswrapper[4824]: I1006 10:16:55.848374 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-65448d5ddd-mfzdl" Oct 06 10:16:55 crc kubenswrapper[4824]: I1006 10:16:55.941218 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/299892ea-e7b0-455b-9277-865866f47b63-config\") pod \"299892ea-e7b0-455b-9277-865866f47b63\" (UID: \"299892ea-e7b0-455b-9277-865866f47b63\") " Oct 06 10:16:55 crc kubenswrapper[4824]: I1006 10:16:55.941295 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/299892ea-e7b0-455b-9277-865866f47b63-combined-ca-bundle\") pod \"299892ea-e7b0-455b-9277-865866f47b63\" (UID: \"299892ea-e7b0-455b-9277-865866f47b63\") " Oct 06 10:16:55 crc kubenswrapper[4824]: I1006 10:16:55.941331 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hwlc2\" (UniqueName: \"kubernetes.io/projected/299892ea-e7b0-455b-9277-865866f47b63-kube-api-access-hwlc2\") pod \"299892ea-e7b0-455b-9277-865866f47b63\" (UID: \"299892ea-e7b0-455b-9277-865866f47b63\") " Oct 06 10:16:55 crc kubenswrapper[4824]: I1006 10:16:55.941442 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/299892ea-e7b0-455b-9277-865866f47b63-ovndb-tls-certs\") pod \"299892ea-e7b0-455b-9277-865866f47b63\" (UID: \"299892ea-e7b0-455b-9277-865866f47b63\") " Oct 06 10:16:55 crc kubenswrapper[4824]: I1006 10:16:55.941697 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/299892ea-e7b0-455b-9277-865866f47b63-httpd-config\") pod \"299892ea-e7b0-455b-9277-865866f47b63\" (UID: \"299892ea-e7b0-455b-9277-865866f47b63\") " Oct 06 10:16:55 crc kubenswrapper[4824]: I1006 10:16:55.977406 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/299892ea-e7b0-455b-9277-865866f47b63-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "299892ea-e7b0-455b-9277-865866f47b63" (UID: "299892ea-e7b0-455b-9277-865866f47b63"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:16:55 crc kubenswrapper[4824]: I1006 10:16:55.977856 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/299892ea-e7b0-455b-9277-865866f47b63-kube-api-access-hwlc2" (OuterVolumeSpecName: "kube-api-access-hwlc2") pod "299892ea-e7b0-455b-9277-865866f47b63" (UID: "299892ea-e7b0-455b-9277-865866f47b63"). InnerVolumeSpecName "kube-api-access-hwlc2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:16:56 crc kubenswrapper[4824]: I1006 10:16:56.042205 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/299892ea-e7b0-455b-9277-865866f47b63-config" (OuterVolumeSpecName: "config") pod "299892ea-e7b0-455b-9277-865866f47b63" (UID: "299892ea-e7b0-455b-9277-865866f47b63"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:16:56 crc kubenswrapper[4824]: I1006 10:16:56.043552 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/299892ea-e7b0-455b-9277-865866f47b63-config\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:56 crc kubenswrapper[4824]: I1006 10:16:56.043635 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hwlc2\" (UniqueName: \"kubernetes.io/projected/299892ea-e7b0-455b-9277-865866f47b63-kube-api-access-hwlc2\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:56 crc kubenswrapper[4824]: I1006 10:16:56.043693 4824 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/299892ea-e7b0-455b-9277-865866f47b63-httpd-config\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:56 crc kubenswrapper[4824]: I1006 10:16:56.054260 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/299892ea-e7b0-455b-9277-865866f47b63-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "299892ea-e7b0-455b-9277-865866f47b63" (UID: "299892ea-e7b0-455b-9277-865866f47b63"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:16:56 crc kubenswrapper[4824]: I1006 10:16:56.056751 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/299892ea-e7b0-455b-9277-865866f47b63-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "299892ea-e7b0-455b-9277-865866f47b63" (UID: "299892ea-e7b0-455b-9277-865866f47b63"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:16:56 crc kubenswrapper[4824]: I1006 10:16:56.145510 4824 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/299892ea-e7b0-455b-9277-865866f47b63-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:56 crc kubenswrapper[4824]: I1006 10:16:56.145699 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/299892ea-e7b0-455b-9277-865866f47b63-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:56 crc kubenswrapper[4824]: I1006 10:16:56.451824 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 06 10:16:56 crc kubenswrapper[4824]: I1006 10:16:56.553451 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a56ad79-83f8-4fe1-be09-fa2be02d7644-combined-ca-bundle\") pod \"4a56ad79-83f8-4fe1-be09-fa2be02d7644\" (UID: \"4a56ad79-83f8-4fe1-be09-fa2be02d7644\") " Oct 06 10:16:56 crc kubenswrapper[4824]: I1006 10:16:56.553527 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4a56ad79-83f8-4fe1-be09-fa2be02d7644-etc-machine-id\") pod \"4a56ad79-83f8-4fe1-be09-fa2be02d7644\" (UID: \"4a56ad79-83f8-4fe1-be09-fa2be02d7644\") " Oct 06 10:16:56 crc kubenswrapper[4824]: I1006 10:16:56.553570 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a56ad79-83f8-4fe1-be09-fa2be02d7644-config-data\") pod \"4a56ad79-83f8-4fe1-be09-fa2be02d7644\" (UID: \"4a56ad79-83f8-4fe1-be09-fa2be02d7644\") " Oct 06 10:16:56 crc kubenswrapper[4824]: I1006 10:16:56.553658 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4a56ad79-83f8-4fe1-be09-fa2be02d7644-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "4a56ad79-83f8-4fe1-be09-fa2be02d7644" (UID: "4a56ad79-83f8-4fe1-be09-fa2be02d7644"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 06 10:16:56 crc kubenswrapper[4824]: I1006 10:16:56.553739 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4a56ad79-83f8-4fe1-be09-fa2be02d7644-scripts\") pod \"4a56ad79-83f8-4fe1-be09-fa2be02d7644\" (UID: \"4a56ad79-83f8-4fe1-be09-fa2be02d7644\") " Oct 06 10:16:56 crc kubenswrapper[4824]: I1006 10:16:56.554022 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4a56ad79-83f8-4fe1-be09-fa2be02d7644-config-data-custom\") pod \"4a56ad79-83f8-4fe1-be09-fa2be02d7644\" (UID: \"4a56ad79-83f8-4fe1-be09-fa2be02d7644\") " Oct 06 10:16:56 crc kubenswrapper[4824]: I1006 10:16:56.554112 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j49q4\" (UniqueName: \"kubernetes.io/projected/4a56ad79-83f8-4fe1-be09-fa2be02d7644-kube-api-access-j49q4\") pod \"4a56ad79-83f8-4fe1-be09-fa2be02d7644\" (UID: \"4a56ad79-83f8-4fe1-be09-fa2be02d7644\") " Oct 06 10:16:56 crc kubenswrapper[4824]: I1006 10:16:56.554764 4824 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4a56ad79-83f8-4fe1-be09-fa2be02d7644-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:56 crc kubenswrapper[4824]: I1006 10:16:56.561083 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a56ad79-83f8-4fe1-be09-fa2be02d7644-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "4a56ad79-83f8-4fe1-be09-fa2be02d7644" (UID: "4a56ad79-83f8-4fe1-be09-fa2be02d7644"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:16:56 crc kubenswrapper[4824]: I1006 10:16:56.561191 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4a56ad79-83f8-4fe1-be09-fa2be02d7644-kube-api-access-j49q4" (OuterVolumeSpecName: "kube-api-access-j49q4") pod "4a56ad79-83f8-4fe1-be09-fa2be02d7644" (UID: "4a56ad79-83f8-4fe1-be09-fa2be02d7644"). InnerVolumeSpecName "kube-api-access-j49q4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:16:56 crc kubenswrapper[4824]: I1006 10:16:56.562210 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a56ad79-83f8-4fe1-be09-fa2be02d7644-scripts" (OuterVolumeSpecName: "scripts") pod "4a56ad79-83f8-4fe1-be09-fa2be02d7644" (UID: "4a56ad79-83f8-4fe1-be09-fa2be02d7644"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:16:56 crc kubenswrapper[4824]: I1006 10:16:56.610572 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a56ad79-83f8-4fe1-be09-fa2be02d7644-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4a56ad79-83f8-4fe1-be09-fa2be02d7644" (UID: "4a56ad79-83f8-4fe1-be09-fa2be02d7644"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:16:56 crc kubenswrapper[4824]: I1006 10:16:56.657170 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4a56ad79-83f8-4fe1-be09-fa2be02d7644-scripts\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:56 crc kubenswrapper[4824]: I1006 10:16:56.657205 4824 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4a56ad79-83f8-4fe1-be09-fa2be02d7644-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:56 crc kubenswrapper[4824]: I1006 10:16:56.657218 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j49q4\" (UniqueName: \"kubernetes.io/projected/4a56ad79-83f8-4fe1-be09-fa2be02d7644-kube-api-access-j49q4\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:56 crc kubenswrapper[4824]: I1006 10:16:56.657227 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a56ad79-83f8-4fe1-be09-fa2be02d7644-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:56 crc kubenswrapper[4824]: I1006 10:16:56.679208 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a56ad79-83f8-4fe1-be09-fa2be02d7644-config-data" (OuterVolumeSpecName: "config-data") pod "4a56ad79-83f8-4fe1-be09-fa2be02d7644" (UID: "4a56ad79-83f8-4fe1-be09-fa2be02d7644"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:16:56 crc kubenswrapper[4824]: I1006 10:16:56.735282 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-65448d5ddd-mfzdl" event={"ID":"299892ea-e7b0-455b-9277-865866f47b63","Type":"ContainerDied","Data":"28a0a8ae4d05860d5fe7df5368cd78b30c60ffcde19db5dff14cce8952b9d3aa"} Oct 06 10:16:56 crc kubenswrapper[4824]: I1006 10:16:56.735329 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-65448d5ddd-mfzdl" Oct 06 10:16:56 crc kubenswrapper[4824]: I1006 10:16:56.735346 4824 scope.go:117] "RemoveContainer" containerID="d3552bb35c20a93f32e2830bccd410137d4b0c1ca566b67532d960b5ca86d33a" Oct 06 10:16:56 crc kubenswrapper[4824]: I1006 10:16:56.737603 4824 generic.go:334] "Generic (PLEG): container finished" podID="4a56ad79-83f8-4fe1-be09-fa2be02d7644" containerID="3e0c8ceb85bea720a29c6e4e40be493316944e7dbb5ef9fb850e8f56b782e7c2" exitCode=0 Oct 06 10:16:56 crc kubenswrapper[4824]: I1006 10:16:56.737648 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"4a56ad79-83f8-4fe1-be09-fa2be02d7644","Type":"ContainerDied","Data":"3e0c8ceb85bea720a29c6e4e40be493316944e7dbb5ef9fb850e8f56b782e7c2"} Oct 06 10:16:56 crc kubenswrapper[4824]: I1006 10:16:56.737675 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"4a56ad79-83f8-4fe1-be09-fa2be02d7644","Type":"ContainerDied","Data":"9824f2cc9d5eb11f41080dc542eaaf7b5d2b7099d462c38b6c1cd386c4819c5f"} Oct 06 10:16:56 crc kubenswrapper[4824]: I1006 10:16:56.737734 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 06 10:16:56 crc kubenswrapper[4824]: I1006 10:16:56.758814 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a56ad79-83f8-4fe1-be09-fa2be02d7644-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:56 crc kubenswrapper[4824]: I1006 10:16:56.776897 4824 scope.go:117] "RemoveContainer" containerID="2cb51964aca5d0c9ee841b955b8b7366929eb144c4e2e4aca4376e930b9de5cf" Oct 06 10:16:56 crc kubenswrapper[4824]: I1006 10:16:56.840419 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-65448d5ddd-mfzdl"] Oct 06 10:16:56 crc kubenswrapper[4824]: I1006 10:16:56.845390 4824 scope.go:117] "RemoveContainer" containerID="a4bb0261a913d9416e0e86f150ab44a4aa20058631428603f67dbce53ccaee6b" Oct 06 10:16:56 crc kubenswrapper[4824]: I1006 10:16:56.855056 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-65448d5ddd-mfzdl"] Oct 06 10:16:56 crc kubenswrapper[4824]: I1006 10:16:56.891400 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 06 10:16:56 crc kubenswrapper[4824]: I1006 10:16:56.898935 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 06 10:16:56 crc kubenswrapper[4824]: I1006 10:16:56.909995 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Oct 06 10:16:56 crc kubenswrapper[4824]: E1006 10:16:56.914410 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="299892ea-e7b0-455b-9277-865866f47b63" containerName="neutron-api" Oct 06 10:16:56 crc kubenswrapper[4824]: I1006 10:16:56.914444 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="299892ea-e7b0-455b-9277-865866f47b63" containerName="neutron-api" Oct 06 10:16:56 crc kubenswrapper[4824]: E1006 10:16:56.914465 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="edde940b-8e99-477c-bdd3-9e767af9a4bc" containerName="dnsmasq-dns" Oct 06 10:16:56 crc kubenswrapper[4824]: I1006 10:16:56.914472 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="edde940b-8e99-477c-bdd3-9e767af9a4bc" containerName="dnsmasq-dns" Oct 06 10:16:56 crc kubenswrapper[4824]: E1006 10:16:56.914502 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="299892ea-e7b0-455b-9277-865866f47b63" containerName="neutron-httpd" Oct 06 10:16:56 crc kubenswrapper[4824]: I1006 10:16:56.914508 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="299892ea-e7b0-455b-9277-865866f47b63" containerName="neutron-httpd" Oct 06 10:16:56 crc kubenswrapper[4824]: E1006 10:16:56.914517 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a56ad79-83f8-4fe1-be09-fa2be02d7644" containerName="cinder-scheduler" Oct 06 10:16:56 crc kubenswrapper[4824]: I1006 10:16:56.914522 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a56ad79-83f8-4fe1-be09-fa2be02d7644" containerName="cinder-scheduler" Oct 06 10:16:56 crc kubenswrapper[4824]: E1006 10:16:56.914538 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="edde940b-8e99-477c-bdd3-9e767af9a4bc" containerName="init" Oct 06 10:16:56 crc kubenswrapper[4824]: I1006 10:16:56.914545 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="edde940b-8e99-477c-bdd3-9e767af9a4bc" containerName="init" Oct 06 10:16:56 crc kubenswrapper[4824]: E1006 10:16:56.914560 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a56ad79-83f8-4fe1-be09-fa2be02d7644" containerName="probe" Oct 06 10:16:56 crc kubenswrapper[4824]: I1006 10:16:56.914566 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a56ad79-83f8-4fe1-be09-fa2be02d7644" containerName="probe" Oct 06 10:16:56 crc kubenswrapper[4824]: I1006 10:16:56.914800 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="299892ea-e7b0-455b-9277-865866f47b63" containerName="neutron-httpd" Oct 06 10:16:56 crc kubenswrapper[4824]: I1006 10:16:56.914821 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="299892ea-e7b0-455b-9277-865866f47b63" containerName="neutron-api" Oct 06 10:16:56 crc kubenswrapper[4824]: I1006 10:16:56.914831 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="4a56ad79-83f8-4fe1-be09-fa2be02d7644" containerName="probe" Oct 06 10:16:56 crc kubenswrapper[4824]: I1006 10:16:56.914851 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="4a56ad79-83f8-4fe1-be09-fa2be02d7644" containerName="cinder-scheduler" Oct 06 10:16:56 crc kubenswrapper[4824]: I1006 10:16:56.914866 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="edde940b-8e99-477c-bdd3-9e767af9a4bc" containerName="dnsmasq-dns" Oct 06 10:16:56 crc kubenswrapper[4824]: I1006 10:16:56.922368 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 06 10:16:56 crc kubenswrapper[4824]: I1006 10:16:56.922489 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 06 10:16:56 crc kubenswrapper[4824]: I1006 10:16:56.925949 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Oct 06 10:16:56 crc kubenswrapper[4824]: I1006 10:16:56.928800 4824 scope.go:117] "RemoveContainer" containerID="3e0c8ceb85bea720a29c6e4e40be493316944e7dbb5ef9fb850e8f56b782e7c2" Oct 06 10:16:56 crc kubenswrapper[4824]: I1006 10:16:56.959344 4824 scope.go:117] "RemoveContainer" containerID="a4bb0261a913d9416e0e86f150ab44a4aa20058631428603f67dbce53ccaee6b" Oct 06 10:16:56 crc kubenswrapper[4824]: E1006 10:16:56.959869 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a4bb0261a913d9416e0e86f150ab44a4aa20058631428603f67dbce53ccaee6b\": container with ID starting with a4bb0261a913d9416e0e86f150ab44a4aa20058631428603f67dbce53ccaee6b not found: ID does not exist" containerID="a4bb0261a913d9416e0e86f150ab44a4aa20058631428603f67dbce53ccaee6b" Oct 06 10:16:56 crc kubenswrapper[4824]: I1006 10:16:56.959898 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a4bb0261a913d9416e0e86f150ab44a4aa20058631428603f67dbce53ccaee6b"} err="failed to get container status \"a4bb0261a913d9416e0e86f150ab44a4aa20058631428603f67dbce53ccaee6b\": rpc error: code = NotFound desc = could not find container \"a4bb0261a913d9416e0e86f150ab44a4aa20058631428603f67dbce53ccaee6b\": container with ID starting with a4bb0261a913d9416e0e86f150ab44a4aa20058631428603f67dbce53ccaee6b not found: ID does not exist" Oct 06 10:16:56 crc kubenswrapper[4824]: I1006 10:16:56.959919 4824 scope.go:117] "RemoveContainer" containerID="3e0c8ceb85bea720a29c6e4e40be493316944e7dbb5ef9fb850e8f56b782e7c2" Oct 06 10:16:56 crc kubenswrapper[4824]: E1006 10:16:56.960398 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3e0c8ceb85bea720a29c6e4e40be493316944e7dbb5ef9fb850e8f56b782e7c2\": container with ID starting with 3e0c8ceb85bea720a29c6e4e40be493316944e7dbb5ef9fb850e8f56b782e7c2 not found: ID does not exist" containerID="3e0c8ceb85bea720a29c6e4e40be493316944e7dbb5ef9fb850e8f56b782e7c2" Oct 06 10:16:56 crc kubenswrapper[4824]: I1006 10:16:56.960417 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3e0c8ceb85bea720a29c6e4e40be493316944e7dbb5ef9fb850e8f56b782e7c2"} err="failed to get container status \"3e0c8ceb85bea720a29c6e4e40be493316944e7dbb5ef9fb850e8f56b782e7c2\": rpc error: code = NotFound desc = could not find container \"3e0c8ceb85bea720a29c6e4e40be493316944e7dbb5ef9fb850e8f56b782e7c2\": container with ID starting with 3e0c8ceb85bea720a29c6e4e40be493316944e7dbb5ef9fb850e8f56b782e7c2 not found: ID does not exist" Oct 06 10:16:57 crc kubenswrapper[4824]: I1006 10:16:57.072896 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/18194590-7963-41e8-9253-dcf486d2cea4-scripts\") pod \"cinder-scheduler-0\" (UID: \"18194590-7963-41e8-9253-dcf486d2cea4\") " pod="openstack/cinder-scheduler-0" Oct 06 10:16:57 crc kubenswrapper[4824]: I1006 10:16:57.073043 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18194590-7963-41e8-9253-dcf486d2cea4-config-data\") pod \"cinder-scheduler-0\" (UID: \"18194590-7963-41e8-9253-dcf486d2cea4\") " pod="openstack/cinder-scheduler-0" Oct 06 10:16:57 crc kubenswrapper[4824]: I1006 10:16:57.073105 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/18194590-7963-41e8-9253-dcf486d2cea4-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"18194590-7963-41e8-9253-dcf486d2cea4\") " pod="openstack/cinder-scheduler-0" Oct 06 10:16:57 crc kubenswrapper[4824]: I1006 10:16:57.073369 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jmhgg\" (UniqueName: \"kubernetes.io/projected/18194590-7963-41e8-9253-dcf486d2cea4-kube-api-access-jmhgg\") pod \"cinder-scheduler-0\" (UID: \"18194590-7963-41e8-9253-dcf486d2cea4\") " pod="openstack/cinder-scheduler-0" Oct 06 10:16:57 crc kubenswrapper[4824]: I1006 10:16:57.073584 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/18194590-7963-41e8-9253-dcf486d2cea4-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"18194590-7963-41e8-9253-dcf486d2cea4\") " pod="openstack/cinder-scheduler-0" Oct 06 10:16:57 crc kubenswrapper[4824]: I1006 10:16:57.073709 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18194590-7963-41e8-9253-dcf486d2cea4-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"18194590-7963-41e8-9253-dcf486d2cea4\") " pod="openstack/cinder-scheduler-0" Oct 06 10:16:57 crc kubenswrapper[4824]: I1006 10:16:57.176088 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/18194590-7963-41e8-9253-dcf486d2cea4-scripts\") pod \"cinder-scheduler-0\" (UID: \"18194590-7963-41e8-9253-dcf486d2cea4\") " pod="openstack/cinder-scheduler-0" Oct 06 10:16:57 crc kubenswrapper[4824]: I1006 10:16:57.176173 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18194590-7963-41e8-9253-dcf486d2cea4-config-data\") pod \"cinder-scheduler-0\" (UID: \"18194590-7963-41e8-9253-dcf486d2cea4\") " pod="openstack/cinder-scheduler-0" Oct 06 10:16:57 crc kubenswrapper[4824]: I1006 10:16:57.176205 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/18194590-7963-41e8-9253-dcf486d2cea4-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"18194590-7963-41e8-9253-dcf486d2cea4\") " pod="openstack/cinder-scheduler-0" Oct 06 10:16:57 crc kubenswrapper[4824]: I1006 10:16:57.176315 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jmhgg\" (UniqueName: \"kubernetes.io/projected/18194590-7963-41e8-9253-dcf486d2cea4-kube-api-access-jmhgg\") pod \"cinder-scheduler-0\" (UID: \"18194590-7963-41e8-9253-dcf486d2cea4\") " pod="openstack/cinder-scheduler-0" Oct 06 10:16:57 crc kubenswrapper[4824]: I1006 10:16:57.176363 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/18194590-7963-41e8-9253-dcf486d2cea4-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"18194590-7963-41e8-9253-dcf486d2cea4\") " pod="openstack/cinder-scheduler-0" Oct 06 10:16:57 crc kubenswrapper[4824]: I1006 10:16:57.176393 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18194590-7963-41e8-9253-dcf486d2cea4-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"18194590-7963-41e8-9253-dcf486d2cea4\") " pod="openstack/cinder-scheduler-0" Oct 06 10:16:57 crc kubenswrapper[4824]: I1006 10:16:57.176488 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/18194590-7963-41e8-9253-dcf486d2cea4-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"18194590-7963-41e8-9253-dcf486d2cea4\") " pod="openstack/cinder-scheduler-0" Oct 06 10:16:57 crc kubenswrapper[4824]: I1006 10:16:57.183281 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/18194590-7963-41e8-9253-dcf486d2cea4-scripts\") pod \"cinder-scheduler-0\" (UID: \"18194590-7963-41e8-9253-dcf486d2cea4\") " pod="openstack/cinder-scheduler-0" Oct 06 10:16:57 crc kubenswrapper[4824]: I1006 10:16:57.191612 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18194590-7963-41e8-9253-dcf486d2cea4-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"18194590-7963-41e8-9253-dcf486d2cea4\") " pod="openstack/cinder-scheduler-0" Oct 06 10:16:57 crc kubenswrapper[4824]: I1006 10:16:57.192261 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/18194590-7963-41e8-9253-dcf486d2cea4-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"18194590-7963-41e8-9253-dcf486d2cea4\") " pod="openstack/cinder-scheduler-0" Oct 06 10:16:57 crc kubenswrapper[4824]: I1006 10:16:57.198256 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18194590-7963-41e8-9253-dcf486d2cea4-config-data\") pod \"cinder-scheduler-0\" (UID: \"18194590-7963-41e8-9253-dcf486d2cea4\") " pod="openstack/cinder-scheduler-0" Oct 06 10:16:57 crc kubenswrapper[4824]: I1006 10:16:57.206118 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jmhgg\" (UniqueName: \"kubernetes.io/projected/18194590-7963-41e8-9253-dcf486d2cea4-kube-api-access-jmhgg\") pod \"cinder-scheduler-0\" (UID: \"18194590-7963-41e8-9253-dcf486d2cea4\") " pod="openstack/cinder-scheduler-0" Oct 06 10:16:57 crc kubenswrapper[4824]: I1006 10:16:57.238367 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 06 10:16:57 crc kubenswrapper[4824]: I1006 10:16:57.305543 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="299892ea-e7b0-455b-9277-865866f47b63" path="/var/lib/kubelet/pods/299892ea-e7b0-455b-9277-865866f47b63/volumes" Oct 06 10:16:57 crc kubenswrapper[4824]: I1006 10:16:57.306865 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4a56ad79-83f8-4fe1-be09-fa2be02d7644" path="/var/lib/kubelet/pods/4a56ad79-83f8-4fe1-be09-fa2be02d7644/volumes" Oct 06 10:16:57 crc kubenswrapper[4824]: I1006 10:16:57.836575 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 06 10:16:57 crc kubenswrapper[4824]: I1006 10:16:57.861398 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-78c847f9f4-tv8zc" podUID="68ef92a4-c4bc-4acc-be58-e14010395b8e" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.149:8443/dashboard/auth/login/?next=/dashboard/\": read tcp 10.217.0.2:37444->10.217.0.149:8443: read: connection reset by peer" Oct 06 10:16:58 crc kubenswrapper[4824]: I1006 10:16:58.775034 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"18194590-7963-41e8-9253-dcf486d2cea4","Type":"ContainerStarted","Data":"a2113b72ae8b3ac2ff2c365faf70b61a4766d58647515d5aad396bef53270db0"} Oct 06 10:16:58 crc kubenswrapper[4824]: I1006 10:16:58.776128 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"18194590-7963-41e8-9253-dcf486d2cea4","Type":"ContainerStarted","Data":"b306b259763b4ed6ce1e68f2281a580bae4a93c50bce41b0838acad12dfa2820"} Oct 06 10:16:58 crc kubenswrapper[4824]: I1006 10:16:58.779693 4824 generic.go:334] "Generic (PLEG): container finished" podID="68ef92a4-c4bc-4acc-be58-e14010395b8e" containerID="62bfee4f7781b10bfd4f3b2388238d938ea44501a5d20380123b779025d75b32" exitCode=0 Oct 06 10:16:58 crc kubenswrapper[4824]: I1006 10:16:58.779757 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-78c847f9f4-tv8zc" event={"ID":"68ef92a4-c4bc-4acc-be58-e14010395b8e","Type":"ContainerDied","Data":"62bfee4f7781b10bfd4f3b2388238d938ea44501a5d20380123b779025d75b32"} Oct 06 10:16:59 crc kubenswrapper[4824]: I1006 10:16:59.008264 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-5f478db59d-8kntd" podUID="1e792caf-ac02-4e6c-9841-ed58133f1918" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.162:9311/healthcheck\": read tcp 10.217.0.2:35828->10.217.0.162:9311: read: connection reset by peer" Oct 06 10:16:59 crc kubenswrapper[4824]: I1006 10:16:59.008814 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-5f478db59d-8kntd" podUID="1e792caf-ac02-4e6c-9841-ed58133f1918" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.162:9311/healthcheck\": dial tcp 10.217.0.162:9311: connect: connection refused" Oct 06 10:16:59 crc kubenswrapper[4824]: I1006 10:16:59.008950 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-5f478db59d-8kntd" Oct 06 10:16:59 crc kubenswrapper[4824]: I1006 10:16:59.008931 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-5f478db59d-8kntd" podUID="1e792caf-ac02-4e6c-9841-ed58133f1918" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.162:9311/healthcheck\": read tcp 10.217.0.2:36722->10.217.0.162:9311: read: connection reset by peer" Oct 06 10:16:59 crc kubenswrapper[4824]: I1006 10:16:59.559438 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5f478db59d-8kntd" Oct 06 10:16:59 crc kubenswrapper[4824]: I1006 10:16:59.644706 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e792caf-ac02-4e6c-9841-ed58133f1918-config-data\") pod \"1e792caf-ac02-4e6c-9841-ed58133f1918\" (UID: \"1e792caf-ac02-4e6c-9841-ed58133f1918\") " Oct 06 10:16:59 crc kubenswrapper[4824]: I1006 10:16:59.644900 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vgb9v\" (UniqueName: \"kubernetes.io/projected/1e792caf-ac02-4e6c-9841-ed58133f1918-kube-api-access-vgb9v\") pod \"1e792caf-ac02-4e6c-9841-ed58133f1918\" (UID: \"1e792caf-ac02-4e6c-9841-ed58133f1918\") " Oct 06 10:16:59 crc kubenswrapper[4824]: I1006 10:16:59.644950 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1e792caf-ac02-4e6c-9841-ed58133f1918-config-data-custom\") pod \"1e792caf-ac02-4e6c-9841-ed58133f1918\" (UID: \"1e792caf-ac02-4e6c-9841-ed58133f1918\") " Oct 06 10:16:59 crc kubenswrapper[4824]: I1006 10:16:59.645054 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1e792caf-ac02-4e6c-9841-ed58133f1918-logs\") pod \"1e792caf-ac02-4e6c-9841-ed58133f1918\" (UID: \"1e792caf-ac02-4e6c-9841-ed58133f1918\") " Oct 06 10:16:59 crc kubenswrapper[4824]: I1006 10:16:59.645218 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e792caf-ac02-4e6c-9841-ed58133f1918-combined-ca-bundle\") pod \"1e792caf-ac02-4e6c-9841-ed58133f1918\" (UID: \"1e792caf-ac02-4e6c-9841-ed58133f1918\") " Oct 06 10:16:59 crc kubenswrapper[4824]: I1006 10:16:59.646774 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1e792caf-ac02-4e6c-9841-ed58133f1918-logs" (OuterVolumeSpecName: "logs") pod "1e792caf-ac02-4e6c-9841-ed58133f1918" (UID: "1e792caf-ac02-4e6c-9841-ed58133f1918"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:16:59 crc kubenswrapper[4824]: I1006 10:16:59.664437 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e792caf-ac02-4e6c-9841-ed58133f1918-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "1e792caf-ac02-4e6c-9841-ed58133f1918" (UID: "1e792caf-ac02-4e6c-9841-ed58133f1918"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:16:59 crc kubenswrapper[4824]: I1006 10:16:59.668247 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1e792caf-ac02-4e6c-9841-ed58133f1918-kube-api-access-vgb9v" (OuterVolumeSpecName: "kube-api-access-vgb9v") pod "1e792caf-ac02-4e6c-9841-ed58133f1918" (UID: "1e792caf-ac02-4e6c-9841-ed58133f1918"). InnerVolumeSpecName "kube-api-access-vgb9v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:16:59 crc kubenswrapper[4824]: I1006 10:16:59.689308 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e792caf-ac02-4e6c-9841-ed58133f1918-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1e792caf-ac02-4e6c-9841-ed58133f1918" (UID: "1e792caf-ac02-4e6c-9841-ed58133f1918"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:16:59 crc kubenswrapper[4824]: I1006 10:16:59.710056 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e792caf-ac02-4e6c-9841-ed58133f1918-config-data" (OuterVolumeSpecName: "config-data") pod "1e792caf-ac02-4e6c-9841-ed58133f1918" (UID: "1e792caf-ac02-4e6c-9841-ed58133f1918"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:16:59 crc kubenswrapper[4824]: I1006 10:16:59.748160 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e792caf-ac02-4e6c-9841-ed58133f1918-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:59 crc kubenswrapper[4824]: I1006 10:16:59.748371 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vgb9v\" (UniqueName: \"kubernetes.io/projected/1e792caf-ac02-4e6c-9841-ed58133f1918-kube-api-access-vgb9v\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:59 crc kubenswrapper[4824]: I1006 10:16:59.748464 4824 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1e792caf-ac02-4e6c-9841-ed58133f1918-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:59 crc kubenswrapper[4824]: I1006 10:16:59.748516 4824 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1e792caf-ac02-4e6c-9841-ed58133f1918-logs\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:59 crc kubenswrapper[4824]: I1006 10:16:59.748563 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e792caf-ac02-4e6c-9841-ed58133f1918-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:59 crc kubenswrapper[4824]: I1006 10:16:59.797338 4824 generic.go:334] "Generic (PLEG): container finished" podID="1e792caf-ac02-4e6c-9841-ed58133f1918" containerID="717654dbd581395068e7062c4a749a857288b21cb683c2bb83c1b3355e2232ca" exitCode=0 Oct 06 10:16:59 crc kubenswrapper[4824]: I1006 10:16:59.797471 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5f478db59d-8kntd" event={"ID":"1e792caf-ac02-4e6c-9841-ed58133f1918","Type":"ContainerDied","Data":"717654dbd581395068e7062c4a749a857288b21cb683c2bb83c1b3355e2232ca"} Oct 06 10:16:59 crc kubenswrapper[4824]: I1006 10:16:59.797518 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5f478db59d-8kntd" event={"ID":"1e792caf-ac02-4e6c-9841-ed58133f1918","Type":"ContainerDied","Data":"48a22e6159cff9549064e391a6121b0e4b88133f150e41f49a193cc664a59156"} Oct 06 10:16:59 crc kubenswrapper[4824]: I1006 10:16:59.797548 4824 scope.go:117] "RemoveContainer" containerID="717654dbd581395068e7062c4a749a857288b21cb683c2bb83c1b3355e2232ca" Oct 06 10:16:59 crc kubenswrapper[4824]: I1006 10:16:59.798082 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5f478db59d-8kntd" Oct 06 10:16:59 crc kubenswrapper[4824]: I1006 10:16:59.807169 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"18194590-7963-41e8-9253-dcf486d2cea4","Type":"ContainerStarted","Data":"5c0738d12a8cfb250bbad9ffbd244a4632ca5ee5932e523e043dc103d7fe5e41"} Oct 06 10:16:59 crc kubenswrapper[4824]: I1006 10:16:59.846684 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.846654119 podStartE2EDuration="3.846654119s" podCreationTimestamp="2025-10-06 10:16:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:16:59.837954244 +0000 UTC m=+1129.202377095" watchObservedRunningTime="2025-10-06 10:16:59.846654119 +0000 UTC m=+1129.211076980" Oct 06 10:16:59 crc kubenswrapper[4824]: I1006 10:16:59.916046 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-5f478db59d-8kntd"] Oct 06 10:16:59 crc kubenswrapper[4824]: I1006 10:16:59.933259 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-5f478db59d-8kntd"] Oct 06 10:16:59 crc kubenswrapper[4824]: I1006 10:16:59.942672 4824 scope.go:117] "RemoveContainer" containerID="38b1e0c9677bb96ef4d1872779ef930ca0e79ec2b31786d8860e86c3613db462" Oct 06 10:16:59 crc kubenswrapper[4824]: I1006 10:16:59.966419 4824 scope.go:117] "RemoveContainer" containerID="717654dbd581395068e7062c4a749a857288b21cb683c2bb83c1b3355e2232ca" Oct 06 10:16:59 crc kubenswrapper[4824]: E1006 10:16:59.966992 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"717654dbd581395068e7062c4a749a857288b21cb683c2bb83c1b3355e2232ca\": container with ID starting with 717654dbd581395068e7062c4a749a857288b21cb683c2bb83c1b3355e2232ca not found: ID does not exist" containerID="717654dbd581395068e7062c4a749a857288b21cb683c2bb83c1b3355e2232ca" Oct 06 10:16:59 crc kubenswrapper[4824]: I1006 10:16:59.967043 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"717654dbd581395068e7062c4a749a857288b21cb683c2bb83c1b3355e2232ca"} err="failed to get container status \"717654dbd581395068e7062c4a749a857288b21cb683c2bb83c1b3355e2232ca\": rpc error: code = NotFound desc = could not find container \"717654dbd581395068e7062c4a749a857288b21cb683c2bb83c1b3355e2232ca\": container with ID starting with 717654dbd581395068e7062c4a749a857288b21cb683c2bb83c1b3355e2232ca not found: ID does not exist" Oct 06 10:16:59 crc kubenswrapper[4824]: I1006 10:16:59.967071 4824 scope.go:117] "RemoveContainer" containerID="38b1e0c9677bb96ef4d1872779ef930ca0e79ec2b31786d8860e86c3613db462" Oct 06 10:16:59 crc kubenswrapper[4824]: E1006 10:16:59.967581 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"38b1e0c9677bb96ef4d1872779ef930ca0e79ec2b31786d8860e86c3613db462\": container with ID starting with 38b1e0c9677bb96ef4d1872779ef930ca0e79ec2b31786d8860e86c3613db462 not found: ID does not exist" containerID="38b1e0c9677bb96ef4d1872779ef930ca0e79ec2b31786d8860e86c3613db462" Oct 06 10:16:59 crc kubenswrapper[4824]: I1006 10:16:59.967617 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"38b1e0c9677bb96ef4d1872779ef930ca0e79ec2b31786d8860e86c3613db462"} err="failed to get container status \"38b1e0c9677bb96ef4d1872779ef930ca0e79ec2b31786d8860e86c3613db462\": rpc error: code = NotFound desc = could not find container \"38b1e0c9677bb96ef4d1872779ef930ca0e79ec2b31786d8860e86c3613db462\": container with ID starting with 38b1e0c9677bb96ef4d1872779ef930ca0e79ec2b31786d8860e86c3613db462 not found: ID does not exist" Oct 06 10:17:01 crc kubenswrapper[4824]: I1006 10:17:01.296349 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1e792caf-ac02-4e6c-9841-ed58133f1918" path="/var/lib/kubelet/pods/1e792caf-ac02-4e6c-9841-ed58133f1918/volumes" Oct 06 10:17:01 crc kubenswrapper[4824]: I1006 10:17:01.328663 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Oct 06 10:17:01 crc kubenswrapper[4824]: I1006 10:17:01.970174 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-5f446cffd4-v2pxw" Oct 06 10:17:01 crc kubenswrapper[4824]: I1006 10:17:01.971404 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-5f446cffd4-v2pxw" Oct 06 10:17:02 crc kubenswrapper[4824]: I1006 10:17:02.238959 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Oct 06 10:17:02 crc kubenswrapper[4824]: I1006 10:17:02.683179 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-fbf8cb574-lrj2t" Oct 06 10:17:03 crc kubenswrapper[4824]: I1006 10:17:03.868362 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Oct 06 10:17:03 crc kubenswrapper[4824]: E1006 10:17:03.868713 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e792caf-ac02-4e6c-9841-ed58133f1918" containerName="barbican-api" Oct 06 10:17:03 crc kubenswrapper[4824]: I1006 10:17:03.868724 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e792caf-ac02-4e6c-9841-ed58133f1918" containerName="barbican-api" Oct 06 10:17:03 crc kubenswrapper[4824]: E1006 10:17:03.868744 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e792caf-ac02-4e6c-9841-ed58133f1918" containerName="barbican-api-log" Oct 06 10:17:03 crc kubenswrapper[4824]: I1006 10:17:03.868750 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e792caf-ac02-4e6c-9841-ed58133f1918" containerName="barbican-api-log" Oct 06 10:17:03 crc kubenswrapper[4824]: I1006 10:17:03.868910 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="1e792caf-ac02-4e6c-9841-ed58133f1918" containerName="barbican-api-log" Oct 06 10:17:03 crc kubenswrapper[4824]: I1006 10:17:03.868927 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="1e792caf-ac02-4e6c-9841-ed58133f1918" containerName="barbican-api" Oct 06 10:17:03 crc kubenswrapper[4824]: I1006 10:17:03.872741 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 06 10:17:03 crc kubenswrapper[4824]: I1006 10:17:03.877248 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-ghn27" Oct 06 10:17:03 crc kubenswrapper[4824]: I1006 10:17:03.877283 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Oct 06 10:17:03 crc kubenswrapper[4824]: I1006 10:17:03.886169 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Oct 06 10:17:03 crc kubenswrapper[4824]: I1006 10:17:03.890299 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 06 10:17:03 crc kubenswrapper[4824]: I1006 10:17:03.939670 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2df776b-a96a-4fc3-baaf-a888f8d947a2-combined-ca-bundle\") pod \"openstackclient\" (UID: \"b2df776b-a96a-4fc3-baaf-a888f8d947a2\") " pod="openstack/openstackclient" Oct 06 10:17:03 crc kubenswrapper[4824]: I1006 10:17:03.939775 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-92lr6\" (UniqueName: \"kubernetes.io/projected/b2df776b-a96a-4fc3-baaf-a888f8d947a2-kube-api-access-92lr6\") pod \"openstackclient\" (UID: \"b2df776b-a96a-4fc3-baaf-a888f8d947a2\") " pod="openstack/openstackclient" Oct 06 10:17:03 crc kubenswrapper[4824]: I1006 10:17:03.940019 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/b2df776b-a96a-4fc3-baaf-a888f8d947a2-openstack-config-secret\") pod \"openstackclient\" (UID: \"b2df776b-a96a-4fc3-baaf-a888f8d947a2\") " pod="openstack/openstackclient" Oct 06 10:17:03 crc kubenswrapper[4824]: I1006 10:17:03.940138 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/b2df776b-a96a-4fc3-baaf-a888f8d947a2-openstack-config\") pod \"openstackclient\" (UID: \"b2df776b-a96a-4fc3-baaf-a888f8d947a2\") " pod="openstack/openstackclient" Oct 06 10:17:04 crc kubenswrapper[4824]: I1006 10:17:04.041929 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/b2df776b-a96a-4fc3-baaf-a888f8d947a2-openstack-config\") pod \"openstackclient\" (UID: \"b2df776b-a96a-4fc3-baaf-a888f8d947a2\") " pod="openstack/openstackclient" Oct 06 10:17:04 crc kubenswrapper[4824]: I1006 10:17:04.042153 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-92lr6\" (UniqueName: \"kubernetes.io/projected/b2df776b-a96a-4fc3-baaf-a888f8d947a2-kube-api-access-92lr6\") pod \"openstackclient\" (UID: \"b2df776b-a96a-4fc3-baaf-a888f8d947a2\") " pod="openstack/openstackclient" Oct 06 10:17:04 crc kubenswrapper[4824]: I1006 10:17:04.042195 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2df776b-a96a-4fc3-baaf-a888f8d947a2-combined-ca-bundle\") pod \"openstackclient\" (UID: \"b2df776b-a96a-4fc3-baaf-a888f8d947a2\") " pod="openstack/openstackclient" Oct 06 10:17:04 crc kubenswrapper[4824]: I1006 10:17:04.042336 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/b2df776b-a96a-4fc3-baaf-a888f8d947a2-openstack-config-secret\") pod \"openstackclient\" (UID: \"b2df776b-a96a-4fc3-baaf-a888f8d947a2\") " pod="openstack/openstackclient" Oct 06 10:17:04 crc kubenswrapper[4824]: I1006 10:17:04.043271 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/b2df776b-a96a-4fc3-baaf-a888f8d947a2-openstack-config\") pod \"openstackclient\" (UID: \"b2df776b-a96a-4fc3-baaf-a888f8d947a2\") " pod="openstack/openstackclient" Oct 06 10:17:04 crc kubenswrapper[4824]: I1006 10:17:04.055859 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/b2df776b-a96a-4fc3-baaf-a888f8d947a2-openstack-config-secret\") pod \"openstackclient\" (UID: \"b2df776b-a96a-4fc3-baaf-a888f8d947a2\") " pod="openstack/openstackclient" Oct 06 10:17:04 crc kubenswrapper[4824]: I1006 10:17:04.058483 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2df776b-a96a-4fc3-baaf-a888f8d947a2-combined-ca-bundle\") pod \"openstackclient\" (UID: \"b2df776b-a96a-4fc3-baaf-a888f8d947a2\") " pod="openstack/openstackclient" Oct 06 10:17:04 crc kubenswrapper[4824]: I1006 10:17:04.068307 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-92lr6\" (UniqueName: \"kubernetes.io/projected/b2df776b-a96a-4fc3-baaf-a888f8d947a2-kube-api-access-92lr6\") pod \"openstackclient\" (UID: \"b2df776b-a96a-4fc3-baaf-a888f8d947a2\") " pod="openstack/openstackclient" Oct 06 10:17:04 crc kubenswrapper[4824]: I1006 10:17:04.132227 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Oct 06 10:17:04 crc kubenswrapper[4824]: I1006 10:17:04.133651 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 06 10:17:04 crc kubenswrapper[4824]: I1006 10:17:04.146642 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Oct 06 10:17:04 crc kubenswrapper[4824]: I1006 10:17:04.232593 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Oct 06 10:17:04 crc kubenswrapper[4824]: I1006 10:17:04.234678 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 06 10:17:04 crc kubenswrapper[4824]: I1006 10:17:04.248377 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6cdeb7d5-11b7-4e30-92fd-5c88b658ea92-combined-ca-bundle\") pod \"openstackclient\" (UID: \"6cdeb7d5-11b7-4e30-92fd-5c88b658ea92\") " pod="openstack/openstackclient" Oct 06 10:17:04 crc kubenswrapper[4824]: I1006 10:17:04.248780 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/6cdeb7d5-11b7-4e30-92fd-5c88b658ea92-openstack-config\") pod \"openstackclient\" (UID: \"6cdeb7d5-11b7-4e30-92fd-5c88b658ea92\") " pod="openstack/openstackclient" Oct 06 10:17:04 crc kubenswrapper[4824]: I1006 10:17:04.249039 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2978g\" (UniqueName: \"kubernetes.io/projected/6cdeb7d5-11b7-4e30-92fd-5c88b658ea92-kube-api-access-2978g\") pod \"openstackclient\" (UID: \"6cdeb7d5-11b7-4e30-92fd-5c88b658ea92\") " pod="openstack/openstackclient" Oct 06 10:17:04 crc kubenswrapper[4824]: I1006 10:17:04.249070 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/6cdeb7d5-11b7-4e30-92fd-5c88b658ea92-openstack-config-secret\") pod \"openstackclient\" (UID: \"6cdeb7d5-11b7-4e30-92fd-5c88b658ea92\") " pod="openstack/openstackclient" Oct 06 10:17:04 crc kubenswrapper[4824]: I1006 10:17:04.261328 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 06 10:17:04 crc kubenswrapper[4824]: E1006 10:17:04.328313 4824 log.go:32] "RunPodSandbox from runtime service failed" err=< Oct 06 10:17:04 crc kubenswrapper[4824]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_openstackclient_openstack_b2df776b-a96a-4fc3-baaf-a888f8d947a2_0(e03604d838fd8d5301a18a6ac4c7c53703129ae8520ba7a42c927765a6b6494a): error adding pod openstack_openstackclient to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"e03604d838fd8d5301a18a6ac4c7c53703129ae8520ba7a42c927765a6b6494a" Netns:"/var/run/netns/769744b0-58b6-49f0-b4ad-68383749ef41" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openstack;K8S_POD_NAME=openstackclient;K8S_POD_INFRA_CONTAINER_ID=e03604d838fd8d5301a18a6ac4c7c53703129ae8520ba7a42c927765a6b6494a;K8S_POD_UID=b2df776b-a96a-4fc3-baaf-a888f8d947a2" Path:"" ERRORED: error configuring pod [openstack/openstackclient] networking: Multus: [openstack/openstackclient/b2df776b-a96a-4fc3-baaf-a888f8d947a2]: expected pod UID "b2df776b-a96a-4fc3-baaf-a888f8d947a2" but got "6cdeb7d5-11b7-4e30-92fd-5c88b658ea92" from Kube API Oct 06 10:17:04 crc kubenswrapper[4824]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Oct 06 10:17:04 crc kubenswrapper[4824]: > Oct 06 10:17:04 crc kubenswrapper[4824]: E1006 10:17:04.328398 4824 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err=< Oct 06 10:17:04 crc kubenswrapper[4824]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_openstackclient_openstack_b2df776b-a96a-4fc3-baaf-a888f8d947a2_0(e03604d838fd8d5301a18a6ac4c7c53703129ae8520ba7a42c927765a6b6494a): error adding pod openstack_openstackclient to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"e03604d838fd8d5301a18a6ac4c7c53703129ae8520ba7a42c927765a6b6494a" Netns:"/var/run/netns/769744b0-58b6-49f0-b4ad-68383749ef41" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openstack;K8S_POD_NAME=openstackclient;K8S_POD_INFRA_CONTAINER_ID=e03604d838fd8d5301a18a6ac4c7c53703129ae8520ba7a42c927765a6b6494a;K8S_POD_UID=b2df776b-a96a-4fc3-baaf-a888f8d947a2" Path:"" ERRORED: error configuring pod [openstack/openstackclient] networking: Multus: [openstack/openstackclient/b2df776b-a96a-4fc3-baaf-a888f8d947a2]: expected pod UID "b2df776b-a96a-4fc3-baaf-a888f8d947a2" but got "6cdeb7d5-11b7-4e30-92fd-5c88b658ea92" from Kube API Oct 06 10:17:04 crc kubenswrapper[4824]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Oct 06 10:17:04 crc kubenswrapper[4824]: > pod="openstack/openstackclient" Oct 06 10:17:04 crc kubenswrapper[4824]: I1006 10:17:04.350357 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/6cdeb7d5-11b7-4e30-92fd-5c88b658ea92-openstack-config\") pod \"openstackclient\" (UID: \"6cdeb7d5-11b7-4e30-92fd-5c88b658ea92\") " pod="openstack/openstackclient" Oct 06 10:17:04 crc kubenswrapper[4824]: I1006 10:17:04.350560 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2978g\" (UniqueName: \"kubernetes.io/projected/6cdeb7d5-11b7-4e30-92fd-5c88b658ea92-kube-api-access-2978g\") pod \"openstackclient\" (UID: \"6cdeb7d5-11b7-4e30-92fd-5c88b658ea92\") " pod="openstack/openstackclient" Oct 06 10:17:04 crc kubenswrapper[4824]: I1006 10:17:04.350598 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/6cdeb7d5-11b7-4e30-92fd-5c88b658ea92-openstack-config-secret\") pod \"openstackclient\" (UID: \"6cdeb7d5-11b7-4e30-92fd-5c88b658ea92\") " pod="openstack/openstackclient" Oct 06 10:17:04 crc kubenswrapper[4824]: I1006 10:17:04.350636 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6cdeb7d5-11b7-4e30-92fd-5c88b658ea92-combined-ca-bundle\") pod \"openstackclient\" (UID: \"6cdeb7d5-11b7-4e30-92fd-5c88b658ea92\") " pod="openstack/openstackclient" Oct 06 10:17:04 crc kubenswrapper[4824]: I1006 10:17:04.354427 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/6cdeb7d5-11b7-4e30-92fd-5c88b658ea92-openstack-config\") pod \"openstackclient\" (UID: \"6cdeb7d5-11b7-4e30-92fd-5c88b658ea92\") " pod="openstack/openstackclient" Oct 06 10:17:04 crc kubenswrapper[4824]: I1006 10:17:04.358605 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/6cdeb7d5-11b7-4e30-92fd-5c88b658ea92-openstack-config-secret\") pod \"openstackclient\" (UID: \"6cdeb7d5-11b7-4e30-92fd-5c88b658ea92\") " pod="openstack/openstackclient" Oct 06 10:17:04 crc kubenswrapper[4824]: I1006 10:17:04.360735 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6cdeb7d5-11b7-4e30-92fd-5c88b658ea92-combined-ca-bundle\") pod \"openstackclient\" (UID: \"6cdeb7d5-11b7-4e30-92fd-5c88b658ea92\") " pod="openstack/openstackclient" Oct 06 10:17:04 crc kubenswrapper[4824]: I1006 10:17:04.374801 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2978g\" (UniqueName: \"kubernetes.io/projected/6cdeb7d5-11b7-4e30-92fd-5c88b658ea92-kube-api-access-2978g\") pod \"openstackclient\" (UID: \"6cdeb7d5-11b7-4e30-92fd-5c88b658ea92\") " pod="openstack/openstackclient" Oct 06 10:17:04 crc kubenswrapper[4824]: I1006 10:17:04.642199 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 06 10:17:04 crc kubenswrapper[4824]: I1006 10:17:04.874224 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 06 10:17:04 crc kubenswrapper[4824]: I1006 10:17:04.879052 4824 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="b2df776b-a96a-4fc3-baaf-a888f8d947a2" podUID="6cdeb7d5-11b7-4e30-92fd-5c88b658ea92" Oct 06 10:17:04 crc kubenswrapper[4824]: I1006 10:17:04.889304 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 06 10:17:04 crc kubenswrapper[4824]: I1006 10:17:04.963379 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2df776b-a96a-4fc3-baaf-a888f8d947a2-combined-ca-bundle\") pod \"b2df776b-a96a-4fc3-baaf-a888f8d947a2\" (UID: \"b2df776b-a96a-4fc3-baaf-a888f8d947a2\") " Oct 06 10:17:04 crc kubenswrapper[4824]: I1006 10:17:04.963483 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-92lr6\" (UniqueName: \"kubernetes.io/projected/b2df776b-a96a-4fc3-baaf-a888f8d947a2-kube-api-access-92lr6\") pod \"b2df776b-a96a-4fc3-baaf-a888f8d947a2\" (UID: \"b2df776b-a96a-4fc3-baaf-a888f8d947a2\") " Oct 06 10:17:04 crc kubenswrapper[4824]: I1006 10:17:04.963741 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/b2df776b-a96a-4fc3-baaf-a888f8d947a2-openstack-config\") pod \"b2df776b-a96a-4fc3-baaf-a888f8d947a2\" (UID: \"b2df776b-a96a-4fc3-baaf-a888f8d947a2\") " Oct 06 10:17:04 crc kubenswrapper[4824]: I1006 10:17:04.964791 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b2df776b-a96a-4fc3-baaf-a888f8d947a2-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "b2df776b-a96a-4fc3-baaf-a888f8d947a2" (UID: "b2df776b-a96a-4fc3-baaf-a888f8d947a2"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:17:04 crc kubenswrapper[4824]: I1006 10:17:04.964822 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/b2df776b-a96a-4fc3-baaf-a888f8d947a2-openstack-config-secret\") pod \"b2df776b-a96a-4fc3-baaf-a888f8d947a2\" (UID: \"b2df776b-a96a-4fc3-baaf-a888f8d947a2\") " Oct 06 10:17:04 crc kubenswrapper[4824]: I1006 10:17:04.965517 4824 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/b2df776b-a96a-4fc3-baaf-a888f8d947a2-openstack-config\") on node \"crc\" DevicePath \"\"" Oct 06 10:17:04 crc kubenswrapper[4824]: I1006 10:17:04.970890 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b2df776b-a96a-4fc3-baaf-a888f8d947a2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b2df776b-a96a-4fc3-baaf-a888f8d947a2" (UID: "b2df776b-a96a-4fc3-baaf-a888f8d947a2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:17:04 crc kubenswrapper[4824]: I1006 10:17:04.971057 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b2df776b-a96a-4fc3-baaf-a888f8d947a2-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "b2df776b-a96a-4fc3-baaf-a888f8d947a2" (UID: "b2df776b-a96a-4fc3-baaf-a888f8d947a2"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:17:04 crc kubenswrapper[4824]: I1006 10:17:04.976573 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b2df776b-a96a-4fc3-baaf-a888f8d947a2-kube-api-access-92lr6" (OuterVolumeSpecName: "kube-api-access-92lr6") pod "b2df776b-a96a-4fc3-baaf-a888f8d947a2" (UID: "b2df776b-a96a-4fc3-baaf-a888f8d947a2"). InnerVolumeSpecName "kube-api-access-92lr6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:17:05 crc kubenswrapper[4824]: I1006 10:17:05.066929 4824 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/b2df776b-a96a-4fc3-baaf-a888f8d947a2-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Oct 06 10:17:05 crc kubenswrapper[4824]: I1006 10:17:05.066969 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2df776b-a96a-4fc3-baaf-a888f8d947a2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 10:17:05 crc kubenswrapper[4824]: I1006 10:17:05.066993 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-92lr6\" (UniqueName: \"kubernetes.io/projected/b2df776b-a96a-4fc3-baaf-a888f8d947a2-kube-api-access-92lr6\") on node \"crc\" DevicePath \"\"" Oct 06 10:17:05 crc kubenswrapper[4824]: I1006 10:17:05.134055 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 06 10:17:05 crc kubenswrapper[4824]: I1006 10:17:05.289129 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b2df776b-a96a-4fc3-baaf-a888f8d947a2" path="/var/lib/kubelet/pods/b2df776b-a96a-4fc3-baaf-a888f8d947a2/volumes" Oct 06 10:17:05 crc kubenswrapper[4824]: I1006 10:17:05.884641 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"6cdeb7d5-11b7-4e30-92fd-5c88b658ea92","Type":"ContainerStarted","Data":"c6d8f063ab720f7c2fbbf0cce4c1e1fe74722841bf3a0d395edb5cdd3be16e47"} Oct 06 10:17:05 crc kubenswrapper[4824]: I1006 10:17:05.884678 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 06 10:17:05 crc kubenswrapper[4824]: I1006 10:17:05.891353 4824 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="b2df776b-a96a-4fc3-baaf-a888f8d947a2" podUID="6cdeb7d5-11b7-4e30-92fd-5c88b658ea92" Oct 06 10:17:06 crc kubenswrapper[4824]: I1006 10:17:06.233147 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-78c847f9f4-tv8zc" podUID="68ef92a4-c4bc-4acc-be58-e14010395b8e" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.149:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.149:8443: connect: connection refused" Oct 06 10:17:07 crc kubenswrapper[4824]: I1006 10:17:07.623852 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Oct 06 10:17:08 crc kubenswrapper[4824]: I1006 10:17:08.974688 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 06 10:17:08 crc kubenswrapper[4824]: I1006 10:17:08.976760 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="54d572d5-25f5-4531-bfc0-5a75e1a90cc6" containerName="ceilometer-central-agent" containerID="cri-o://1bff648dcf7b79ca89921e2e518b5cb358f03fa3f398af9219f6fb069693118b" gracePeriod=30 Oct 06 10:17:08 crc kubenswrapper[4824]: I1006 10:17:08.976847 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="54d572d5-25f5-4531-bfc0-5a75e1a90cc6" containerName="proxy-httpd" containerID="cri-o://0300131a20e09d3243c9587d3d719b3b8e6fdbb7eb8b5fd65050788326cf67ff" gracePeriod=30 Oct 06 10:17:08 crc kubenswrapper[4824]: I1006 10:17:08.976931 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="54d572d5-25f5-4531-bfc0-5a75e1a90cc6" containerName="ceilometer-notification-agent" containerID="cri-o://6a35262fed9eb98e214f87a986dd0b40d5221f00dfeef9089eabbd903577021d" gracePeriod=30 Oct 06 10:17:08 crc kubenswrapper[4824]: I1006 10:17:08.976946 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="54d572d5-25f5-4531-bfc0-5a75e1a90cc6" containerName="sg-core" containerID="cri-o://465d1fb4fc5bd81e1a6167185933ebfaf1a5fa5bfe19dccc85f947a72337e87f" gracePeriod=30 Oct 06 10:17:08 crc kubenswrapper[4824]: I1006 10:17:08.982921 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="54d572d5-25f5-4531-bfc0-5a75e1a90cc6" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.167:3000/\": read tcp 10.217.0.2:55434->10.217.0.167:3000: read: connection reset by peer" Oct 06 10:17:09 crc kubenswrapper[4824]: I1006 10:17:09.204497 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-78d766d697-s7wzp"] Oct 06 10:17:09 crc kubenswrapper[4824]: I1006 10:17:09.207121 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-78d766d697-s7wzp" Oct 06 10:17:09 crc kubenswrapper[4824]: I1006 10:17:09.209822 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Oct 06 10:17:09 crc kubenswrapper[4824]: I1006 10:17:09.209960 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Oct 06 10:17:09 crc kubenswrapper[4824]: I1006 10:17:09.210865 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Oct 06 10:17:09 crc kubenswrapper[4824]: I1006 10:17:09.227463 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-78d766d697-s7wzp"] Oct 06 10:17:09 crc kubenswrapper[4824]: I1006 10:17:09.278050 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/925ed890-0e01-4422-a2f8-4871bf90087d-internal-tls-certs\") pod \"swift-proxy-78d766d697-s7wzp\" (UID: \"925ed890-0e01-4422-a2f8-4871bf90087d\") " pod="openstack/swift-proxy-78d766d697-s7wzp" Oct 06 10:17:09 crc kubenswrapper[4824]: I1006 10:17:09.278176 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/925ed890-0e01-4422-a2f8-4871bf90087d-combined-ca-bundle\") pod \"swift-proxy-78d766d697-s7wzp\" (UID: \"925ed890-0e01-4422-a2f8-4871bf90087d\") " pod="openstack/swift-proxy-78d766d697-s7wzp" Oct 06 10:17:09 crc kubenswrapper[4824]: I1006 10:17:09.278204 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/925ed890-0e01-4422-a2f8-4871bf90087d-log-httpd\") pod \"swift-proxy-78d766d697-s7wzp\" (UID: \"925ed890-0e01-4422-a2f8-4871bf90087d\") " pod="openstack/swift-proxy-78d766d697-s7wzp" Oct 06 10:17:09 crc kubenswrapper[4824]: I1006 10:17:09.278246 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/925ed890-0e01-4422-a2f8-4871bf90087d-config-data\") pod \"swift-proxy-78d766d697-s7wzp\" (UID: \"925ed890-0e01-4422-a2f8-4871bf90087d\") " pod="openstack/swift-proxy-78d766d697-s7wzp" Oct 06 10:17:09 crc kubenswrapper[4824]: I1006 10:17:09.278278 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/925ed890-0e01-4422-a2f8-4871bf90087d-etc-swift\") pod \"swift-proxy-78d766d697-s7wzp\" (UID: \"925ed890-0e01-4422-a2f8-4871bf90087d\") " pod="openstack/swift-proxy-78d766d697-s7wzp" Oct 06 10:17:09 crc kubenswrapper[4824]: I1006 10:17:09.278302 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/925ed890-0e01-4422-a2f8-4871bf90087d-public-tls-certs\") pod \"swift-proxy-78d766d697-s7wzp\" (UID: \"925ed890-0e01-4422-a2f8-4871bf90087d\") " pod="openstack/swift-proxy-78d766d697-s7wzp" Oct 06 10:17:09 crc kubenswrapper[4824]: I1006 10:17:09.278319 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jx5zd\" (UniqueName: \"kubernetes.io/projected/925ed890-0e01-4422-a2f8-4871bf90087d-kube-api-access-jx5zd\") pod \"swift-proxy-78d766d697-s7wzp\" (UID: \"925ed890-0e01-4422-a2f8-4871bf90087d\") " pod="openstack/swift-proxy-78d766d697-s7wzp" Oct 06 10:17:09 crc kubenswrapper[4824]: I1006 10:17:09.278363 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/925ed890-0e01-4422-a2f8-4871bf90087d-run-httpd\") pod \"swift-proxy-78d766d697-s7wzp\" (UID: \"925ed890-0e01-4422-a2f8-4871bf90087d\") " pod="openstack/swift-proxy-78d766d697-s7wzp" Oct 06 10:17:09 crc kubenswrapper[4824]: I1006 10:17:09.379369 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/925ed890-0e01-4422-a2f8-4871bf90087d-combined-ca-bundle\") pod \"swift-proxy-78d766d697-s7wzp\" (UID: \"925ed890-0e01-4422-a2f8-4871bf90087d\") " pod="openstack/swift-proxy-78d766d697-s7wzp" Oct 06 10:17:09 crc kubenswrapper[4824]: I1006 10:17:09.379416 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/925ed890-0e01-4422-a2f8-4871bf90087d-log-httpd\") pod \"swift-proxy-78d766d697-s7wzp\" (UID: \"925ed890-0e01-4422-a2f8-4871bf90087d\") " pod="openstack/swift-proxy-78d766d697-s7wzp" Oct 06 10:17:09 crc kubenswrapper[4824]: I1006 10:17:09.379463 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/925ed890-0e01-4422-a2f8-4871bf90087d-config-data\") pod \"swift-proxy-78d766d697-s7wzp\" (UID: \"925ed890-0e01-4422-a2f8-4871bf90087d\") " pod="openstack/swift-proxy-78d766d697-s7wzp" Oct 06 10:17:09 crc kubenswrapper[4824]: I1006 10:17:09.379498 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/925ed890-0e01-4422-a2f8-4871bf90087d-etc-swift\") pod \"swift-proxy-78d766d697-s7wzp\" (UID: \"925ed890-0e01-4422-a2f8-4871bf90087d\") " pod="openstack/swift-proxy-78d766d697-s7wzp" Oct 06 10:17:09 crc kubenswrapper[4824]: I1006 10:17:09.379544 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/925ed890-0e01-4422-a2f8-4871bf90087d-public-tls-certs\") pod \"swift-proxy-78d766d697-s7wzp\" (UID: \"925ed890-0e01-4422-a2f8-4871bf90087d\") " pod="openstack/swift-proxy-78d766d697-s7wzp" Oct 06 10:17:09 crc kubenswrapper[4824]: I1006 10:17:09.379563 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jx5zd\" (UniqueName: \"kubernetes.io/projected/925ed890-0e01-4422-a2f8-4871bf90087d-kube-api-access-jx5zd\") pod \"swift-proxy-78d766d697-s7wzp\" (UID: \"925ed890-0e01-4422-a2f8-4871bf90087d\") " pod="openstack/swift-proxy-78d766d697-s7wzp" Oct 06 10:17:09 crc kubenswrapper[4824]: I1006 10:17:09.379753 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/925ed890-0e01-4422-a2f8-4871bf90087d-run-httpd\") pod \"swift-proxy-78d766d697-s7wzp\" (UID: \"925ed890-0e01-4422-a2f8-4871bf90087d\") " pod="openstack/swift-proxy-78d766d697-s7wzp" Oct 06 10:17:09 crc kubenswrapper[4824]: I1006 10:17:09.379780 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/925ed890-0e01-4422-a2f8-4871bf90087d-internal-tls-certs\") pod \"swift-proxy-78d766d697-s7wzp\" (UID: \"925ed890-0e01-4422-a2f8-4871bf90087d\") " pod="openstack/swift-proxy-78d766d697-s7wzp" Oct 06 10:17:09 crc kubenswrapper[4824]: I1006 10:17:09.382656 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/925ed890-0e01-4422-a2f8-4871bf90087d-log-httpd\") pod \"swift-proxy-78d766d697-s7wzp\" (UID: \"925ed890-0e01-4422-a2f8-4871bf90087d\") " pod="openstack/swift-proxy-78d766d697-s7wzp" Oct 06 10:17:09 crc kubenswrapper[4824]: I1006 10:17:09.387897 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/925ed890-0e01-4422-a2f8-4871bf90087d-run-httpd\") pod \"swift-proxy-78d766d697-s7wzp\" (UID: \"925ed890-0e01-4422-a2f8-4871bf90087d\") " pod="openstack/swift-proxy-78d766d697-s7wzp" Oct 06 10:17:09 crc kubenswrapper[4824]: I1006 10:17:09.388059 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/925ed890-0e01-4422-a2f8-4871bf90087d-public-tls-certs\") pod \"swift-proxy-78d766d697-s7wzp\" (UID: \"925ed890-0e01-4422-a2f8-4871bf90087d\") " pod="openstack/swift-proxy-78d766d697-s7wzp" Oct 06 10:17:09 crc kubenswrapper[4824]: I1006 10:17:09.390806 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/925ed890-0e01-4422-a2f8-4871bf90087d-etc-swift\") pod \"swift-proxy-78d766d697-s7wzp\" (UID: \"925ed890-0e01-4422-a2f8-4871bf90087d\") " pod="openstack/swift-proxy-78d766d697-s7wzp" Oct 06 10:17:09 crc kubenswrapper[4824]: I1006 10:17:09.391199 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/925ed890-0e01-4422-a2f8-4871bf90087d-internal-tls-certs\") pod \"swift-proxy-78d766d697-s7wzp\" (UID: \"925ed890-0e01-4422-a2f8-4871bf90087d\") " pod="openstack/swift-proxy-78d766d697-s7wzp" Oct 06 10:17:09 crc kubenswrapper[4824]: I1006 10:17:09.396821 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/925ed890-0e01-4422-a2f8-4871bf90087d-combined-ca-bundle\") pod \"swift-proxy-78d766d697-s7wzp\" (UID: \"925ed890-0e01-4422-a2f8-4871bf90087d\") " pod="openstack/swift-proxy-78d766d697-s7wzp" Oct 06 10:17:09 crc kubenswrapper[4824]: I1006 10:17:09.397840 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/925ed890-0e01-4422-a2f8-4871bf90087d-config-data\") pod \"swift-proxy-78d766d697-s7wzp\" (UID: \"925ed890-0e01-4422-a2f8-4871bf90087d\") " pod="openstack/swift-proxy-78d766d697-s7wzp" Oct 06 10:17:09 crc kubenswrapper[4824]: I1006 10:17:09.399646 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jx5zd\" (UniqueName: \"kubernetes.io/projected/925ed890-0e01-4422-a2f8-4871bf90087d-kube-api-access-jx5zd\") pod \"swift-proxy-78d766d697-s7wzp\" (UID: \"925ed890-0e01-4422-a2f8-4871bf90087d\") " pod="openstack/swift-proxy-78d766d697-s7wzp" Oct 06 10:17:09 crc kubenswrapper[4824]: I1006 10:17:09.552720 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-78d766d697-s7wzp" Oct 06 10:17:09 crc kubenswrapper[4824]: I1006 10:17:09.942944 4824 generic.go:334] "Generic (PLEG): container finished" podID="54d572d5-25f5-4531-bfc0-5a75e1a90cc6" containerID="0300131a20e09d3243c9587d3d719b3b8e6fdbb7eb8b5fd65050788326cf67ff" exitCode=0 Oct 06 10:17:09 crc kubenswrapper[4824]: I1006 10:17:09.943002 4824 generic.go:334] "Generic (PLEG): container finished" podID="54d572d5-25f5-4531-bfc0-5a75e1a90cc6" containerID="465d1fb4fc5bd81e1a6167185933ebfaf1a5fa5bfe19dccc85f947a72337e87f" exitCode=2 Oct 06 10:17:09 crc kubenswrapper[4824]: I1006 10:17:09.943011 4824 generic.go:334] "Generic (PLEG): container finished" podID="54d572d5-25f5-4531-bfc0-5a75e1a90cc6" containerID="1bff648dcf7b79ca89921e2e518b5cb358f03fa3f398af9219f6fb069693118b" exitCode=0 Oct 06 10:17:09 crc kubenswrapper[4824]: I1006 10:17:09.943031 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"54d572d5-25f5-4531-bfc0-5a75e1a90cc6","Type":"ContainerDied","Data":"0300131a20e09d3243c9587d3d719b3b8e6fdbb7eb8b5fd65050788326cf67ff"} Oct 06 10:17:09 crc kubenswrapper[4824]: I1006 10:17:09.943058 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"54d572d5-25f5-4531-bfc0-5a75e1a90cc6","Type":"ContainerDied","Data":"465d1fb4fc5bd81e1a6167185933ebfaf1a5fa5bfe19dccc85f947a72337e87f"} Oct 06 10:17:09 crc kubenswrapper[4824]: I1006 10:17:09.943069 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"54d572d5-25f5-4531-bfc0-5a75e1a90cc6","Type":"ContainerDied","Data":"1bff648dcf7b79ca89921e2e518b5cb358f03fa3f398af9219f6fb069693118b"} Oct 06 10:17:10 crc kubenswrapper[4824]: I1006 10:17:10.700557 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-lgz2q"] Oct 06 10:17:10 crc kubenswrapper[4824]: I1006 10:17:10.702022 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-lgz2q" Oct 06 10:17:10 crc kubenswrapper[4824]: I1006 10:17:10.733354 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-lgz2q"] Oct 06 10:17:10 crc kubenswrapper[4824]: I1006 10:17:10.794569 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-qwq7x"] Oct 06 10:17:10 crc kubenswrapper[4824]: I1006 10:17:10.795845 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-qwq7x" Oct 06 10:17:10 crc kubenswrapper[4824]: I1006 10:17:10.811502 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wwcsv\" (UniqueName: \"kubernetes.io/projected/547f97c5-ea4d-4fbc-9f6a-d21ebd33d58f-kube-api-access-wwcsv\") pod \"nova-api-db-create-lgz2q\" (UID: \"547f97c5-ea4d-4fbc-9f6a-d21ebd33d58f\") " pod="openstack/nova-api-db-create-lgz2q" Oct 06 10:17:10 crc kubenswrapper[4824]: I1006 10:17:10.817585 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-qwq7x"] Oct 06 10:17:10 crc kubenswrapper[4824]: I1006 10:17:10.881179 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-48qqz"] Oct 06 10:17:10 crc kubenswrapper[4824]: I1006 10:17:10.882596 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-48qqz" Oct 06 10:17:10 crc kubenswrapper[4824]: I1006 10:17:10.911434 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-48qqz"] Oct 06 10:17:10 crc kubenswrapper[4824]: I1006 10:17:10.913501 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-48v6b\" (UniqueName: \"kubernetes.io/projected/c492c641-1fe6-44d1-8b6b-0933efbef65d-kube-api-access-48v6b\") pod \"nova-cell0-db-create-qwq7x\" (UID: \"c492c641-1fe6-44d1-8b6b-0933efbef65d\") " pod="openstack/nova-cell0-db-create-qwq7x" Oct 06 10:17:10 crc kubenswrapper[4824]: I1006 10:17:10.913592 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wwcsv\" (UniqueName: \"kubernetes.io/projected/547f97c5-ea4d-4fbc-9f6a-d21ebd33d58f-kube-api-access-wwcsv\") pod \"nova-api-db-create-lgz2q\" (UID: \"547f97c5-ea4d-4fbc-9f6a-d21ebd33d58f\") " pod="openstack/nova-api-db-create-lgz2q" Oct 06 10:17:10 crc kubenswrapper[4824]: I1006 10:17:10.932700 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wwcsv\" (UniqueName: \"kubernetes.io/projected/547f97c5-ea4d-4fbc-9f6a-d21ebd33d58f-kube-api-access-wwcsv\") pod \"nova-api-db-create-lgz2q\" (UID: \"547f97c5-ea4d-4fbc-9f6a-d21ebd33d58f\") " pod="openstack/nova-api-db-create-lgz2q" Oct 06 10:17:11 crc kubenswrapper[4824]: I1006 10:17:11.016067 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ndsjt\" (UniqueName: \"kubernetes.io/projected/6d4717f3-9382-4a1b-a6a3-d0f0f37880c7-kube-api-access-ndsjt\") pod \"nova-cell1-db-create-48qqz\" (UID: \"6d4717f3-9382-4a1b-a6a3-d0f0f37880c7\") " pod="openstack/nova-cell1-db-create-48qqz" Oct 06 10:17:11 crc kubenswrapper[4824]: I1006 10:17:11.016481 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-48v6b\" (UniqueName: \"kubernetes.io/projected/c492c641-1fe6-44d1-8b6b-0933efbef65d-kube-api-access-48v6b\") pod \"nova-cell0-db-create-qwq7x\" (UID: \"c492c641-1fe6-44d1-8b6b-0933efbef65d\") " pod="openstack/nova-cell0-db-create-qwq7x" Oct 06 10:17:11 crc kubenswrapper[4824]: I1006 10:17:11.033312 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-lgz2q" Oct 06 10:17:11 crc kubenswrapper[4824]: I1006 10:17:11.048718 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-48v6b\" (UniqueName: \"kubernetes.io/projected/c492c641-1fe6-44d1-8b6b-0933efbef65d-kube-api-access-48v6b\") pod \"nova-cell0-db-create-qwq7x\" (UID: \"c492c641-1fe6-44d1-8b6b-0933efbef65d\") " pod="openstack/nova-cell0-db-create-qwq7x" Oct 06 10:17:11 crc kubenswrapper[4824]: I1006 10:17:11.114075 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-qwq7x" Oct 06 10:17:11 crc kubenswrapper[4824]: I1006 10:17:11.118766 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ndsjt\" (UniqueName: \"kubernetes.io/projected/6d4717f3-9382-4a1b-a6a3-d0f0f37880c7-kube-api-access-ndsjt\") pod \"nova-cell1-db-create-48qqz\" (UID: \"6d4717f3-9382-4a1b-a6a3-d0f0f37880c7\") " pod="openstack/nova-cell1-db-create-48qqz" Oct 06 10:17:11 crc kubenswrapper[4824]: I1006 10:17:11.138502 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ndsjt\" (UniqueName: \"kubernetes.io/projected/6d4717f3-9382-4a1b-a6a3-d0f0f37880c7-kube-api-access-ndsjt\") pod \"nova-cell1-db-create-48qqz\" (UID: \"6d4717f3-9382-4a1b-a6a3-d0f0f37880c7\") " pod="openstack/nova-cell1-db-create-48qqz" Oct 06 10:17:11 crc kubenswrapper[4824]: I1006 10:17:11.207285 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-48qqz" Oct 06 10:17:13 crc kubenswrapper[4824]: I1006 10:17:13.993011 4824 generic.go:334] "Generic (PLEG): container finished" podID="54d572d5-25f5-4531-bfc0-5a75e1a90cc6" containerID="6a35262fed9eb98e214f87a986dd0b40d5221f00dfeef9089eabbd903577021d" exitCode=0 Oct 06 10:17:13 crc kubenswrapper[4824]: I1006 10:17:13.993342 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"54d572d5-25f5-4531-bfc0-5a75e1a90cc6","Type":"ContainerDied","Data":"6a35262fed9eb98e214f87a986dd0b40d5221f00dfeef9089eabbd903577021d"} Oct 06 10:17:15 crc kubenswrapper[4824]: I1006 10:17:15.030679 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 06 10:17:15 crc kubenswrapper[4824]: I1006 10:17:15.044823 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"54d572d5-25f5-4531-bfc0-5a75e1a90cc6","Type":"ContainerDied","Data":"ab72b288fa642b607e8a99fd684161f7670f804ca0215c1aa53da21e02ef7d9b"} Oct 06 10:17:15 crc kubenswrapper[4824]: I1006 10:17:15.049868 4824 scope.go:117] "RemoveContainer" containerID="0300131a20e09d3243c9587d3d719b3b8e6fdbb7eb8b5fd65050788326cf67ff" Oct 06 10:17:15 crc kubenswrapper[4824]: I1006 10:17:15.090372 4824 scope.go:117] "RemoveContainer" containerID="465d1fb4fc5bd81e1a6167185933ebfaf1a5fa5bfe19dccc85f947a72337e87f" Oct 06 10:17:15 crc kubenswrapper[4824]: I1006 10:17:15.116151 4824 scope.go:117] "RemoveContainer" containerID="6a35262fed9eb98e214f87a986dd0b40d5221f00dfeef9089eabbd903577021d" Oct 06 10:17:15 crc kubenswrapper[4824]: I1006 10:17:15.140838 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54d572d5-25f5-4531-bfc0-5a75e1a90cc6-combined-ca-bundle\") pod \"54d572d5-25f5-4531-bfc0-5a75e1a90cc6\" (UID: \"54d572d5-25f5-4531-bfc0-5a75e1a90cc6\") " Oct 06 10:17:15 crc kubenswrapper[4824]: I1006 10:17:15.140952 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/54d572d5-25f5-4531-bfc0-5a75e1a90cc6-config-data\") pod \"54d572d5-25f5-4531-bfc0-5a75e1a90cc6\" (UID: \"54d572d5-25f5-4531-bfc0-5a75e1a90cc6\") " Oct 06 10:17:15 crc kubenswrapper[4824]: I1006 10:17:15.141006 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-29xlz\" (UniqueName: \"kubernetes.io/projected/54d572d5-25f5-4531-bfc0-5a75e1a90cc6-kube-api-access-29xlz\") pod \"54d572d5-25f5-4531-bfc0-5a75e1a90cc6\" (UID: \"54d572d5-25f5-4531-bfc0-5a75e1a90cc6\") " Oct 06 10:17:15 crc kubenswrapper[4824]: I1006 10:17:15.141030 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/54d572d5-25f5-4531-bfc0-5a75e1a90cc6-scripts\") pod \"54d572d5-25f5-4531-bfc0-5a75e1a90cc6\" (UID: \"54d572d5-25f5-4531-bfc0-5a75e1a90cc6\") " Oct 06 10:17:15 crc kubenswrapper[4824]: I1006 10:17:15.142339 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/54d572d5-25f5-4531-bfc0-5a75e1a90cc6-run-httpd\") pod \"54d572d5-25f5-4531-bfc0-5a75e1a90cc6\" (UID: \"54d572d5-25f5-4531-bfc0-5a75e1a90cc6\") " Oct 06 10:17:15 crc kubenswrapper[4824]: I1006 10:17:15.142459 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/54d572d5-25f5-4531-bfc0-5a75e1a90cc6-log-httpd\") pod \"54d572d5-25f5-4531-bfc0-5a75e1a90cc6\" (UID: \"54d572d5-25f5-4531-bfc0-5a75e1a90cc6\") " Oct 06 10:17:15 crc kubenswrapper[4824]: I1006 10:17:15.142503 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/54d572d5-25f5-4531-bfc0-5a75e1a90cc6-sg-core-conf-yaml\") pod \"54d572d5-25f5-4531-bfc0-5a75e1a90cc6\" (UID: \"54d572d5-25f5-4531-bfc0-5a75e1a90cc6\") " Oct 06 10:17:15 crc kubenswrapper[4824]: I1006 10:17:15.143122 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/54d572d5-25f5-4531-bfc0-5a75e1a90cc6-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "54d572d5-25f5-4531-bfc0-5a75e1a90cc6" (UID: "54d572d5-25f5-4531-bfc0-5a75e1a90cc6"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:17:15 crc kubenswrapper[4824]: I1006 10:17:15.143515 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/54d572d5-25f5-4531-bfc0-5a75e1a90cc6-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "54d572d5-25f5-4531-bfc0-5a75e1a90cc6" (UID: "54d572d5-25f5-4531-bfc0-5a75e1a90cc6"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:17:15 crc kubenswrapper[4824]: I1006 10:17:15.148828 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/54d572d5-25f5-4531-bfc0-5a75e1a90cc6-kube-api-access-29xlz" (OuterVolumeSpecName: "kube-api-access-29xlz") pod "54d572d5-25f5-4531-bfc0-5a75e1a90cc6" (UID: "54d572d5-25f5-4531-bfc0-5a75e1a90cc6"). InnerVolumeSpecName "kube-api-access-29xlz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:17:15 crc kubenswrapper[4824]: I1006 10:17:15.150448 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/54d572d5-25f5-4531-bfc0-5a75e1a90cc6-scripts" (OuterVolumeSpecName: "scripts") pod "54d572d5-25f5-4531-bfc0-5a75e1a90cc6" (UID: "54d572d5-25f5-4531-bfc0-5a75e1a90cc6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:17:15 crc kubenswrapper[4824]: I1006 10:17:15.154681 4824 scope.go:117] "RemoveContainer" containerID="1bff648dcf7b79ca89921e2e518b5cb358f03fa3f398af9219f6fb069693118b" Oct 06 10:17:15 crc kubenswrapper[4824]: I1006 10:17:15.194906 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/54d572d5-25f5-4531-bfc0-5a75e1a90cc6-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "54d572d5-25f5-4531-bfc0-5a75e1a90cc6" (UID: "54d572d5-25f5-4531-bfc0-5a75e1a90cc6"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:17:15 crc kubenswrapper[4824]: I1006 10:17:15.246448 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/54d572d5-25f5-4531-bfc0-5a75e1a90cc6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "54d572d5-25f5-4531-bfc0-5a75e1a90cc6" (UID: "54d572d5-25f5-4531-bfc0-5a75e1a90cc6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:17:15 crc kubenswrapper[4824]: I1006 10:17:15.247641 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54d572d5-25f5-4531-bfc0-5a75e1a90cc6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 10:17:15 crc kubenswrapper[4824]: I1006 10:17:15.247717 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-29xlz\" (UniqueName: \"kubernetes.io/projected/54d572d5-25f5-4531-bfc0-5a75e1a90cc6-kube-api-access-29xlz\") on node \"crc\" DevicePath \"\"" Oct 06 10:17:15 crc kubenswrapper[4824]: I1006 10:17:15.247770 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/54d572d5-25f5-4531-bfc0-5a75e1a90cc6-scripts\") on node \"crc\" DevicePath \"\"" Oct 06 10:17:15 crc kubenswrapper[4824]: I1006 10:17:15.247818 4824 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/54d572d5-25f5-4531-bfc0-5a75e1a90cc6-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 06 10:17:15 crc kubenswrapper[4824]: I1006 10:17:15.247889 4824 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/54d572d5-25f5-4531-bfc0-5a75e1a90cc6-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 06 10:17:15 crc kubenswrapper[4824]: I1006 10:17:15.247942 4824 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/54d572d5-25f5-4531-bfc0-5a75e1a90cc6-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 06 10:17:15 crc kubenswrapper[4824]: I1006 10:17:15.275768 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/54d572d5-25f5-4531-bfc0-5a75e1a90cc6-config-data" (OuterVolumeSpecName: "config-data") pod "54d572d5-25f5-4531-bfc0-5a75e1a90cc6" (UID: "54d572d5-25f5-4531-bfc0-5a75e1a90cc6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:17:15 crc kubenswrapper[4824]: I1006 10:17:15.351432 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/54d572d5-25f5-4531-bfc0-5a75e1a90cc6-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 10:17:15 crc kubenswrapper[4824]: I1006 10:17:15.429498 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-lgz2q"] Oct 06 10:17:15 crc kubenswrapper[4824]: I1006 10:17:15.436813 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-48qqz"] Oct 06 10:17:15 crc kubenswrapper[4824]: W1006 10:17:15.441800 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod547f97c5_ea4d_4fbc_9f6a_d21ebd33d58f.slice/crio-9984ef6e6b0767d57c3e9b1b51ab03c2beceb61627fa8a5a70204793e6ab9a52 WatchSource:0}: Error finding container 9984ef6e6b0767d57c3e9b1b51ab03c2beceb61627fa8a5a70204793e6ab9a52: Status 404 returned error can't find the container with id 9984ef6e6b0767d57c3e9b1b51ab03c2beceb61627fa8a5a70204793e6ab9a52 Oct 06 10:17:15 crc kubenswrapper[4824]: I1006 10:17:15.484278 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-78d766d697-s7wzp"] Oct 06 10:17:15 crc kubenswrapper[4824]: I1006 10:17:15.538940 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-qwq7x"] Oct 06 10:17:16 crc kubenswrapper[4824]: I1006 10:17:16.055839 4824 generic.go:334] "Generic (PLEG): container finished" podID="547f97c5-ea4d-4fbc-9f6a-d21ebd33d58f" containerID="0197e3d5f2022700f4e6658e19fec631b535f5202acfa0bf383ab4c2a7407e42" exitCode=0 Oct 06 10:17:16 crc kubenswrapper[4824]: I1006 10:17:16.055909 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-lgz2q" event={"ID":"547f97c5-ea4d-4fbc-9f6a-d21ebd33d58f","Type":"ContainerDied","Data":"0197e3d5f2022700f4e6658e19fec631b535f5202acfa0bf383ab4c2a7407e42"} Oct 06 10:17:16 crc kubenswrapper[4824]: I1006 10:17:16.057207 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-lgz2q" event={"ID":"547f97c5-ea4d-4fbc-9f6a-d21ebd33d58f","Type":"ContainerStarted","Data":"9984ef6e6b0767d57c3e9b1b51ab03c2beceb61627fa8a5a70204793e6ab9a52"} Oct 06 10:17:16 crc kubenswrapper[4824]: I1006 10:17:16.059080 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-78d766d697-s7wzp" event={"ID":"925ed890-0e01-4422-a2f8-4871bf90087d","Type":"ContainerStarted","Data":"f3a3615a457e939da31de5437694415b2998b78158a64cc4bf4e858b546aed2a"} Oct 06 10:17:16 crc kubenswrapper[4824]: I1006 10:17:16.059201 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-78d766d697-s7wzp" event={"ID":"925ed890-0e01-4422-a2f8-4871bf90087d","Type":"ContainerStarted","Data":"47ed02d9b3d907848be8a0959ae4c0ec9c90b8e6425ec5bb11ea535c159ba82b"} Oct 06 10:17:16 crc kubenswrapper[4824]: I1006 10:17:16.060633 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 06 10:17:16 crc kubenswrapper[4824]: I1006 10:17:16.062164 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"6cdeb7d5-11b7-4e30-92fd-5c88b658ea92","Type":"ContainerStarted","Data":"1eb57fdceb7c113c15b0ba39d2ac9ecb5e7d3d930fa0a3e2f44a088a75b5e342"} Oct 06 10:17:16 crc kubenswrapper[4824]: I1006 10:17:16.077390 4824 generic.go:334] "Generic (PLEG): container finished" podID="6d4717f3-9382-4a1b-a6a3-d0f0f37880c7" containerID="87baaf8f01793ebfb0d0c10a1534bef63ae25ed902ab10b41d81f273be74418b" exitCode=0 Oct 06 10:17:16 crc kubenswrapper[4824]: I1006 10:17:16.077453 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-48qqz" event={"ID":"6d4717f3-9382-4a1b-a6a3-d0f0f37880c7","Type":"ContainerDied","Data":"87baaf8f01793ebfb0d0c10a1534bef63ae25ed902ab10b41d81f273be74418b"} Oct 06 10:17:16 crc kubenswrapper[4824]: I1006 10:17:16.077479 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-48qqz" event={"ID":"6d4717f3-9382-4a1b-a6a3-d0f0f37880c7","Type":"ContainerStarted","Data":"9afb359cfc2bbfa38cbe551f380ac3fca63e4651ce676bb98c325ef54d4baad4"} Oct 06 10:17:16 crc kubenswrapper[4824]: I1006 10:17:16.087533 4824 generic.go:334] "Generic (PLEG): container finished" podID="c492c641-1fe6-44d1-8b6b-0933efbef65d" containerID="cd64c0d30963bc4055ec2f373fd1f3c755ad602d382d96c789000d1db60c8a91" exitCode=0 Oct 06 10:17:16 crc kubenswrapper[4824]: I1006 10:17:16.087768 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-qwq7x" event={"ID":"c492c641-1fe6-44d1-8b6b-0933efbef65d","Type":"ContainerDied","Data":"cd64c0d30963bc4055ec2f373fd1f3c755ad602d382d96c789000d1db60c8a91"} Oct 06 10:17:16 crc kubenswrapper[4824]: I1006 10:17:16.087803 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-qwq7x" event={"ID":"c492c641-1fe6-44d1-8b6b-0933efbef65d","Type":"ContainerStarted","Data":"55768d4883e23a64b71f9d259b29c90afc3b2750f4aab37e298ee829a3229eb4"} Oct 06 10:17:16 crc kubenswrapper[4824]: I1006 10:17:16.107119 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.490375032 podStartE2EDuration="12.107099406s" podCreationTimestamp="2025-10-06 10:17:04 +0000 UTC" firstStartedPulling="2025-10-06 10:17:05.158178388 +0000 UTC m=+1134.522601259" lastFinishedPulling="2025-10-06 10:17:14.774902772 +0000 UTC m=+1144.139325633" observedRunningTime="2025-10-06 10:17:16.104670316 +0000 UTC m=+1145.469093207" watchObservedRunningTime="2025-10-06 10:17:16.107099406 +0000 UTC m=+1145.471522267" Oct 06 10:17:16 crc kubenswrapper[4824]: I1006 10:17:16.164190 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 06 10:17:16 crc kubenswrapper[4824]: I1006 10:17:16.169998 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 06 10:17:16 crc kubenswrapper[4824]: I1006 10:17:16.195613 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 06 10:17:16 crc kubenswrapper[4824]: E1006 10:17:16.196155 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54d572d5-25f5-4531-bfc0-5a75e1a90cc6" containerName="ceilometer-notification-agent" Oct 06 10:17:16 crc kubenswrapper[4824]: I1006 10:17:16.196172 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="54d572d5-25f5-4531-bfc0-5a75e1a90cc6" containerName="ceilometer-notification-agent" Oct 06 10:17:16 crc kubenswrapper[4824]: E1006 10:17:16.196203 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54d572d5-25f5-4531-bfc0-5a75e1a90cc6" containerName="proxy-httpd" Oct 06 10:17:16 crc kubenswrapper[4824]: I1006 10:17:16.196209 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="54d572d5-25f5-4531-bfc0-5a75e1a90cc6" containerName="proxy-httpd" Oct 06 10:17:16 crc kubenswrapper[4824]: E1006 10:17:16.196230 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54d572d5-25f5-4531-bfc0-5a75e1a90cc6" containerName="ceilometer-central-agent" Oct 06 10:17:16 crc kubenswrapper[4824]: I1006 10:17:16.196236 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="54d572d5-25f5-4531-bfc0-5a75e1a90cc6" containerName="ceilometer-central-agent" Oct 06 10:17:16 crc kubenswrapper[4824]: E1006 10:17:16.196258 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54d572d5-25f5-4531-bfc0-5a75e1a90cc6" containerName="sg-core" Oct 06 10:17:16 crc kubenswrapper[4824]: I1006 10:17:16.196264 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="54d572d5-25f5-4531-bfc0-5a75e1a90cc6" containerName="sg-core" Oct 06 10:17:16 crc kubenswrapper[4824]: I1006 10:17:16.196461 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="54d572d5-25f5-4531-bfc0-5a75e1a90cc6" containerName="ceilometer-notification-agent" Oct 06 10:17:16 crc kubenswrapper[4824]: I1006 10:17:16.196474 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="54d572d5-25f5-4531-bfc0-5a75e1a90cc6" containerName="ceilometer-central-agent" Oct 06 10:17:16 crc kubenswrapper[4824]: I1006 10:17:16.196485 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="54d572d5-25f5-4531-bfc0-5a75e1a90cc6" containerName="sg-core" Oct 06 10:17:16 crc kubenswrapper[4824]: I1006 10:17:16.196494 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="54d572d5-25f5-4531-bfc0-5a75e1a90cc6" containerName="proxy-httpd" Oct 06 10:17:16 crc kubenswrapper[4824]: I1006 10:17:16.198257 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 06 10:17:16 crc kubenswrapper[4824]: I1006 10:17:16.202495 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 06 10:17:16 crc kubenswrapper[4824]: I1006 10:17:16.202623 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 06 10:17:16 crc kubenswrapper[4824]: I1006 10:17:16.204253 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 06 10:17:16 crc kubenswrapper[4824]: I1006 10:17:16.232769 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-78c847f9f4-tv8zc" podUID="68ef92a4-c4bc-4acc-be58-e14010395b8e" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.149:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.149:8443: connect: connection refused" Oct 06 10:17:16 crc kubenswrapper[4824]: I1006 10:17:16.233049 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-78c847f9f4-tv8zc" Oct 06 10:17:16 crc kubenswrapper[4824]: I1006 10:17:16.268383 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/11aa77f4-266f-492a-a859-147f00900e7d-config-data\") pod \"ceilometer-0\" (UID: \"11aa77f4-266f-492a-a859-147f00900e7d\") " pod="openstack/ceilometer-0" Oct 06 10:17:16 crc kubenswrapper[4824]: I1006 10:17:16.268426 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w5d86\" (UniqueName: \"kubernetes.io/projected/11aa77f4-266f-492a-a859-147f00900e7d-kube-api-access-w5d86\") pod \"ceilometer-0\" (UID: \"11aa77f4-266f-492a-a859-147f00900e7d\") " pod="openstack/ceilometer-0" Oct 06 10:17:16 crc kubenswrapper[4824]: I1006 10:17:16.268459 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/11aa77f4-266f-492a-a859-147f00900e7d-scripts\") pod \"ceilometer-0\" (UID: \"11aa77f4-266f-492a-a859-147f00900e7d\") " pod="openstack/ceilometer-0" Oct 06 10:17:16 crc kubenswrapper[4824]: I1006 10:17:16.268512 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/11aa77f4-266f-492a-a859-147f00900e7d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"11aa77f4-266f-492a-a859-147f00900e7d\") " pod="openstack/ceilometer-0" Oct 06 10:17:16 crc kubenswrapper[4824]: I1006 10:17:16.268532 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/11aa77f4-266f-492a-a859-147f00900e7d-log-httpd\") pod \"ceilometer-0\" (UID: \"11aa77f4-266f-492a-a859-147f00900e7d\") " pod="openstack/ceilometer-0" Oct 06 10:17:16 crc kubenswrapper[4824]: I1006 10:17:16.268571 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11aa77f4-266f-492a-a859-147f00900e7d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"11aa77f4-266f-492a-a859-147f00900e7d\") " pod="openstack/ceilometer-0" Oct 06 10:17:16 crc kubenswrapper[4824]: I1006 10:17:16.268587 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/11aa77f4-266f-492a-a859-147f00900e7d-run-httpd\") pod \"ceilometer-0\" (UID: \"11aa77f4-266f-492a-a859-147f00900e7d\") " pod="openstack/ceilometer-0" Oct 06 10:17:16 crc kubenswrapper[4824]: I1006 10:17:16.370709 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11aa77f4-266f-492a-a859-147f00900e7d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"11aa77f4-266f-492a-a859-147f00900e7d\") " pod="openstack/ceilometer-0" Oct 06 10:17:16 crc kubenswrapper[4824]: I1006 10:17:16.370777 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/11aa77f4-266f-492a-a859-147f00900e7d-run-httpd\") pod \"ceilometer-0\" (UID: \"11aa77f4-266f-492a-a859-147f00900e7d\") " pod="openstack/ceilometer-0" Oct 06 10:17:16 crc kubenswrapper[4824]: I1006 10:17:16.370967 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/11aa77f4-266f-492a-a859-147f00900e7d-config-data\") pod \"ceilometer-0\" (UID: \"11aa77f4-266f-492a-a859-147f00900e7d\") " pod="openstack/ceilometer-0" Oct 06 10:17:16 crc kubenswrapper[4824]: I1006 10:17:16.371029 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w5d86\" (UniqueName: \"kubernetes.io/projected/11aa77f4-266f-492a-a859-147f00900e7d-kube-api-access-w5d86\") pod \"ceilometer-0\" (UID: \"11aa77f4-266f-492a-a859-147f00900e7d\") " pod="openstack/ceilometer-0" Oct 06 10:17:16 crc kubenswrapper[4824]: I1006 10:17:16.371212 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/11aa77f4-266f-492a-a859-147f00900e7d-run-httpd\") pod \"ceilometer-0\" (UID: \"11aa77f4-266f-492a-a859-147f00900e7d\") " pod="openstack/ceilometer-0" Oct 06 10:17:16 crc kubenswrapper[4824]: I1006 10:17:16.371861 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/11aa77f4-266f-492a-a859-147f00900e7d-scripts\") pod \"ceilometer-0\" (UID: \"11aa77f4-266f-492a-a859-147f00900e7d\") " pod="openstack/ceilometer-0" Oct 06 10:17:16 crc kubenswrapper[4824]: I1006 10:17:16.372008 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/11aa77f4-266f-492a-a859-147f00900e7d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"11aa77f4-266f-492a-a859-147f00900e7d\") " pod="openstack/ceilometer-0" Oct 06 10:17:16 crc kubenswrapper[4824]: I1006 10:17:16.372045 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/11aa77f4-266f-492a-a859-147f00900e7d-log-httpd\") pod \"ceilometer-0\" (UID: \"11aa77f4-266f-492a-a859-147f00900e7d\") " pod="openstack/ceilometer-0" Oct 06 10:17:16 crc kubenswrapper[4824]: I1006 10:17:16.372647 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/11aa77f4-266f-492a-a859-147f00900e7d-log-httpd\") pod \"ceilometer-0\" (UID: \"11aa77f4-266f-492a-a859-147f00900e7d\") " pod="openstack/ceilometer-0" Oct 06 10:17:16 crc kubenswrapper[4824]: I1006 10:17:16.377679 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/11aa77f4-266f-492a-a859-147f00900e7d-config-data\") pod \"ceilometer-0\" (UID: \"11aa77f4-266f-492a-a859-147f00900e7d\") " pod="openstack/ceilometer-0" Oct 06 10:17:16 crc kubenswrapper[4824]: I1006 10:17:16.378214 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/11aa77f4-266f-492a-a859-147f00900e7d-scripts\") pod \"ceilometer-0\" (UID: \"11aa77f4-266f-492a-a859-147f00900e7d\") " pod="openstack/ceilometer-0" Oct 06 10:17:16 crc kubenswrapper[4824]: I1006 10:17:16.378219 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/11aa77f4-266f-492a-a859-147f00900e7d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"11aa77f4-266f-492a-a859-147f00900e7d\") " pod="openstack/ceilometer-0" Oct 06 10:17:16 crc kubenswrapper[4824]: I1006 10:17:16.378394 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11aa77f4-266f-492a-a859-147f00900e7d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"11aa77f4-266f-492a-a859-147f00900e7d\") " pod="openstack/ceilometer-0" Oct 06 10:17:16 crc kubenswrapper[4824]: I1006 10:17:16.396645 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w5d86\" (UniqueName: \"kubernetes.io/projected/11aa77f4-266f-492a-a859-147f00900e7d-kube-api-access-w5d86\") pod \"ceilometer-0\" (UID: \"11aa77f4-266f-492a-a859-147f00900e7d\") " pod="openstack/ceilometer-0" Oct 06 10:17:16 crc kubenswrapper[4824]: I1006 10:17:16.521040 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 06 10:17:17 crc kubenswrapper[4824]: I1006 10:17:17.013147 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 06 10:17:17 crc kubenswrapper[4824]: I1006 10:17:17.103579 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-78d766d697-s7wzp" event={"ID":"925ed890-0e01-4422-a2f8-4871bf90087d","Type":"ContainerStarted","Data":"4d6538132d8c04b1f4fa04859a4748b66014c89e774d4b1445a4d8763edff928"} Oct 06 10:17:17 crc kubenswrapper[4824]: I1006 10:17:17.105045 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-78d766d697-s7wzp" Oct 06 10:17:17 crc kubenswrapper[4824]: I1006 10:17:17.105082 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-78d766d697-s7wzp" Oct 06 10:17:17 crc kubenswrapper[4824]: I1006 10:17:17.112636 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"11aa77f4-266f-492a-a859-147f00900e7d","Type":"ContainerStarted","Data":"60642f1ef3bf58d2b19941565de96d05b2a3287b21f75bf626bc47fe2c4ddf1c"} Oct 06 10:17:17 crc kubenswrapper[4824]: I1006 10:17:17.131585 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-78d766d697-s7wzp" podStartSLOduration=8.131540257 podStartE2EDuration="8.131540257s" podCreationTimestamp="2025-10-06 10:17:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:17:17.130689046 +0000 UTC m=+1146.495111917" watchObservedRunningTime="2025-10-06 10:17:17.131540257 +0000 UTC m=+1146.495963118" Oct 06 10:17:17 crc kubenswrapper[4824]: I1006 10:17:17.296317 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="54d572d5-25f5-4531-bfc0-5a75e1a90cc6" path="/var/lib/kubelet/pods/54d572d5-25f5-4531-bfc0-5a75e1a90cc6/volumes" Oct 06 10:17:17 crc kubenswrapper[4824]: I1006 10:17:17.742622 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-lgz2q" Oct 06 10:17:17 crc kubenswrapper[4824]: I1006 10:17:17.921534 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wwcsv\" (UniqueName: \"kubernetes.io/projected/547f97c5-ea4d-4fbc-9f6a-d21ebd33d58f-kube-api-access-wwcsv\") pod \"547f97c5-ea4d-4fbc-9f6a-d21ebd33d58f\" (UID: \"547f97c5-ea4d-4fbc-9f6a-d21ebd33d58f\") " Oct 06 10:17:17 crc kubenswrapper[4824]: I1006 10:17:17.941828 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/547f97c5-ea4d-4fbc-9f6a-d21ebd33d58f-kube-api-access-wwcsv" (OuterVolumeSpecName: "kube-api-access-wwcsv") pod "547f97c5-ea4d-4fbc-9f6a-d21ebd33d58f" (UID: "547f97c5-ea4d-4fbc-9f6a-d21ebd33d58f"). InnerVolumeSpecName "kube-api-access-wwcsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:17:18 crc kubenswrapper[4824]: I1006 10:17:18.024367 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wwcsv\" (UniqueName: \"kubernetes.io/projected/547f97c5-ea4d-4fbc-9f6a-d21ebd33d58f-kube-api-access-wwcsv\") on node \"crc\" DevicePath \"\"" Oct 06 10:17:18 crc kubenswrapper[4824]: I1006 10:17:18.124636 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-lgz2q" event={"ID":"547f97c5-ea4d-4fbc-9f6a-d21ebd33d58f","Type":"ContainerDied","Data":"9984ef6e6b0767d57c3e9b1b51ab03c2beceb61627fa8a5a70204793e6ab9a52"} Oct 06 10:17:18 crc kubenswrapper[4824]: I1006 10:17:18.124685 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-lgz2q" Oct 06 10:17:18 crc kubenswrapper[4824]: I1006 10:17:18.124721 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9984ef6e6b0767d57c3e9b1b51ab03c2beceb61627fa8a5a70204793e6ab9a52" Oct 06 10:17:18 crc kubenswrapper[4824]: I1006 10:17:18.274469 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-qwq7x" Oct 06 10:17:18 crc kubenswrapper[4824]: I1006 10:17:18.283261 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-48qqz" Oct 06 10:17:18 crc kubenswrapper[4824]: I1006 10:17:18.431883 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-48v6b\" (UniqueName: \"kubernetes.io/projected/c492c641-1fe6-44d1-8b6b-0933efbef65d-kube-api-access-48v6b\") pod \"c492c641-1fe6-44d1-8b6b-0933efbef65d\" (UID: \"c492c641-1fe6-44d1-8b6b-0933efbef65d\") " Oct 06 10:17:18 crc kubenswrapper[4824]: I1006 10:17:18.432068 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ndsjt\" (UniqueName: \"kubernetes.io/projected/6d4717f3-9382-4a1b-a6a3-d0f0f37880c7-kube-api-access-ndsjt\") pod \"6d4717f3-9382-4a1b-a6a3-d0f0f37880c7\" (UID: \"6d4717f3-9382-4a1b-a6a3-d0f0f37880c7\") " Oct 06 10:17:18 crc kubenswrapper[4824]: I1006 10:17:18.437231 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c492c641-1fe6-44d1-8b6b-0933efbef65d-kube-api-access-48v6b" (OuterVolumeSpecName: "kube-api-access-48v6b") pod "c492c641-1fe6-44d1-8b6b-0933efbef65d" (UID: "c492c641-1fe6-44d1-8b6b-0933efbef65d"). InnerVolumeSpecName "kube-api-access-48v6b". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:17:18 crc kubenswrapper[4824]: I1006 10:17:18.437513 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6d4717f3-9382-4a1b-a6a3-d0f0f37880c7-kube-api-access-ndsjt" (OuterVolumeSpecName: "kube-api-access-ndsjt") pod "6d4717f3-9382-4a1b-a6a3-d0f0f37880c7" (UID: "6d4717f3-9382-4a1b-a6a3-d0f0f37880c7"). InnerVolumeSpecName "kube-api-access-ndsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:17:18 crc kubenswrapper[4824]: I1006 10:17:18.538285 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ndsjt\" (UniqueName: \"kubernetes.io/projected/6d4717f3-9382-4a1b-a6a3-d0f0f37880c7-kube-api-access-ndsjt\") on node \"crc\" DevicePath \"\"" Oct 06 10:17:18 crc kubenswrapper[4824]: I1006 10:17:18.538324 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-48v6b\" (UniqueName: \"kubernetes.io/projected/c492c641-1fe6-44d1-8b6b-0933efbef65d-kube-api-access-48v6b\") on node \"crc\" DevicePath \"\"" Oct 06 10:17:19 crc kubenswrapper[4824]: I1006 10:17:19.133675 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-qwq7x" Oct 06 10:17:19 crc kubenswrapper[4824]: I1006 10:17:19.133676 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-qwq7x" event={"ID":"c492c641-1fe6-44d1-8b6b-0933efbef65d","Type":"ContainerDied","Data":"55768d4883e23a64b71f9d259b29c90afc3b2750f4aab37e298ee829a3229eb4"} Oct 06 10:17:19 crc kubenswrapper[4824]: I1006 10:17:19.134121 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="55768d4883e23a64b71f9d259b29c90afc3b2750f4aab37e298ee829a3229eb4" Oct 06 10:17:19 crc kubenswrapper[4824]: I1006 10:17:19.136736 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"11aa77f4-266f-492a-a859-147f00900e7d","Type":"ContainerStarted","Data":"2355e113a1c344098469fe0db5bc6338b9e60c558e6de64c5e22eb3fd0b81b6c"} Oct 06 10:17:19 crc kubenswrapper[4824]: I1006 10:17:19.139500 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-48qqz" Oct 06 10:17:19 crc kubenswrapper[4824]: I1006 10:17:19.143478 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-48qqz" event={"ID":"6d4717f3-9382-4a1b-a6a3-d0f0f37880c7","Type":"ContainerDied","Data":"9afb359cfc2bbfa38cbe551f380ac3fca63e4651ce676bb98c325ef54d4baad4"} Oct 06 10:17:19 crc kubenswrapper[4824]: I1006 10:17:19.143512 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9afb359cfc2bbfa38cbe551f380ac3fca63e4651ce676bb98c325ef54d4baad4" Oct 06 10:17:19 crc kubenswrapper[4824]: I1006 10:17:19.263960 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 06 10:17:20 crc kubenswrapper[4824]: I1006 10:17:20.146575 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"11aa77f4-266f-492a-a859-147f00900e7d","Type":"ContainerStarted","Data":"f61f7288551177fba804cc0b69018f73fec04b977d927c28815981511db78a37"} Oct 06 10:17:21 crc kubenswrapper[4824]: I1006 10:17:21.159867 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"11aa77f4-266f-492a-a859-147f00900e7d","Type":"ContainerStarted","Data":"694d57e11eb58492921ebc1dd3762e380619452042291e449959e0433c3add29"} Oct 06 10:17:22 crc kubenswrapper[4824]: I1006 10:17:22.173028 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"11aa77f4-266f-492a-a859-147f00900e7d","Type":"ContainerStarted","Data":"15ab9953bb9e65b6ace0ab01643fa9490513f74cc7bfc4cdb4a6a00331ab253c"} Oct 06 10:17:22 crc kubenswrapper[4824]: I1006 10:17:22.174767 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 06 10:17:22 crc kubenswrapper[4824]: I1006 10:17:22.173230 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="11aa77f4-266f-492a-a859-147f00900e7d" containerName="sg-core" containerID="cri-o://694d57e11eb58492921ebc1dd3762e380619452042291e449959e0433c3add29" gracePeriod=30 Oct 06 10:17:22 crc kubenswrapper[4824]: I1006 10:17:22.173238 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="11aa77f4-266f-492a-a859-147f00900e7d" containerName="proxy-httpd" containerID="cri-o://15ab9953bb9e65b6ace0ab01643fa9490513f74cc7bfc4cdb4a6a00331ab253c" gracePeriod=30 Oct 06 10:17:22 crc kubenswrapper[4824]: I1006 10:17:22.173272 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="11aa77f4-266f-492a-a859-147f00900e7d" containerName="ceilometer-notification-agent" containerID="cri-o://f61f7288551177fba804cc0b69018f73fec04b977d927c28815981511db78a37" gracePeriod=30 Oct 06 10:17:22 crc kubenswrapper[4824]: I1006 10:17:22.173148 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="11aa77f4-266f-492a-a859-147f00900e7d" containerName="ceilometer-central-agent" containerID="cri-o://2355e113a1c344098469fe0db5bc6338b9e60c558e6de64c5e22eb3fd0b81b6c" gracePeriod=30 Oct 06 10:17:22 crc kubenswrapper[4824]: I1006 10:17:22.217103 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.920838472 podStartE2EDuration="6.217063176s" podCreationTimestamp="2025-10-06 10:17:16 +0000 UTC" firstStartedPulling="2025-10-06 10:17:17.027934634 +0000 UTC m=+1146.392357495" lastFinishedPulling="2025-10-06 10:17:21.324159338 +0000 UTC m=+1150.688582199" observedRunningTime="2025-10-06 10:17:22.197307907 +0000 UTC m=+1151.561730768" watchObservedRunningTime="2025-10-06 10:17:22.217063176 +0000 UTC m=+1151.581486037" Oct 06 10:17:23 crc kubenswrapper[4824]: I1006 10:17:23.198869 4824 generic.go:334] "Generic (PLEG): container finished" podID="11aa77f4-266f-492a-a859-147f00900e7d" containerID="15ab9953bb9e65b6ace0ab01643fa9490513f74cc7bfc4cdb4a6a00331ab253c" exitCode=0 Oct 06 10:17:23 crc kubenswrapper[4824]: I1006 10:17:23.199170 4824 generic.go:334] "Generic (PLEG): container finished" podID="11aa77f4-266f-492a-a859-147f00900e7d" containerID="694d57e11eb58492921ebc1dd3762e380619452042291e449959e0433c3add29" exitCode=2 Oct 06 10:17:23 crc kubenswrapper[4824]: I1006 10:17:23.199181 4824 generic.go:334] "Generic (PLEG): container finished" podID="11aa77f4-266f-492a-a859-147f00900e7d" containerID="f61f7288551177fba804cc0b69018f73fec04b977d927c28815981511db78a37" exitCode=0 Oct 06 10:17:23 crc kubenswrapper[4824]: I1006 10:17:23.198939 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"11aa77f4-266f-492a-a859-147f00900e7d","Type":"ContainerDied","Data":"15ab9953bb9e65b6ace0ab01643fa9490513f74cc7bfc4cdb4a6a00331ab253c"} Oct 06 10:17:23 crc kubenswrapper[4824]: I1006 10:17:23.199218 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"11aa77f4-266f-492a-a859-147f00900e7d","Type":"ContainerDied","Data":"694d57e11eb58492921ebc1dd3762e380619452042291e449959e0433c3add29"} Oct 06 10:17:23 crc kubenswrapper[4824]: I1006 10:17:23.199233 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"11aa77f4-266f-492a-a859-147f00900e7d","Type":"ContainerDied","Data":"f61f7288551177fba804cc0b69018f73fec04b977d927c28815981511db78a37"} Oct 06 10:17:24 crc kubenswrapper[4824]: I1006 10:17:24.580442 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-78d766d697-s7wzp" Oct 06 10:17:24 crc kubenswrapper[4824]: I1006 10:17:24.588872 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-78d766d697-s7wzp" Oct 06 10:17:25 crc kubenswrapper[4824]: I1006 10:17:25.150146 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-78c847f9f4-tv8zc" Oct 06 10:17:25 crc kubenswrapper[4824]: I1006 10:17:25.221234 4824 generic.go:334] "Generic (PLEG): container finished" podID="68ef92a4-c4bc-4acc-be58-e14010395b8e" containerID="e347cab97c01ec84b5735cd302c782ea65d188c9d2c13baf5983c3a2843b3758" exitCode=137 Oct 06 10:17:25 crc kubenswrapper[4824]: I1006 10:17:25.221529 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-78c847f9f4-tv8zc" event={"ID":"68ef92a4-c4bc-4acc-be58-e14010395b8e","Type":"ContainerDied","Data":"e347cab97c01ec84b5735cd302c782ea65d188c9d2c13baf5983c3a2843b3758"} Oct 06 10:17:25 crc kubenswrapper[4824]: I1006 10:17:25.221578 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-78c847f9f4-tv8zc" event={"ID":"68ef92a4-c4bc-4acc-be58-e14010395b8e","Type":"ContainerDied","Data":"7a182aa0ff510239fa3e786b04999f60551ad03be619039603b91a15a219780e"} Oct 06 10:17:25 crc kubenswrapper[4824]: I1006 10:17:25.221628 4824 scope.go:117] "RemoveContainer" containerID="62bfee4f7781b10bfd4f3b2388238d938ea44501a5d20380123b779025d75b32" Oct 06 10:17:25 crc kubenswrapper[4824]: I1006 10:17:25.221691 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-78c847f9f4-tv8zc" Oct 06 10:17:25 crc kubenswrapper[4824]: I1006 10:17:25.277476 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hxgpf\" (UniqueName: \"kubernetes.io/projected/68ef92a4-c4bc-4acc-be58-e14010395b8e-kube-api-access-hxgpf\") pod \"68ef92a4-c4bc-4acc-be58-e14010395b8e\" (UID: \"68ef92a4-c4bc-4acc-be58-e14010395b8e\") " Oct 06 10:17:25 crc kubenswrapper[4824]: I1006 10:17:25.277595 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/68ef92a4-c4bc-4acc-be58-e14010395b8e-horizon-tls-certs\") pod \"68ef92a4-c4bc-4acc-be58-e14010395b8e\" (UID: \"68ef92a4-c4bc-4acc-be58-e14010395b8e\") " Oct 06 10:17:25 crc kubenswrapper[4824]: I1006 10:17:25.277641 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/68ef92a4-c4bc-4acc-be58-e14010395b8e-config-data\") pod \"68ef92a4-c4bc-4acc-be58-e14010395b8e\" (UID: \"68ef92a4-c4bc-4acc-be58-e14010395b8e\") " Oct 06 10:17:25 crc kubenswrapper[4824]: I1006 10:17:25.277770 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/68ef92a4-c4bc-4acc-be58-e14010395b8e-scripts\") pod \"68ef92a4-c4bc-4acc-be58-e14010395b8e\" (UID: \"68ef92a4-c4bc-4acc-be58-e14010395b8e\") " Oct 06 10:17:25 crc kubenswrapper[4824]: I1006 10:17:25.277840 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/68ef92a4-c4bc-4acc-be58-e14010395b8e-logs\") pod \"68ef92a4-c4bc-4acc-be58-e14010395b8e\" (UID: \"68ef92a4-c4bc-4acc-be58-e14010395b8e\") " Oct 06 10:17:25 crc kubenswrapper[4824]: I1006 10:17:25.277995 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68ef92a4-c4bc-4acc-be58-e14010395b8e-combined-ca-bundle\") pod \"68ef92a4-c4bc-4acc-be58-e14010395b8e\" (UID: \"68ef92a4-c4bc-4acc-be58-e14010395b8e\") " Oct 06 10:17:25 crc kubenswrapper[4824]: I1006 10:17:25.278126 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/68ef92a4-c4bc-4acc-be58-e14010395b8e-horizon-secret-key\") pod \"68ef92a4-c4bc-4acc-be58-e14010395b8e\" (UID: \"68ef92a4-c4bc-4acc-be58-e14010395b8e\") " Oct 06 10:17:25 crc kubenswrapper[4824]: I1006 10:17:25.278778 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/68ef92a4-c4bc-4acc-be58-e14010395b8e-logs" (OuterVolumeSpecName: "logs") pod "68ef92a4-c4bc-4acc-be58-e14010395b8e" (UID: "68ef92a4-c4bc-4acc-be58-e14010395b8e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:17:25 crc kubenswrapper[4824]: I1006 10:17:25.285097 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/68ef92a4-c4bc-4acc-be58-e14010395b8e-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "68ef92a4-c4bc-4acc-be58-e14010395b8e" (UID: "68ef92a4-c4bc-4acc-be58-e14010395b8e"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:17:25 crc kubenswrapper[4824]: I1006 10:17:25.318873 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/68ef92a4-c4bc-4acc-be58-e14010395b8e-config-data" (OuterVolumeSpecName: "config-data") pod "68ef92a4-c4bc-4acc-be58-e14010395b8e" (UID: "68ef92a4-c4bc-4acc-be58-e14010395b8e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:17:25 crc kubenswrapper[4824]: I1006 10:17:25.320679 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/68ef92a4-c4bc-4acc-be58-e14010395b8e-scripts" (OuterVolumeSpecName: "scripts") pod "68ef92a4-c4bc-4acc-be58-e14010395b8e" (UID: "68ef92a4-c4bc-4acc-be58-e14010395b8e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:17:25 crc kubenswrapper[4824]: I1006 10:17:25.322775 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/68ef92a4-c4bc-4acc-be58-e14010395b8e-kube-api-access-hxgpf" (OuterVolumeSpecName: "kube-api-access-hxgpf") pod "68ef92a4-c4bc-4acc-be58-e14010395b8e" (UID: "68ef92a4-c4bc-4acc-be58-e14010395b8e"). InnerVolumeSpecName "kube-api-access-hxgpf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:17:25 crc kubenswrapper[4824]: I1006 10:17:25.336924 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/68ef92a4-c4bc-4acc-be58-e14010395b8e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "68ef92a4-c4bc-4acc-be58-e14010395b8e" (UID: "68ef92a4-c4bc-4acc-be58-e14010395b8e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:17:25 crc kubenswrapper[4824]: I1006 10:17:25.352851 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/68ef92a4-c4bc-4acc-be58-e14010395b8e-horizon-tls-certs" (OuterVolumeSpecName: "horizon-tls-certs") pod "68ef92a4-c4bc-4acc-be58-e14010395b8e" (UID: "68ef92a4-c4bc-4acc-be58-e14010395b8e"). InnerVolumeSpecName "horizon-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:17:25 crc kubenswrapper[4824]: I1006 10:17:25.380627 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68ef92a4-c4bc-4acc-be58-e14010395b8e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 10:17:25 crc kubenswrapper[4824]: I1006 10:17:25.380674 4824 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/68ef92a4-c4bc-4acc-be58-e14010395b8e-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 06 10:17:25 crc kubenswrapper[4824]: I1006 10:17:25.380688 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hxgpf\" (UniqueName: \"kubernetes.io/projected/68ef92a4-c4bc-4acc-be58-e14010395b8e-kube-api-access-hxgpf\") on node \"crc\" DevicePath \"\"" Oct 06 10:17:25 crc kubenswrapper[4824]: I1006 10:17:25.380703 4824 reconciler_common.go:293] "Volume detached for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/68ef92a4-c4bc-4acc-be58-e14010395b8e-horizon-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 06 10:17:25 crc kubenswrapper[4824]: I1006 10:17:25.380717 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/68ef92a4-c4bc-4acc-be58-e14010395b8e-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 10:17:25 crc kubenswrapper[4824]: I1006 10:17:25.380730 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/68ef92a4-c4bc-4acc-be58-e14010395b8e-scripts\") on node \"crc\" DevicePath \"\"" Oct 06 10:17:25 crc kubenswrapper[4824]: I1006 10:17:25.380742 4824 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/68ef92a4-c4bc-4acc-be58-e14010395b8e-logs\") on node \"crc\" DevicePath \"\"" Oct 06 10:17:25 crc kubenswrapper[4824]: I1006 10:17:25.405341 4824 scope.go:117] "RemoveContainer" containerID="e347cab97c01ec84b5735cd302c782ea65d188c9d2c13baf5983c3a2843b3758" Oct 06 10:17:25 crc kubenswrapper[4824]: I1006 10:17:25.434262 4824 scope.go:117] "RemoveContainer" containerID="62bfee4f7781b10bfd4f3b2388238d938ea44501a5d20380123b779025d75b32" Oct 06 10:17:25 crc kubenswrapper[4824]: E1006 10:17:25.434779 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"62bfee4f7781b10bfd4f3b2388238d938ea44501a5d20380123b779025d75b32\": container with ID starting with 62bfee4f7781b10bfd4f3b2388238d938ea44501a5d20380123b779025d75b32 not found: ID does not exist" containerID="62bfee4f7781b10bfd4f3b2388238d938ea44501a5d20380123b779025d75b32" Oct 06 10:17:25 crc kubenswrapper[4824]: I1006 10:17:25.434830 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"62bfee4f7781b10bfd4f3b2388238d938ea44501a5d20380123b779025d75b32"} err="failed to get container status \"62bfee4f7781b10bfd4f3b2388238d938ea44501a5d20380123b779025d75b32\": rpc error: code = NotFound desc = could not find container \"62bfee4f7781b10bfd4f3b2388238d938ea44501a5d20380123b779025d75b32\": container with ID starting with 62bfee4f7781b10bfd4f3b2388238d938ea44501a5d20380123b779025d75b32 not found: ID does not exist" Oct 06 10:17:25 crc kubenswrapper[4824]: I1006 10:17:25.434870 4824 scope.go:117] "RemoveContainer" containerID="e347cab97c01ec84b5735cd302c782ea65d188c9d2c13baf5983c3a2843b3758" Oct 06 10:17:25 crc kubenswrapper[4824]: E1006 10:17:25.435735 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e347cab97c01ec84b5735cd302c782ea65d188c9d2c13baf5983c3a2843b3758\": container with ID starting with e347cab97c01ec84b5735cd302c782ea65d188c9d2c13baf5983c3a2843b3758 not found: ID does not exist" containerID="e347cab97c01ec84b5735cd302c782ea65d188c9d2c13baf5983c3a2843b3758" Oct 06 10:17:25 crc kubenswrapper[4824]: I1006 10:17:25.435801 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e347cab97c01ec84b5735cd302c782ea65d188c9d2c13baf5983c3a2843b3758"} err="failed to get container status \"e347cab97c01ec84b5735cd302c782ea65d188c9d2c13baf5983c3a2843b3758\": rpc error: code = NotFound desc = could not find container \"e347cab97c01ec84b5735cd302c782ea65d188c9d2c13baf5983c3a2843b3758\": container with ID starting with e347cab97c01ec84b5735cd302c782ea65d188c9d2c13baf5983c3a2843b3758 not found: ID does not exist" Oct 06 10:17:25 crc kubenswrapper[4824]: I1006 10:17:25.561361 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-78c847f9f4-tv8zc"] Oct 06 10:17:25 crc kubenswrapper[4824]: I1006 10:17:25.571284 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-78c847f9f4-tv8zc"] Oct 06 10:17:27 crc kubenswrapper[4824]: I1006 10:17:27.285479 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="68ef92a4-c4bc-4acc-be58-e14010395b8e" path="/var/lib/kubelet/pods/68ef92a4-c4bc-4acc-be58-e14010395b8e/volumes" Oct 06 10:17:27 crc kubenswrapper[4824]: I1006 10:17:27.921227 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 06 10:17:28 crc kubenswrapper[4824]: I1006 10:17:28.030755 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/11aa77f4-266f-492a-a859-147f00900e7d-sg-core-conf-yaml\") pod \"11aa77f4-266f-492a-a859-147f00900e7d\" (UID: \"11aa77f4-266f-492a-a859-147f00900e7d\") " Oct 06 10:17:28 crc kubenswrapper[4824]: I1006 10:17:28.031108 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/11aa77f4-266f-492a-a859-147f00900e7d-log-httpd\") pod \"11aa77f4-266f-492a-a859-147f00900e7d\" (UID: \"11aa77f4-266f-492a-a859-147f00900e7d\") " Oct 06 10:17:28 crc kubenswrapper[4824]: I1006 10:17:28.031259 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/11aa77f4-266f-492a-a859-147f00900e7d-config-data\") pod \"11aa77f4-266f-492a-a859-147f00900e7d\" (UID: \"11aa77f4-266f-492a-a859-147f00900e7d\") " Oct 06 10:17:28 crc kubenswrapper[4824]: I1006 10:17:28.031309 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11aa77f4-266f-492a-a859-147f00900e7d-combined-ca-bundle\") pod \"11aa77f4-266f-492a-a859-147f00900e7d\" (UID: \"11aa77f4-266f-492a-a859-147f00900e7d\") " Oct 06 10:17:28 crc kubenswrapper[4824]: I1006 10:17:28.031380 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w5d86\" (UniqueName: \"kubernetes.io/projected/11aa77f4-266f-492a-a859-147f00900e7d-kube-api-access-w5d86\") pod \"11aa77f4-266f-492a-a859-147f00900e7d\" (UID: \"11aa77f4-266f-492a-a859-147f00900e7d\") " Oct 06 10:17:28 crc kubenswrapper[4824]: I1006 10:17:28.031411 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/11aa77f4-266f-492a-a859-147f00900e7d-run-httpd\") pod \"11aa77f4-266f-492a-a859-147f00900e7d\" (UID: \"11aa77f4-266f-492a-a859-147f00900e7d\") " Oct 06 10:17:28 crc kubenswrapper[4824]: I1006 10:17:28.031445 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/11aa77f4-266f-492a-a859-147f00900e7d-scripts\") pod \"11aa77f4-266f-492a-a859-147f00900e7d\" (UID: \"11aa77f4-266f-492a-a859-147f00900e7d\") " Oct 06 10:17:28 crc kubenswrapper[4824]: I1006 10:17:28.031924 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/11aa77f4-266f-492a-a859-147f00900e7d-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "11aa77f4-266f-492a-a859-147f00900e7d" (UID: "11aa77f4-266f-492a-a859-147f00900e7d"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:17:28 crc kubenswrapper[4824]: I1006 10:17:28.032072 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/11aa77f4-266f-492a-a859-147f00900e7d-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "11aa77f4-266f-492a-a859-147f00900e7d" (UID: "11aa77f4-266f-492a-a859-147f00900e7d"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:17:28 crc kubenswrapper[4824]: I1006 10:17:28.037864 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/11aa77f4-266f-492a-a859-147f00900e7d-scripts" (OuterVolumeSpecName: "scripts") pod "11aa77f4-266f-492a-a859-147f00900e7d" (UID: "11aa77f4-266f-492a-a859-147f00900e7d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:17:28 crc kubenswrapper[4824]: I1006 10:17:28.038110 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/11aa77f4-266f-492a-a859-147f00900e7d-kube-api-access-w5d86" (OuterVolumeSpecName: "kube-api-access-w5d86") pod "11aa77f4-266f-492a-a859-147f00900e7d" (UID: "11aa77f4-266f-492a-a859-147f00900e7d"). InnerVolumeSpecName "kube-api-access-w5d86". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:17:28 crc kubenswrapper[4824]: I1006 10:17:28.069564 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/11aa77f4-266f-492a-a859-147f00900e7d-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "11aa77f4-266f-492a-a859-147f00900e7d" (UID: "11aa77f4-266f-492a-a859-147f00900e7d"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:17:28 crc kubenswrapper[4824]: I1006 10:17:28.125093 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/11aa77f4-266f-492a-a859-147f00900e7d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "11aa77f4-266f-492a-a859-147f00900e7d" (UID: "11aa77f4-266f-492a-a859-147f00900e7d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:17:28 crc kubenswrapper[4824]: I1006 10:17:28.133225 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11aa77f4-266f-492a-a859-147f00900e7d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 10:17:28 crc kubenswrapper[4824]: I1006 10:17:28.133264 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w5d86\" (UniqueName: \"kubernetes.io/projected/11aa77f4-266f-492a-a859-147f00900e7d-kube-api-access-w5d86\") on node \"crc\" DevicePath \"\"" Oct 06 10:17:28 crc kubenswrapper[4824]: I1006 10:17:28.133276 4824 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/11aa77f4-266f-492a-a859-147f00900e7d-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 06 10:17:28 crc kubenswrapper[4824]: I1006 10:17:28.133285 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/11aa77f4-266f-492a-a859-147f00900e7d-scripts\") on node \"crc\" DevicePath \"\"" Oct 06 10:17:28 crc kubenswrapper[4824]: I1006 10:17:28.133295 4824 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/11aa77f4-266f-492a-a859-147f00900e7d-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 06 10:17:28 crc kubenswrapper[4824]: I1006 10:17:28.133304 4824 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/11aa77f4-266f-492a-a859-147f00900e7d-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 06 10:17:28 crc kubenswrapper[4824]: I1006 10:17:28.154347 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/11aa77f4-266f-492a-a859-147f00900e7d-config-data" (OuterVolumeSpecName: "config-data") pod "11aa77f4-266f-492a-a859-147f00900e7d" (UID: "11aa77f4-266f-492a-a859-147f00900e7d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:17:28 crc kubenswrapper[4824]: I1006 10:17:28.234514 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/11aa77f4-266f-492a-a859-147f00900e7d-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 10:17:28 crc kubenswrapper[4824]: I1006 10:17:28.253130 4824 generic.go:334] "Generic (PLEG): container finished" podID="11aa77f4-266f-492a-a859-147f00900e7d" containerID="2355e113a1c344098469fe0db5bc6338b9e60c558e6de64c5e22eb3fd0b81b6c" exitCode=0 Oct 06 10:17:28 crc kubenswrapper[4824]: I1006 10:17:28.253174 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"11aa77f4-266f-492a-a859-147f00900e7d","Type":"ContainerDied","Data":"2355e113a1c344098469fe0db5bc6338b9e60c558e6de64c5e22eb3fd0b81b6c"} Oct 06 10:17:28 crc kubenswrapper[4824]: I1006 10:17:28.253206 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"11aa77f4-266f-492a-a859-147f00900e7d","Type":"ContainerDied","Data":"60642f1ef3bf58d2b19941565de96d05b2a3287b21f75bf626bc47fe2c4ddf1c"} Oct 06 10:17:28 crc kubenswrapper[4824]: I1006 10:17:28.253225 4824 scope.go:117] "RemoveContainer" containerID="15ab9953bb9e65b6ace0ab01643fa9490513f74cc7bfc4cdb4a6a00331ab253c" Oct 06 10:17:28 crc kubenswrapper[4824]: I1006 10:17:28.253225 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 06 10:17:28 crc kubenswrapper[4824]: I1006 10:17:28.274696 4824 scope.go:117] "RemoveContainer" containerID="694d57e11eb58492921ebc1dd3762e380619452042291e449959e0433c3add29" Oct 06 10:17:28 crc kubenswrapper[4824]: I1006 10:17:28.285004 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 06 10:17:28 crc kubenswrapper[4824]: I1006 10:17:28.298430 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 06 10:17:28 crc kubenswrapper[4824]: I1006 10:17:28.300508 4824 scope.go:117] "RemoveContainer" containerID="f61f7288551177fba804cc0b69018f73fec04b977d927c28815981511db78a37" Oct 06 10:17:28 crc kubenswrapper[4824]: I1006 10:17:28.328255 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 06 10:17:28 crc kubenswrapper[4824]: E1006 10:17:28.328762 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11aa77f4-266f-492a-a859-147f00900e7d" containerName="ceilometer-central-agent" Oct 06 10:17:28 crc kubenswrapper[4824]: I1006 10:17:28.328780 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="11aa77f4-266f-492a-a859-147f00900e7d" containerName="ceilometer-central-agent" Oct 06 10:17:28 crc kubenswrapper[4824]: E1006 10:17:28.328795 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11aa77f4-266f-492a-a859-147f00900e7d" containerName="sg-core" Oct 06 10:17:28 crc kubenswrapper[4824]: I1006 10:17:28.328804 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="11aa77f4-266f-492a-a859-147f00900e7d" containerName="sg-core" Oct 06 10:17:28 crc kubenswrapper[4824]: E1006 10:17:28.328831 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="547f97c5-ea4d-4fbc-9f6a-d21ebd33d58f" containerName="mariadb-database-create" Oct 06 10:17:28 crc kubenswrapper[4824]: I1006 10:17:28.328838 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="547f97c5-ea4d-4fbc-9f6a-d21ebd33d58f" containerName="mariadb-database-create" Oct 06 10:17:28 crc kubenswrapper[4824]: E1006 10:17:28.328854 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11aa77f4-266f-492a-a859-147f00900e7d" containerName="ceilometer-notification-agent" Oct 06 10:17:28 crc kubenswrapper[4824]: I1006 10:17:28.328861 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="11aa77f4-266f-492a-a859-147f00900e7d" containerName="ceilometer-notification-agent" Oct 06 10:17:28 crc kubenswrapper[4824]: E1006 10:17:28.328875 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11aa77f4-266f-492a-a859-147f00900e7d" containerName="proxy-httpd" Oct 06 10:17:28 crc kubenswrapper[4824]: I1006 10:17:28.328881 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="11aa77f4-266f-492a-a859-147f00900e7d" containerName="proxy-httpd" Oct 06 10:17:28 crc kubenswrapper[4824]: E1006 10:17:28.328910 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68ef92a4-c4bc-4acc-be58-e14010395b8e" containerName="horizon-log" Oct 06 10:17:28 crc kubenswrapper[4824]: I1006 10:17:28.328917 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="68ef92a4-c4bc-4acc-be58-e14010395b8e" containerName="horizon-log" Oct 06 10:17:28 crc kubenswrapper[4824]: E1006 10:17:28.328926 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68ef92a4-c4bc-4acc-be58-e14010395b8e" containerName="horizon" Oct 06 10:17:28 crc kubenswrapper[4824]: I1006 10:17:28.328934 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="68ef92a4-c4bc-4acc-be58-e14010395b8e" containerName="horizon" Oct 06 10:17:28 crc kubenswrapper[4824]: E1006 10:17:28.328949 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c492c641-1fe6-44d1-8b6b-0933efbef65d" containerName="mariadb-database-create" Oct 06 10:17:28 crc kubenswrapper[4824]: I1006 10:17:28.328956 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="c492c641-1fe6-44d1-8b6b-0933efbef65d" containerName="mariadb-database-create" Oct 06 10:17:28 crc kubenswrapper[4824]: E1006 10:17:28.329002 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d4717f3-9382-4a1b-a6a3-d0f0f37880c7" containerName="mariadb-database-create" Oct 06 10:17:28 crc kubenswrapper[4824]: I1006 10:17:28.329009 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d4717f3-9382-4a1b-a6a3-d0f0f37880c7" containerName="mariadb-database-create" Oct 06 10:17:28 crc kubenswrapper[4824]: I1006 10:17:28.329200 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="11aa77f4-266f-492a-a859-147f00900e7d" containerName="proxy-httpd" Oct 06 10:17:28 crc kubenswrapper[4824]: I1006 10:17:28.329234 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="68ef92a4-c4bc-4acc-be58-e14010395b8e" containerName="horizon" Oct 06 10:17:28 crc kubenswrapper[4824]: I1006 10:17:28.329245 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="11aa77f4-266f-492a-a859-147f00900e7d" containerName="ceilometer-notification-agent" Oct 06 10:17:28 crc kubenswrapper[4824]: I1006 10:17:28.329253 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="547f97c5-ea4d-4fbc-9f6a-d21ebd33d58f" containerName="mariadb-database-create" Oct 06 10:17:28 crc kubenswrapper[4824]: I1006 10:17:28.329264 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="68ef92a4-c4bc-4acc-be58-e14010395b8e" containerName="horizon-log" Oct 06 10:17:28 crc kubenswrapper[4824]: I1006 10:17:28.329272 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="11aa77f4-266f-492a-a859-147f00900e7d" containerName="sg-core" Oct 06 10:17:28 crc kubenswrapper[4824]: I1006 10:17:28.329288 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="6d4717f3-9382-4a1b-a6a3-d0f0f37880c7" containerName="mariadb-database-create" Oct 06 10:17:28 crc kubenswrapper[4824]: I1006 10:17:28.329318 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="11aa77f4-266f-492a-a859-147f00900e7d" containerName="ceilometer-central-agent" Oct 06 10:17:28 crc kubenswrapper[4824]: I1006 10:17:28.329327 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="c492c641-1fe6-44d1-8b6b-0933efbef65d" containerName="mariadb-database-create" Oct 06 10:17:28 crc kubenswrapper[4824]: I1006 10:17:28.331410 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 06 10:17:28 crc kubenswrapper[4824]: I1006 10:17:28.332149 4824 scope.go:117] "RemoveContainer" containerID="2355e113a1c344098469fe0db5bc6338b9e60c558e6de64c5e22eb3fd0b81b6c" Oct 06 10:17:28 crc kubenswrapper[4824]: I1006 10:17:28.333783 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 06 10:17:28 crc kubenswrapper[4824]: I1006 10:17:28.334117 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 06 10:17:28 crc kubenswrapper[4824]: I1006 10:17:28.338758 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 06 10:17:28 crc kubenswrapper[4824]: I1006 10:17:28.370425 4824 scope.go:117] "RemoveContainer" containerID="15ab9953bb9e65b6ace0ab01643fa9490513f74cc7bfc4cdb4a6a00331ab253c" Oct 06 10:17:28 crc kubenswrapper[4824]: E1006 10:17:28.375379 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"15ab9953bb9e65b6ace0ab01643fa9490513f74cc7bfc4cdb4a6a00331ab253c\": container with ID starting with 15ab9953bb9e65b6ace0ab01643fa9490513f74cc7bfc4cdb4a6a00331ab253c not found: ID does not exist" containerID="15ab9953bb9e65b6ace0ab01643fa9490513f74cc7bfc4cdb4a6a00331ab253c" Oct 06 10:17:28 crc kubenswrapper[4824]: I1006 10:17:28.375420 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"15ab9953bb9e65b6ace0ab01643fa9490513f74cc7bfc4cdb4a6a00331ab253c"} err="failed to get container status \"15ab9953bb9e65b6ace0ab01643fa9490513f74cc7bfc4cdb4a6a00331ab253c\": rpc error: code = NotFound desc = could not find container \"15ab9953bb9e65b6ace0ab01643fa9490513f74cc7bfc4cdb4a6a00331ab253c\": container with ID starting with 15ab9953bb9e65b6ace0ab01643fa9490513f74cc7bfc4cdb4a6a00331ab253c not found: ID does not exist" Oct 06 10:17:28 crc kubenswrapper[4824]: I1006 10:17:28.375445 4824 scope.go:117] "RemoveContainer" containerID="694d57e11eb58492921ebc1dd3762e380619452042291e449959e0433c3add29" Oct 06 10:17:28 crc kubenswrapper[4824]: E1006 10:17:28.375888 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"694d57e11eb58492921ebc1dd3762e380619452042291e449959e0433c3add29\": container with ID starting with 694d57e11eb58492921ebc1dd3762e380619452042291e449959e0433c3add29 not found: ID does not exist" containerID="694d57e11eb58492921ebc1dd3762e380619452042291e449959e0433c3add29" Oct 06 10:17:28 crc kubenswrapper[4824]: I1006 10:17:28.375907 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"694d57e11eb58492921ebc1dd3762e380619452042291e449959e0433c3add29"} err="failed to get container status \"694d57e11eb58492921ebc1dd3762e380619452042291e449959e0433c3add29\": rpc error: code = NotFound desc = could not find container \"694d57e11eb58492921ebc1dd3762e380619452042291e449959e0433c3add29\": container with ID starting with 694d57e11eb58492921ebc1dd3762e380619452042291e449959e0433c3add29 not found: ID does not exist" Oct 06 10:17:28 crc kubenswrapper[4824]: I1006 10:17:28.375919 4824 scope.go:117] "RemoveContainer" containerID="f61f7288551177fba804cc0b69018f73fec04b977d927c28815981511db78a37" Oct 06 10:17:28 crc kubenswrapper[4824]: E1006 10:17:28.376141 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f61f7288551177fba804cc0b69018f73fec04b977d927c28815981511db78a37\": container with ID starting with f61f7288551177fba804cc0b69018f73fec04b977d927c28815981511db78a37 not found: ID does not exist" containerID="f61f7288551177fba804cc0b69018f73fec04b977d927c28815981511db78a37" Oct 06 10:17:28 crc kubenswrapper[4824]: I1006 10:17:28.376196 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f61f7288551177fba804cc0b69018f73fec04b977d927c28815981511db78a37"} err="failed to get container status \"f61f7288551177fba804cc0b69018f73fec04b977d927c28815981511db78a37\": rpc error: code = NotFound desc = could not find container \"f61f7288551177fba804cc0b69018f73fec04b977d927c28815981511db78a37\": container with ID starting with f61f7288551177fba804cc0b69018f73fec04b977d927c28815981511db78a37 not found: ID does not exist" Oct 06 10:17:28 crc kubenswrapper[4824]: I1006 10:17:28.376213 4824 scope.go:117] "RemoveContainer" containerID="2355e113a1c344098469fe0db5bc6338b9e60c558e6de64c5e22eb3fd0b81b6c" Oct 06 10:17:28 crc kubenswrapper[4824]: E1006 10:17:28.377166 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2355e113a1c344098469fe0db5bc6338b9e60c558e6de64c5e22eb3fd0b81b6c\": container with ID starting with 2355e113a1c344098469fe0db5bc6338b9e60c558e6de64c5e22eb3fd0b81b6c not found: ID does not exist" containerID="2355e113a1c344098469fe0db5bc6338b9e60c558e6de64c5e22eb3fd0b81b6c" Oct 06 10:17:28 crc kubenswrapper[4824]: I1006 10:17:28.377207 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2355e113a1c344098469fe0db5bc6338b9e60c558e6de64c5e22eb3fd0b81b6c"} err="failed to get container status \"2355e113a1c344098469fe0db5bc6338b9e60c558e6de64c5e22eb3fd0b81b6c\": rpc error: code = NotFound desc = could not find container \"2355e113a1c344098469fe0db5bc6338b9e60c558e6de64c5e22eb3fd0b81b6c\": container with ID starting with 2355e113a1c344098469fe0db5bc6338b9e60c558e6de64c5e22eb3fd0b81b6c not found: ID does not exist" Oct 06 10:17:28 crc kubenswrapper[4824]: I1006 10:17:28.437856 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/76e8c268-3c97-4c7f-9f91-a06880778160-scripts\") pod \"ceilometer-0\" (UID: \"76e8c268-3c97-4c7f-9f91-a06880778160\") " pod="openstack/ceilometer-0" Oct 06 10:17:28 crc kubenswrapper[4824]: I1006 10:17:28.437922 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/76e8c268-3c97-4c7f-9f91-a06880778160-log-httpd\") pod \"ceilometer-0\" (UID: \"76e8c268-3c97-4c7f-9f91-a06880778160\") " pod="openstack/ceilometer-0" Oct 06 10:17:28 crc kubenswrapper[4824]: I1006 10:17:28.437999 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76e8c268-3c97-4c7f-9f91-a06880778160-config-data\") pod \"ceilometer-0\" (UID: \"76e8c268-3c97-4c7f-9f91-a06880778160\") " pod="openstack/ceilometer-0" Oct 06 10:17:28 crc kubenswrapper[4824]: I1006 10:17:28.438037 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/76e8c268-3c97-4c7f-9f91-a06880778160-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"76e8c268-3c97-4c7f-9f91-a06880778160\") " pod="openstack/ceilometer-0" Oct 06 10:17:28 crc kubenswrapper[4824]: I1006 10:17:28.438081 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76e8c268-3c97-4c7f-9f91-a06880778160-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"76e8c268-3c97-4c7f-9f91-a06880778160\") " pod="openstack/ceilometer-0" Oct 06 10:17:28 crc kubenswrapper[4824]: I1006 10:17:28.438103 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/76e8c268-3c97-4c7f-9f91-a06880778160-run-httpd\") pod \"ceilometer-0\" (UID: \"76e8c268-3c97-4c7f-9f91-a06880778160\") " pod="openstack/ceilometer-0" Oct 06 10:17:28 crc kubenswrapper[4824]: I1006 10:17:28.438221 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9jhb9\" (UniqueName: \"kubernetes.io/projected/76e8c268-3c97-4c7f-9f91-a06880778160-kube-api-access-9jhb9\") pod \"ceilometer-0\" (UID: \"76e8c268-3c97-4c7f-9f91-a06880778160\") " pod="openstack/ceilometer-0" Oct 06 10:17:28 crc kubenswrapper[4824]: I1006 10:17:28.540047 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76e8c268-3c97-4c7f-9f91-a06880778160-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"76e8c268-3c97-4c7f-9f91-a06880778160\") " pod="openstack/ceilometer-0" Oct 06 10:17:28 crc kubenswrapper[4824]: I1006 10:17:28.540317 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/76e8c268-3c97-4c7f-9f91-a06880778160-run-httpd\") pod \"ceilometer-0\" (UID: \"76e8c268-3c97-4c7f-9f91-a06880778160\") " pod="openstack/ceilometer-0" Oct 06 10:17:28 crc kubenswrapper[4824]: I1006 10:17:28.540504 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9jhb9\" (UniqueName: \"kubernetes.io/projected/76e8c268-3c97-4c7f-9f91-a06880778160-kube-api-access-9jhb9\") pod \"ceilometer-0\" (UID: \"76e8c268-3c97-4c7f-9f91-a06880778160\") " pod="openstack/ceilometer-0" Oct 06 10:17:28 crc kubenswrapper[4824]: I1006 10:17:28.540605 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/76e8c268-3c97-4c7f-9f91-a06880778160-scripts\") pod \"ceilometer-0\" (UID: \"76e8c268-3c97-4c7f-9f91-a06880778160\") " pod="openstack/ceilometer-0" Oct 06 10:17:28 crc kubenswrapper[4824]: I1006 10:17:28.540701 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/76e8c268-3c97-4c7f-9f91-a06880778160-log-httpd\") pod \"ceilometer-0\" (UID: \"76e8c268-3c97-4c7f-9f91-a06880778160\") " pod="openstack/ceilometer-0" Oct 06 10:17:28 crc kubenswrapper[4824]: I1006 10:17:28.540813 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76e8c268-3c97-4c7f-9f91-a06880778160-config-data\") pod \"ceilometer-0\" (UID: \"76e8c268-3c97-4c7f-9f91-a06880778160\") " pod="openstack/ceilometer-0" Oct 06 10:17:28 crc kubenswrapper[4824]: I1006 10:17:28.540924 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/76e8c268-3c97-4c7f-9f91-a06880778160-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"76e8c268-3c97-4c7f-9f91-a06880778160\") " pod="openstack/ceilometer-0" Oct 06 10:17:28 crc kubenswrapper[4824]: I1006 10:17:28.542005 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/76e8c268-3c97-4c7f-9f91-a06880778160-log-httpd\") pod \"ceilometer-0\" (UID: \"76e8c268-3c97-4c7f-9f91-a06880778160\") " pod="openstack/ceilometer-0" Oct 06 10:17:28 crc kubenswrapper[4824]: I1006 10:17:28.542403 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/76e8c268-3c97-4c7f-9f91-a06880778160-run-httpd\") pod \"ceilometer-0\" (UID: \"76e8c268-3c97-4c7f-9f91-a06880778160\") " pod="openstack/ceilometer-0" Oct 06 10:17:28 crc kubenswrapper[4824]: I1006 10:17:28.546308 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/76e8c268-3c97-4c7f-9f91-a06880778160-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"76e8c268-3c97-4c7f-9f91-a06880778160\") " pod="openstack/ceilometer-0" Oct 06 10:17:28 crc kubenswrapper[4824]: I1006 10:17:28.546647 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76e8c268-3c97-4c7f-9f91-a06880778160-config-data\") pod \"ceilometer-0\" (UID: \"76e8c268-3c97-4c7f-9f91-a06880778160\") " pod="openstack/ceilometer-0" Oct 06 10:17:28 crc kubenswrapper[4824]: I1006 10:17:28.547245 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/76e8c268-3c97-4c7f-9f91-a06880778160-scripts\") pod \"ceilometer-0\" (UID: \"76e8c268-3c97-4c7f-9f91-a06880778160\") " pod="openstack/ceilometer-0" Oct 06 10:17:28 crc kubenswrapper[4824]: I1006 10:17:28.549042 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76e8c268-3c97-4c7f-9f91-a06880778160-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"76e8c268-3c97-4c7f-9f91-a06880778160\") " pod="openstack/ceilometer-0" Oct 06 10:17:28 crc kubenswrapper[4824]: I1006 10:17:28.561768 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9jhb9\" (UniqueName: \"kubernetes.io/projected/76e8c268-3c97-4c7f-9f91-a06880778160-kube-api-access-9jhb9\") pod \"ceilometer-0\" (UID: \"76e8c268-3c97-4c7f-9f91-a06880778160\") " pod="openstack/ceilometer-0" Oct 06 10:17:28 crc kubenswrapper[4824]: I1006 10:17:28.677329 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 06 10:17:28 crc kubenswrapper[4824]: I1006 10:17:28.852535 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 06 10:17:29 crc kubenswrapper[4824]: I1006 10:17:29.155577 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 06 10:17:29 crc kubenswrapper[4824]: W1006 10:17:29.158821 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod76e8c268_3c97_4c7f_9f91_a06880778160.slice/crio-68b2610b541b3e3f0020d4469973472a80e816f403ec3ac561c0c9c851112a45 WatchSource:0}: Error finding container 68b2610b541b3e3f0020d4469973472a80e816f403ec3ac561c0c9c851112a45: Status 404 returned error can't find the container with id 68b2610b541b3e3f0020d4469973472a80e816f403ec3ac561c0c9c851112a45 Oct 06 10:17:29 crc kubenswrapper[4824]: I1006 10:17:29.263429 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"76e8c268-3c97-4c7f-9f91-a06880778160","Type":"ContainerStarted","Data":"68b2610b541b3e3f0020d4469973472a80e816f403ec3ac561c0c9c851112a45"} Oct 06 10:17:29 crc kubenswrapper[4824]: I1006 10:17:29.286684 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="11aa77f4-266f-492a-a859-147f00900e7d" path="/var/lib/kubelet/pods/11aa77f4-266f-492a-a859-147f00900e7d/volumes" Oct 06 10:17:30 crc kubenswrapper[4824]: I1006 10:17:30.277929 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"76e8c268-3c97-4c7f-9f91-a06880778160","Type":"ContainerStarted","Data":"144d9cca53691b88b9390df3fd5c1a2fccd54826c76cd61650cbaa3aca0bc2a2"} Oct 06 10:17:30 crc kubenswrapper[4824]: I1006 10:17:30.872540 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-c73c-account-create-q9hfk"] Oct 06 10:17:30 crc kubenswrapper[4824]: I1006 10:17:30.874049 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-c73c-account-create-q9hfk" Oct 06 10:17:30 crc kubenswrapper[4824]: I1006 10:17:30.876990 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Oct 06 10:17:30 crc kubenswrapper[4824]: I1006 10:17:30.893144 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-c73c-account-create-q9hfk"] Oct 06 10:17:30 crc kubenswrapper[4824]: I1006 10:17:30.988039 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4qm65\" (UniqueName: \"kubernetes.io/projected/4007653a-5c39-4285-8ba3-e0b3caed5640-kube-api-access-4qm65\") pod \"nova-api-c73c-account-create-q9hfk\" (UID: \"4007653a-5c39-4285-8ba3-e0b3caed5640\") " pod="openstack/nova-api-c73c-account-create-q9hfk" Oct 06 10:17:31 crc kubenswrapper[4824]: I1006 10:17:31.078424 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-e2d6-account-create-rxhwz"] Oct 06 10:17:31 crc kubenswrapper[4824]: I1006 10:17:31.079497 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-e2d6-account-create-rxhwz" Oct 06 10:17:31 crc kubenswrapper[4824]: I1006 10:17:31.086103 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Oct 06 10:17:31 crc kubenswrapper[4824]: I1006 10:17:31.090165 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4qm65\" (UniqueName: \"kubernetes.io/projected/4007653a-5c39-4285-8ba3-e0b3caed5640-kube-api-access-4qm65\") pod \"nova-api-c73c-account-create-q9hfk\" (UID: \"4007653a-5c39-4285-8ba3-e0b3caed5640\") " pod="openstack/nova-api-c73c-account-create-q9hfk" Oct 06 10:17:31 crc kubenswrapper[4824]: I1006 10:17:31.094282 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-e2d6-account-create-rxhwz"] Oct 06 10:17:31 crc kubenswrapper[4824]: I1006 10:17:31.113507 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4qm65\" (UniqueName: \"kubernetes.io/projected/4007653a-5c39-4285-8ba3-e0b3caed5640-kube-api-access-4qm65\") pod \"nova-api-c73c-account-create-q9hfk\" (UID: \"4007653a-5c39-4285-8ba3-e0b3caed5640\") " pod="openstack/nova-api-c73c-account-create-q9hfk" Oct 06 10:17:31 crc kubenswrapper[4824]: I1006 10:17:31.191734 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g4kff\" (UniqueName: \"kubernetes.io/projected/7dd053a5-58ba-4f1c-9d32-e639fa8b1450-kube-api-access-g4kff\") pod \"nova-cell0-e2d6-account-create-rxhwz\" (UID: \"7dd053a5-58ba-4f1c-9d32-e639fa8b1450\") " pod="openstack/nova-cell0-e2d6-account-create-rxhwz" Oct 06 10:17:31 crc kubenswrapper[4824]: I1006 10:17:31.201575 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-c73c-account-create-q9hfk" Oct 06 10:17:31 crc kubenswrapper[4824]: I1006 10:17:31.294591 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g4kff\" (UniqueName: \"kubernetes.io/projected/7dd053a5-58ba-4f1c-9d32-e639fa8b1450-kube-api-access-g4kff\") pod \"nova-cell0-e2d6-account-create-rxhwz\" (UID: \"7dd053a5-58ba-4f1c-9d32-e639fa8b1450\") " pod="openstack/nova-cell0-e2d6-account-create-rxhwz" Oct 06 10:17:31 crc kubenswrapper[4824]: I1006 10:17:31.316691 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-aa65-account-create-l9c7p"] Oct 06 10:17:31 crc kubenswrapper[4824]: I1006 10:17:31.318402 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-aa65-account-create-l9c7p"] Oct 06 10:17:31 crc kubenswrapper[4824]: I1006 10:17:31.318521 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"76e8c268-3c97-4c7f-9f91-a06880778160","Type":"ContainerStarted","Data":"1fd9058c49573e736f1913745838a146ff0a052b9f51b8079a33eb82aea18a30"} Oct 06 10:17:31 crc kubenswrapper[4824]: I1006 10:17:31.318691 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-aa65-account-create-l9c7p" Oct 06 10:17:31 crc kubenswrapper[4824]: I1006 10:17:31.321906 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Oct 06 10:17:31 crc kubenswrapper[4824]: I1006 10:17:31.325317 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g4kff\" (UniqueName: \"kubernetes.io/projected/7dd053a5-58ba-4f1c-9d32-e639fa8b1450-kube-api-access-g4kff\") pod \"nova-cell0-e2d6-account-create-rxhwz\" (UID: \"7dd053a5-58ba-4f1c-9d32-e639fa8b1450\") " pod="openstack/nova-cell0-e2d6-account-create-rxhwz" Oct 06 10:17:31 crc kubenswrapper[4824]: I1006 10:17:31.334025 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 06 10:17:31 crc kubenswrapper[4824]: I1006 10:17:31.334410 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="508d58be-fcaa-4360-ab57-fbaedb728761" containerName="glance-log" containerID="cri-o://2bf864c0a9c8f342accc699a7313528172c1c81ba4c82a5d1cab8dbbfdbe803b" gracePeriod=30 Oct 06 10:17:31 crc kubenswrapper[4824]: I1006 10:17:31.334669 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="508d58be-fcaa-4360-ab57-fbaedb728761" containerName="glance-httpd" containerID="cri-o://e0cc57ebccd50c51a2c1ef0eccf9882e4f8ce8ab483e28f660cc3be288f38f95" gracePeriod=30 Oct 06 10:17:31 crc kubenswrapper[4824]: I1006 10:17:31.399189 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kx7zt\" (UniqueName: \"kubernetes.io/projected/7cbd2d79-2d35-4954-9f61-c9bbbbe9d84e-kube-api-access-kx7zt\") pod \"nova-cell1-aa65-account-create-l9c7p\" (UID: \"7cbd2d79-2d35-4954-9f61-c9bbbbe9d84e\") " pod="openstack/nova-cell1-aa65-account-create-l9c7p" Oct 06 10:17:31 crc kubenswrapper[4824]: I1006 10:17:31.492908 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-e2d6-account-create-rxhwz" Oct 06 10:17:31 crc kubenswrapper[4824]: I1006 10:17:31.501737 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kx7zt\" (UniqueName: \"kubernetes.io/projected/7cbd2d79-2d35-4954-9f61-c9bbbbe9d84e-kube-api-access-kx7zt\") pod \"nova-cell1-aa65-account-create-l9c7p\" (UID: \"7cbd2d79-2d35-4954-9f61-c9bbbbe9d84e\") " pod="openstack/nova-cell1-aa65-account-create-l9c7p" Oct 06 10:17:31 crc kubenswrapper[4824]: I1006 10:17:31.529036 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kx7zt\" (UniqueName: \"kubernetes.io/projected/7cbd2d79-2d35-4954-9f61-c9bbbbe9d84e-kube-api-access-kx7zt\") pod \"nova-cell1-aa65-account-create-l9c7p\" (UID: \"7cbd2d79-2d35-4954-9f61-c9bbbbe9d84e\") " pod="openstack/nova-cell1-aa65-account-create-l9c7p" Oct 06 10:17:31 crc kubenswrapper[4824]: I1006 10:17:31.674400 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-aa65-account-create-l9c7p" Oct 06 10:17:31 crc kubenswrapper[4824]: I1006 10:17:31.857269 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-c73c-account-create-q9hfk"] Oct 06 10:17:32 crc kubenswrapper[4824]: I1006 10:17:32.036057 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-e2d6-account-create-rxhwz"] Oct 06 10:17:32 crc kubenswrapper[4824]: I1006 10:17:32.265965 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-aa65-account-create-l9c7p"] Oct 06 10:17:32 crc kubenswrapper[4824]: W1006 10:17:32.268823 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7cbd2d79_2d35_4954_9f61_c9bbbbe9d84e.slice/crio-583508f60f2e08464dcdd048b7be2ce337a61fa9ff09c8c4f71db470f9698a8d WatchSource:0}: Error finding container 583508f60f2e08464dcdd048b7be2ce337a61fa9ff09c8c4f71db470f9698a8d: Status 404 returned error can't find the container with id 583508f60f2e08464dcdd048b7be2ce337a61fa9ff09c8c4f71db470f9698a8d Oct 06 10:17:32 crc kubenswrapper[4824]: I1006 10:17:32.332537 4824 generic.go:334] "Generic (PLEG): container finished" podID="508d58be-fcaa-4360-ab57-fbaedb728761" containerID="2bf864c0a9c8f342accc699a7313528172c1c81ba4c82a5d1cab8dbbfdbe803b" exitCode=143 Oct 06 10:17:32 crc kubenswrapper[4824]: I1006 10:17:32.332622 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"508d58be-fcaa-4360-ab57-fbaedb728761","Type":"ContainerDied","Data":"2bf864c0a9c8f342accc699a7313528172c1c81ba4c82a5d1cab8dbbfdbe803b"} Oct 06 10:17:32 crc kubenswrapper[4824]: I1006 10:17:32.334066 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-aa65-account-create-l9c7p" event={"ID":"7cbd2d79-2d35-4954-9f61-c9bbbbe9d84e","Type":"ContainerStarted","Data":"583508f60f2e08464dcdd048b7be2ce337a61fa9ff09c8c4f71db470f9698a8d"} Oct 06 10:17:32 crc kubenswrapper[4824]: I1006 10:17:32.335871 4824 generic.go:334] "Generic (PLEG): container finished" podID="4007653a-5c39-4285-8ba3-e0b3caed5640" containerID="ad61d4bd99222d6b606f4e3eb3c1f5d55d69b5ac369233c8514fa1400dca4650" exitCode=0 Oct 06 10:17:32 crc kubenswrapper[4824]: I1006 10:17:32.336120 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-c73c-account-create-q9hfk" event={"ID":"4007653a-5c39-4285-8ba3-e0b3caed5640","Type":"ContainerDied","Data":"ad61d4bd99222d6b606f4e3eb3c1f5d55d69b5ac369233c8514fa1400dca4650"} Oct 06 10:17:32 crc kubenswrapper[4824]: I1006 10:17:32.336176 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-c73c-account-create-q9hfk" event={"ID":"4007653a-5c39-4285-8ba3-e0b3caed5640","Type":"ContainerStarted","Data":"f2c5125706078c13f3f745db9ec786c875c926f1721be643e2570ef6155b2e5f"} Oct 06 10:17:32 crc kubenswrapper[4824]: I1006 10:17:32.340138 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"76e8c268-3c97-4c7f-9f91-a06880778160","Type":"ContainerStarted","Data":"8a525b236722cb1e2fa01fb10a0cad64293948d71e21f9d1d07cc94ee0acd59a"} Oct 06 10:17:32 crc kubenswrapper[4824]: I1006 10:17:32.341825 4824 generic.go:334] "Generic (PLEG): container finished" podID="7dd053a5-58ba-4f1c-9d32-e639fa8b1450" containerID="aed46891458e2473b16dd98b74441f192fc6315b49f05c1ec0728035643b5019" exitCode=0 Oct 06 10:17:32 crc kubenswrapper[4824]: I1006 10:17:32.341895 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-e2d6-account-create-rxhwz" event={"ID":"7dd053a5-58ba-4f1c-9d32-e639fa8b1450","Type":"ContainerDied","Data":"aed46891458e2473b16dd98b74441f192fc6315b49f05c1ec0728035643b5019"} Oct 06 10:17:32 crc kubenswrapper[4824]: I1006 10:17:32.341933 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-e2d6-account-create-rxhwz" event={"ID":"7dd053a5-58ba-4f1c-9d32-e639fa8b1450","Type":"ContainerStarted","Data":"fb31834acde3801408d0e81f9818caa328d4bb034c37b365041baf9b77845b79"} Oct 06 10:17:32 crc kubenswrapper[4824]: I1006 10:17:32.527433 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 06 10:17:32 crc kubenswrapper[4824]: I1006 10:17:32.527804 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="a85f7f99-c70b-40a6-8ea2-839d91a53407" containerName="glance-log" containerID="cri-o://0a813108b412d32c6f3f573b57157b9dca2f4c7447b82953ab11535213306d0a" gracePeriod=30 Oct 06 10:17:32 crc kubenswrapper[4824]: I1006 10:17:32.527896 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="a85f7f99-c70b-40a6-8ea2-839d91a53407" containerName="glance-httpd" containerID="cri-o://866e8b22170a7926f5516f9626a0711f498d48e51135ad184d51704db33d88a6" gracePeriod=30 Oct 06 10:17:33 crc kubenswrapper[4824]: I1006 10:17:33.357822 4824 generic.go:334] "Generic (PLEG): container finished" podID="a85f7f99-c70b-40a6-8ea2-839d91a53407" containerID="0a813108b412d32c6f3f573b57157b9dca2f4c7447b82953ab11535213306d0a" exitCode=143 Oct 06 10:17:33 crc kubenswrapper[4824]: I1006 10:17:33.358248 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"a85f7f99-c70b-40a6-8ea2-839d91a53407","Type":"ContainerDied","Data":"0a813108b412d32c6f3f573b57157b9dca2f4c7447b82953ab11535213306d0a"} Oct 06 10:17:33 crc kubenswrapper[4824]: I1006 10:17:33.366247 4824 generic.go:334] "Generic (PLEG): container finished" podID="7cbd2d79-2d35-4954-9f61-c9bbbbe9d84e" containerID="380cdc21f47ff79047047cde2a74671a6cd209a886c1a0dc68647e0acebb1e32" exitCode=0 Oct 06 10:17:33 crc kubenswrapper[4824]: I1006 10:17:33.366324 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-aa65-account-create-l9c7p" event={"ID":"7cbd2d79-2d35-4954-9f61-c9bbbbe9d84e","Type":"ContainerDied","Data":"380cdc21f47ff79047047cde2a74671a6cd209a886c1a0dc68647e0acebb1e32"} Oct 06 10:17:33 crc kubenswrapper[4824]: I1006 10:17:33.371155 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="76e8c268-3c97-4c7f-9f91-a06880778160" containerName="ceilometer-central-agent" containerID="cri-o://144d9cca53691b88b9390df3fd5c1a2fccd54826c76cd61650cbaa3aca0bc2a2" gracePeriod=30 Oct 06 10:17:33 crc kubenswrapper[4824]: I1006 10:17:33.371378 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"76e8c268-3c97-4c7f-9f91-a06880778160","Type":"ContainerStarted","Data":"2a5c65dc4113f85362f010990c9335a88b385f42366cef3a900ae692d8a15729"} Oct 06 10:17:33 crc kubenswrapper[4824]: I1006 10:17:33.371514 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 06 10:17:33 crc kubenswrapper[4824]: I1006 10:17:33.371565 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="76e8c268-3c97-4c7f-9f91-a06880778160" containerName="proxy-httpd" containerID="cri-o://2a5c65dc4113f85362f010990c9335a88b385f42366cef3a900ae692d8a15729" gracePeriod=30 Oct 06 10:17:33 crc kubenswrapper[4824]: I1006 10:17:33.371610 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="76e8c268-3c97-4c7f-9f91-a06880778160" containerName="sg-core" containerID="cri-o://8a525b236722cb1e2fa01fb10a0cad64293948d71e21f9d1d07cc94ee0acd59a" gracePeriod=30 Oct 06 10:17:33 crc kubenswrapper[4824]: I1006 10:17:33.371654 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="76e8c268-3c97-4c7f-9f91-a06880778160" containerName="ceilometer-notification-agent" containerID="cri-o://1fd9058c49573e736f1913745838a146ff0a052b9f51b8079a33eb82aea18a30" gracePeriod=30 Oct 06 10:17:33 crc kubenswrapper[4824]: I1006 10:17:33.831506 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-e2d6-account-create-rxhwz" Oct 06 10:17:33 crc kubenswrapper[4824]: I1006 10:17:33.841889 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-c73c-account-create-q9hfk" Oct 06 10:17:33 crc kubenswrapper[4824]: I1006 10:17:33.857208 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.439208064 podStartE2EDuration="5.857189573s" podCreationTimestamp="2025-10-06 10:17:28 +0000 UTC" firstStartedPulling="2025-10-06 10:17:29.161294472 +0000 UTC m=+1158.525717333" lastFinishedPulling="2025-10-06 10:17:32.579275981 +0000 UTC m=+1161.943698842" observedRunningTime="2025-10-06 10:17:33.435729297 +0000 UTC m=+1162.800152158" watchObservedRunningTime="2025-10-06 10:17:33.857189573 +0000 UTC m=+1163.221612444" Oct 06 10:17:33 crc kubenswrapper[4824]: I1006 10:17:33.966804 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g4kff\" (UniqueName: \"kubernetes.io/projected/7dd053a5-58ba-4f1c-9d32-e639fa8b1450-kube-api-access-g4kff\") pod \"7dd053a5-58ba-4f1c-9d32-e639fa8b1450\" (UID: \"7dd053a5-58ba-4f1c-9d32-e639fa8b1450\") " Oct 06 10:17:33 crc kubenswrapper[4824]: I1006 10:17:33.966896 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4qm65\" (UniqueName: \"kubernetes.io/projected/4007653a-5c39-4285-8ba3-e0b3caed5640-kube-api-access-4qm65\") pod \"4007653a-5c39-4285-8ba3-e0b3caed5640\" (UID: \"4007653a-5c39-4285-8ba3-e0b3caed5640\") " Oct 06 10:17:33 crc kubenswrapper[4824]: I1006 10:17:33.975864 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7dd053a5-58ba-4f1c-9d32-e639fa8b1450-kube-api-access-g4kff" (OuterVolumeSpecName: "kube-api-access-g4kff") pod "7dd053a5-58ba-4f1c-9d32-e639fa8b1450" (UID: "7dd053a5-58ba-4f1c-9d32-e639fa8b1450"). InnerVolumeSpecName "kube-api-access-g4kff". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:17:33 crc kubenswrapper[4824]: I1006 10:17:33.976521 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4007653a-5c39-4285-8ba3-e0b3caed5640-kube-api-access-4qm65" (OuterVolumeSpecName: "kube-api-access-4qm65") pod "4007653a-5c39-4285-8ba3-e0b3caed5640" (UID: "4007653a-5c39-4285-8ba3-e0b3caed5640"). InnerVolumeSpecName "kube-api-access-4qm65". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:17:34 crc kubenswrapper[4824]: I1006 10:17:34.069240 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g4kff\" (UniqueName: \"kubernetes.io/projected/7dd053a5-58ba-4f1c-9d32-e639fa8b1450-kube-api-access-g4kff\") on node \"crc\" DevicePath \"\"" Oct 06 10:17:34 crc kubenswrapper[4824]: I1006 10:17:34.069275 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4qm65\" (UniqueName: \"kubernetes.io/projected/4007653a-5c39-4285-8ba3-e0b3caed5640-kube-api-access-4qm65\") on node \"crc\" DevicePath \"\"" Oct 06 10:17:34 crc kubenswrapper[4824]: I1006 10:17:34.381887 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-c73c-account-create-q9hfk" Oct 06 10:17:34 crc kubenswrapper[4824]: I1006 10:17:34.382387 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-c73c-account-create-q9hfk" event={"ID":"4007653a-5c39-4285-8ba3-e0b3caed5640","Type":"ContainerDied","Data":"f2c5125706078c13f3f745db9ec786c875c926f1721be643e2570ef6155b2e5f"} Oct 06 10:17:34 crc kubenswrapper[4824]: I1006 10:17:34.382425 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f2c5125706078c13f3f745db9ec786c875c926f1721be643e2570ef6155b2e5f" Oct 06 10:17:34 crc kubenswrapper[4824]: I1006 10:17:34.386074 4824 generic.go:334] "Generic (PLEG): container finished" podID="76e8c268-3c97-4c7f-9f91-a06880778160" containerID="2a5c65dc4113f85362f010990c9335a88b385f42366cef3a900ae692d8a15729" exitCode=0 Oct 06 10:17:34 crc kubenswrapper[4824]: I1006 10:17:34.386113 4824 generic.go:334] "Generic (PLEG): container finished" podID="76e8c268-3c97-4c7f-9f91-a06880778160" containerID="8a525b236722cb1e2fa01fb10a0cad64293948d71e21f9d1d07cc94ee0acd59a" exitCode=2 Oct 06 10:17:34 crc kubenswrapper[4824]: I1006 10:17:34.386126 4824 generic.go:334] "Generic (PLEG): container finished" podID="76e8c268-3c97-4c7f-9f91-a06880778160" containerID="1fd9058c49573e736f1913745838a146ff0a052b9f51b8079a33eb82aea18a30" exitCode=0 Oct 06 10:17:34 crc kubenswrapper[4824]: I1006 10:17:34.386157 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"76e8c268-3c97-4c7f-9f91-a06880778160","Type":"ContainerDied","Data":"2a5c65dc4113f85362f010990c9335a88b385f42366cef3a900ae692d8a15729"} Oct 06 10:17:34 crc kubenswrapper[4824]: I1006 10:17:34.386207 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"76e8c268-3c97-4c7f-9f91-a06880778160","Type":"ContainerDied","Data":"8a525b236722cb1e2fa01fb10a0cad64293948d71e21f9d1d07cc94ee0acd59a"} Oct 06 10:17:34 crc kubenswrapper[4824]: I1006 10:17:34.386227 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"76e8c268-3c97-4c7f-9f91-a06880778160","Type":"ContainerDied","Data":"1fd9058c49573e736f1913745838a146ff0a052b9f51b8079a33eb82aea18a30"} Oct 06 10:17:34 crc kubenswrapper[4824]: I1006 10:17:34.387997 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-e2d6-account-create-rxhwz" Oct 06 10:17:34 crc kubenswrapper[4824]: I1006 10:17:34.387969 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-e2d6-account-create-rxhwz" event={"ID":"7dd053a5-58ba-4f1c-9d32-e639fa8b1450","Type":"ContainerDied","Data":"fb31834acde3801408d0e81f9818caa328d4bb034c37b365041baf9b77845b79"} Oct 06 10:17:34 crc kubenswrapper[4824]: I1006 10:17:34.388057 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fb31834acde3801408d0e81f9818caa328d4bb034c37b365041baf9b77845b79" Oct 06 10:17:34 crc kubenswrapper[4824]: I1006 10:17:34.751612 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-aa65-account-create-l9c7p" Oct 06 10:17:34 crc kubenswrapper[4824]: I1006 10:17:34.898715 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kx7zt\" (UniqueName: \"kubernetes.io/projected/7cbd2d79-2d35-4954-9f61-c9bbbbe9d84e-kube-api-access-kx7zt\") pod \"7cbd2d79-2d35-4954-9f61-c9bbbbe9d84e\" (UID: \"7cbd2d79-2d35-4954-9f61-c9bbbbe9d84e\") " Oct 06 10:17:34 crc kubenswrapper[4824]: I1006 10:17:34.922471 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7cbd2d79-2d35-4954-9f61-c9bbbbe9d84e-kube-api-access-kx7zt" (OuterVolumeSpecName: "kube-api-access-kx7zt") pod "7cbd2d79-2d35-4954-9f61-c9bbbbe9d84e" (UID: "7cbd2d79-2d35-4954-9f61-c9bbbbe9d84e"). InnerVolumeSpecName "kube-api-access-kx7zt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:17:35 crc kubenswrapper[4824]: I1006 10:17:35.001081 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kx7zt\" (UniqueName: \"kubernetes.io/projected/7cbd2d79-2d35-4954-9f61-c9bbbbe9d84e-kube-api-access-kx7zt\") on node \"crc\" DevicePath \"\"" Oct 06 10:17:35 crc kubenswrapper[4824]: I1006 10:17:35.084398 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 06 10:17:35 crc kubenswrapper[4824]: I1006 10:17:35.204348 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/508d58be-fcaa-4360-ab57-fbaedb728761-config-data\") pod \"508d58be-fcaa-4360-ab57-fbaedb728761\" (UID: \"508d58be-fcaa-4360-ab57-fbaedb728761\") " Oct 06 10:17:35 crc kubenswrapper[4824]: I1006 10:17:35.204467 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/508d58be-fcaa-4360-ab57-fbaedb728761-public-tls-certs\") pod \"508d58be-fcaa-4360-ab57-fbaedb728761\" (UID: \"508d58be-fcaa-4360-ab57-fbaedb728761\") " Oct 06 10:17:35 crc kubenswrapper[4824]: I1006 10:17:35.204588 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"508d58be-fcaa-4360-ab57-fbaedb728761\" (UID: \"508d58be-fcaa-4360-ab57-fbaedb728761\") " Oct 06 10:17:35 crc kubenswrapper[4824]: I1006 10:17:35.204692 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/508d58be-fcaa-4360-ab57-fbaedb728761-scripts\") pod \"508d58be-fcaa-4360-ab57-fbaedb728761\" (UID: \"508d58be-fcaa-4360-ab57-fbaedb728761\") " Oct 06 10:17:35 crc kubenswrapper[4824]: I1006 10:17:35.204741 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f5hbv\" (UniqueName: \"kubernetes.io/projected/508d58be-fcaa-4360-ab57-fbaedb728761-kube-api-access-f5hbv\") pod \"508d58be-fcaa-4360-ab57-fbaedb728761\" (UID: \"508d58be-fcaa-4360-ab57-fbaedb728761\") " Oct 06 10:17:35 crc kubenswrapper[4824]: I1006 10:17:35.204815 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/508d58be-fcaa-4360-ab57-fbaedb728761-combined-ca-bundle\") pod \"508d58be-fcaa-4360-ab57-fbaedb728761\" (UID: \"508d58be-fcaa-4360-ab57-fbaedb728761\") " Oct 06 10:17:35 crc kubenswrapper[4824]: I1006 10:17:35.204843 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/508d58be-fcaa-4360-ab57-fbaedb728761-httpd-run\") pod \"508d58be-fcaa-4360-ab57-fbaedb728761\" (UID: \"508d58be-fcaa-4360-ab57-fbaedb728761\") " Oct 06 10:17:35 crc kubenswrapper[4824]: I1006 10:17:35.204877 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/508d58be-fcaa-4360-ab57-fbaedb728761-logs\") pod \"508d58be-fcaa-4360-ab57-fbaedb728761\" (UID: \"508d58be-fcaa-4360-ab57-fbaedb728761\") " Oct 06 10:17:35 crc kubenswrapper[4824]: I1006 10:17:35.205934 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/508d58be-fcaa-4360-ab57-fbaedb728761-logs" (OuterVolumeSpecName: "logs") pod "508d58be-fcaa-4360-ab57-fbaedb728761" (UID: "508d58be-fcaa-4360-ab57-fbaedb728761"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:17:35 crc kubenswrapper[4824]: I1006 10:17:35.207010 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/508d58be-fcaa-4360-ab57-fbaedb728761-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "508d58be-fcaa-4360-ab57-fbaedb728761" (UID: "508d58be-fcaa-4360-ab57-fbaedb728761"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:17:35 crc kubenswrapper[4824]: I1006 10:17:35.212005 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/508d58be-fcaa-4360-ab57-fbaedb728761-kube-api-access-f5hbv" (OuterVolumeSpecName: "kube-api-access-f5hbv") pod "508d58be-fcaa-4360-ab57-fbaedb728761" (UID: "508d58be-fcaa-4360-ab57-fbaedb728761"). InnerVolumeSpecName "kube-api-access-f5hbv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:17:35 crc kubenswrapper[4824]: I1006 10:17:35.212503 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/508d58be-fcaa-4360-ab57-fbaedb728761-scripts" (OuterVolumeSpecName: "scripts") pod "508d58be-fcaa-4360-ab57-fbaedb728761" (UID: "508d58be-fcaa-4360-ab57-fbaedb728761"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:17:35 crc kubenswrapper[4824]: I1006 10:17:35.212683 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "glance") pod "508d58be-fcaa-4360-ab57-fbaedb728761" (UID: "508d58be-fcaa-4360-ab57-fbaedb728761"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 06 10:17:35 crc kubenswrapper[4824]: I1006 10:17:35.258411 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/508d58be-fcaa-4360-ab57-fbaedb728761-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "508d58be-fcaa-4360-ab57-fbaedb728761" (UID: "508d58be-fcaa-4360-ab57-fbaedb728761"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:17:35 crc kubenswrapper[4824]: I1006 10:17:35.269305 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/508d58be-fcaa-4360-ab57-fbaedb728761-config-data" (OuterVolumeSpecName: "config-data") pod "508d58be-fcaa-4360-ab57-fbaedb728761" (UID: "508d58be-fcaa-4360-ab57-fbaedb728761"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:17:35 crc kubenswrapper[4824]: I1006 10:17:35.298303 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/508d58be-fcaa-4360-ab57-fbaedb728761-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "508d58be-fcaa-4360-ab57-fbaedb728761" (UID: "508d58be-fcaa-4360-ab57-fbaedb728761"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:17:35 crc kubenswrapper[4824]: I1006 10:17:35.321660 4824 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Oct 06 10:17:35 crc kubenswrapper[4824]: I1006 10:17:35.327438 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/508d58be-fcaa-4360-ab57-fbaedb728761-scripts\") on node \"crc\" DevicePath \"\"" Oct 06 10:17:35 crc kubenswrapper[4824]: I1006 10:17:35.327945 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f5hbv\" (UniqueName: \"kubernetes.io/projected/508d58be-fcaa-4360-ab57-fbaedb728761-kube-api-access-f5hbv\") on node \"crc\" DevicePath \"\"" Oct 06 10:17:35 crc kubenswrapper[4824]: I1006 10:17:35.328867 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/508d58be-fcaa-4360-ab57-fbaedb728761-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 10:17:35 crc kubenswrapper[4824]: I1006 10:17:35.328902 4824 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/508d58be-fcaa-4360-ab57-fbaedb728761-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 06 10:17:35 crc kubenswrapper[4824]: I1006 10:17:35.328927 4824 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/508d58be-fcaa-4360-ab57-fbaedb728761-logs\") on node \"crc\" DevicePath \"\"" Oct 06 10:17:35 crc kubenswrapper[4824]: I1006 10:17:35.328957 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/508d58be-fcaa-4360-ab57-fbaedb728761-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 10:17:35 crc kubenswrapper[4824]: I1006 10:17:35.329217 4824 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/508d58be-fcaa-4360-ab57-fbaedb728761-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 06 10:17:35 crc kubenswrapper[4824]: I1006 10:17:35.351182 4824 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Oct 06 10:17:35 crc kubenswrapper[4824]: I1006 10:17:35.398616 4824 generic.go:334] "Generic (PLEG): container finished" podID="508d58be-fcaa-4360-ab57-fbaedb728761" containerID="e0cc57ebccd50c51a2c1ef0eccf9882e4f8ce8ab483e28f660cc3be288f38f95" exitCode=0 Oct 06 10:17:35 crc kubenswrapper[4824]: I1006 10:17:35.398806 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"508d58be-fcaa-4360-ab57-fbaedb728761","Type":"ContainerDied","Data":"e0cc57ebccd50c51a2c1ef0eccf9882e4f8ce8ab483e28f660cc3be288f38f95"} Oct 06 10:17:35 crc kubenswrapper[4824]: I1006 10:17:35.398850 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"508d58be-fcaa-4360-ab57-fbaedb728761","Type":"ContainerDied","Data":"66a2d1202ba3ab9a6dcb54d2034450460df4d403cba38005cc0584e842b32557"} Oct 06 10:17:35 crc kubenswrapper[4824]: I1006 10:17:35.398873 4824 scope.go:117] "RemoveContainer" containerID="e0cc57ebccd50c51a2c1ef0eccf9882e4f8ce8ab483e28f660cc3be288f38f95" Oct 06 10:17:35 crc kubenswrapper[4824]: I1006 10:17:35.398910 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 06 10:17:35 crc kubenswrapper[4824]: I1006 10:17:35.404351 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-aa65-account-create-l9c7p" event={"ID":"7cbd2d79-2d35-4954-9f61-c9bbbbe9d84e","Type":"ContainerDied","Data":"583508f60f2e08464dcdd048b7be2ce337a61fa9ff09c8c4f71db470f9698a8d"} Oct 06 10:17:35 crc kubenswrapper[4824]: I1006 10:17:35.404399 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="583508f60f2e08464dcdd048b7be2ce337a61fa9ff09c8c4f71db470f9698a8d" Oct 06 10:17:35 crc kubenswrapper[4824]: I1006 10:17:35.404454 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-aa65-account-create-l9c7p" Oct 06 10:17:35 crc kubenswrapper[4824]: I1006 10:17:35.426821 4824 scope.go:117] "RemoveContainer" containerID="2bf864c0a9c8f342accc699a7313528172c1c81ba4c82a5d1cab8dbbfdbe803b" Oct 06 10:17:35 crc kubenswrapper[4824]: I1006 10:17:35.431854 4824 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Oct 06 10:17:35 crc kubenswrapper[4824]: I1006 10:17:35.446363 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 06 10:17:35 crc kubenswrapper[4824]: I1006 10:17:35.454546 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 06 10:17:35 crc kubenswrapper[4824]: I1006 10:17:35.462211 4824 scope.go:117] "RemoveContainer" containerID="e0cc57ebccd50c51a2c1ef0eccf9882e4f8ce8ab483e28f660cc3be288f38f95" Oct 06 10:17:35 crc kubenswrapper[4824]: E1006 10:17:35.462963 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e0cc57ebccd50c51a2c1ef0eccf9882e4f8ce8ab483e28f660cc3be288f38f95\": container with ID starting with e0cc57ebccd50c51a2c1ef0eccf9882e4f8ce8ab483e28f660cc3be288f38f95 not found: ID does not exist" containerID="e0cc57ebccd50c51a2c1ef0eccf9882e4f8ce8ab483e28f660cc3be288f38f95" Oct 06 10:17:35 crc kubenswrapper[4824]: I1006 10:17:35.463012 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e0cc57ebccd50c51a2c1ef0eccf9882e4f8ce8ab483e28f660cc3be288f38f95"} err="failed to get container status \"e0cc57ebccd50c51a2c1ef0eccf9882e4f8ce8ab483e28f660cc3be288f38f95\": rpc error: code = NotFound desc = could not find container \"e0cc57ebccd50c51a2c1ef0eccf9882e4f8ce8ab483e28f660cc3be288f38f95\": container with ID starting with e0cc57ebccd50c51a2c1ef0eccf9882e4f8ce8ab483e28f660cc3be288f38f95 not found: ID does not exist" Oct 06 10:17:35 crc kubenswrapper[4824]: I1006 10:17:35.463035 4824 scope.go:117] "RemoveContainer" containerID="2bf864c0a9c8f342accc699a7313528172c1c81ba4c82a5d1cab8dbbfdbe803b" Oct 06 10:17:35 crc kubenswrapper[4824]: E1006 10:17:35.464764 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2bf864c0a9c8f342accc699a7313528172c1c81ba4c82a5d1cab8dbbfdbe803b\": container with ID starting with 2bf864c0a9c8f342accc699a7313528172c1c81ba4c82a5d1cab8dbbfdbe803b not found: ID does not exist" containerID="2bf864c0a9c8f342accc699a7313528172c1c81ba4c82a5d1cab8dbbfdbe803b" Oct 06 10:17:35 crc kubenswrapper[4824]: I1006 10:17:35.464830 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2bf864c0a9c8f342accc699a7313528172c1c81ba4c82a5d1cab8dbbfdbe803b"} err="failed to get container status \"2bf864c0a9c8f342accc699a7313528172c1c81ba4c82a5d1cab8dbbfdbe803b\": rpc error: code = NotFound desc = could not find container \"2bf864c0a9c8f342accc699a7313528172c1c81ba4c82a5d1cab8dbbfdbe803b\": container with ID starting with 2bf864c0a9c8f342accc699a7313528172c1c81ba4c82a5d1cab8dbbfdbe803b not found: ID does not exist" Oct 06 10:17:35 crc kubenswrapper[4824]: I1006 10:17:35.483499 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 06 10:17:35 crc kubenswrapper[4824]: E1006 10:17:35.488667 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="508d58be-fcaa-4360-ab57-fbaedb728761" containerName="glance-log" Oct 06 10:17:35 crc kubenswrapper[4824]: I1006 10:17:35.488718 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="508d58be-fcaa-4360-ab57-fbaedb728761" containerName="glance-log" Oct 06 10:17:35 crc kubenswrapper[4824]: E1006 10:17:35.488768 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7dd053a5-58ba-4f1c-9d32-e639fa8b1450" containerName="mariadb-account-create" Oct 06 10:17:35 crc kubenswrapper[4824]: I1006 10:17:35.488777 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="7dd053a5-58ba-4f1c-9d32-e639fa8b1450" containerName="mariadb-account-create" Oct 06 10:17:35 crc kubenswrapper[4824]: E1006 10:17:35.488794 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7cbd2d79-2d35-4954-9f61-c9bbbbe9d84e" containerName="mariadb-account-create" Oct 06 10:17:35 crc kubenswrapper[4824]: I1006 10:17:35.488801 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="7cbd2d79-2d35-4954-9f61-c9bbbbe9d84e" containerName="mariadb-account-create" Oct 06 10:17:35 crc kubenswrapper[4824]: E1006 10:17:35.488822 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="508d58be-fcaa-4360-ab57-fbaedb728761" containerName="glance-httpd" Oct 06 10:17:35 crc kubenswrapper[4824]: I1006 10:17:35.488831 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="508d58be-fcaa-4360-ab57-fbaedb728761" containerName="glance-httpd" Oct 06 10:17:35 crc kubenswrapper[4824]: E1006 10:17:35.488852 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4007653a-5c39-4285-8ba3-e0b3caed5640" containerName="mariadb-account-create" Oct 06 10:17:35 crc kubenswrapper[4824]: I1006 10:17:35.488861 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="4007653a-5c39-4285-8ba3-e0b3caed5640" containerName="mariadb-account-create" Oct 06 10:17:35 crc kubenswrapper[4824]: I1006 10:17:35.489091 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="7cbd2d79-2d35-4954-9f61-c9bbbbe9d84e" containerName="mariadb-account-create" Oct 06 10:17:35 crc kubenswrapper[4824]: I1006 10:17:35.489118 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="508d58be-fcaa-4360-ab57-fbaedb728761" containerName="glance-log" Oct 06 10:17:35 crc kubenswrapper[4824]: I1006 10:17:35.489135 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="508d58be-fcaa-4360-ab57-fbaedb728761" containerName="glance-httpd" Oct 06 10:17:35 crc kubenswrapper[4824]: I1006 10:17:35.489150 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="4007653a-5c39-4285-8ba3-e0b3caed5640" containerName="mariadb-account-create" Oct 06 10:17:35 crc kubenswrapper[4824]: I1006 10:17:35.489161 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="7dd053a5-58ba-4f1c-9d32-e639fa8b1450" containerName="mariadb-account-create" Oct 06 10:17:35 crc kubenswrapper[4824]: I1006 10:17:35.490331 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 06 10:17:35 crc kubenswrapper[4824]: I1006 10:17:35.495686 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 06 10:17:35 crc kubenswrapper[4824]: I1006 10:17:35.500822 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Oct 06 10:17:35 crc kubenswrapper[4824]: I1006 10:17:35.501140 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 06 10:17:35 crc kubenswrapper[4824]: I1006 10:17:35.635963 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4b875c3d-d6b9-47c6-ae58-4ba7b78fb338-logs\") pod \"glance-default-external-api-0\" (UID: \"4b875c3d-d6b9-47c6-ae58-4ba7b78fb338\") " pod="openstack/glance-default-external-api-0" Oct 06 10:17:35 crc kubenswrapper[4824]: I1006 10:17:35.636454 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-frbmb\" (UniqueName: \"kubernetes.io/projected/4b875c3d-d6b9-47c6-ae58-4ba7b78fb338-kube-api-access-frbmb\") pod \"glance-default-external-api-0\" (UID: \"4b875c3d-d6b9-47c6-ae58-4ba7b78fb338\") " pod="openstack/glance-default-external-api-0" Oct 06 10:17:35 crc kubenswrapper[4824]: I1006 10:17:35.636564 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4b875c3d-d6b9-47c6-ae58-4ba7b78fb338-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"4b875c3d-d6b9-47c6-ae58-4ba7b78fb338\") " pod="openstack/glance-default-external-api-0" Oct 06 10:17:35 crc kubenswrapper[4824]: I1006 10:17:35.636643 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4b875c3d-d6b9-47c6-ae58-4ba7b78fb338-scripts\") pod \"glance-default-external-api-0\" (UID: \"4b875c3d-d6b9-47c6-ae58-4ba7b78fb338\") " pod="openstack/glance-default-external-api-0" Oct 06 10:17:35 crc kubenswrapper[4824]: I1006 10:17:35.636676 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"4b875c3d-d6b9-47c6-ae58-4ba7b78fb338\") " pod="openstack/glance-default-external-api-0" Oct 06 10:17:35 crc kubenswrapper[4824]: I1006 10:17:35.636732 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b875c3d-d6b9-47c6-ae58-4ba7b78fb338-config-data\") pod \"glance-default-external-api-0\" (UID: \"4b875c3d-d6b9-47c6-ae58-4ba7b78fb338\") " pod="openstack/glance-default-external-api-0" Oct 06 10:17:35 crc kubenswrapper[4824]: I1006 10:17:35.636763 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4b875c3d-d6b9-47c6-ae58-4ba7b78fb338-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"4b875c3d-d6b9-47c6-ae58-4ba7b78fb338\") " pod="openstack/glance-default-external-api-0" Oct 06 10:17:35 crc kubenswrapper[4824]: I1006 10:17:35.636809 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b875c3d-d6b9-47c6-ae58-4ba7b78fb338-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"4b875c3d-d6b9-47c6-ae58-4ba7b78fb338\") " pod="openstack/glance-default-external-api-0" Oct 06 10:17:35 crc kubenswrapper[4824]: I1006 10:17:35.739417 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4b875c3d-d6b9-47c6-ae58-4ba7b78fb338-scripts\") pod \"glance-default-external-api-0\" (UID: \"4b875c3d-d6b9-47c6-ae58-4ba7b78fb338\") " pod="openstack/glance-default-external-api-0" Oct 06 10:17:35 crc kubenswrapper[4824]: I1006 10:17:35.739476 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"4b875c3d-d6b9-47c6-ae58-4ba7b78fb338\") " pod="openstack/glance-default-external-api-0" Oct 06 10:17:35 crc kubenswrapper[4824]: I1006 10:17:35.739514 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b875c3d-d6b9-47c6-ae58-4ba7b78fb338-config-data\") pod \"glance-default-external-api-0\" (UID: \"4b875c3d-d6b9-47c6-ae58-4ba7b78fb338\") " pod="openstack/glance-default-external-api-0" Oct 06 10:17:35 crc kubenswrapper[4824]: I1006 10:17:35.739545 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4b875c3d-d6b9-47c6-ae58-4ba7b78fb338-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"4b875c3d-d6b9-47c6-ae58-4ba7b78fb338\") " pod="openstack/glance-default-external-api-0" Oct 06 10:17:35 crc kubenswrapper[4824]: I1006 10:17:35.739566 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b875c3d-d6b9-47c6-ae58-4ba7b78fb338-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"4b875c3d-d6b9-47c6-ae58-4ba7b78fb338\") " pod="openstack/glance-default-external-api-0" Oct 06 10:17:35 crc kubenswrapper[4824]: I1006 10:17:35.739637 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4b875c3d-d6b9-47c6-ae58-4ba7b78fb338-logs\") pod \"glance-default-external-api-0\" (UID: \"4b875c3d-d6b9-47c6-ae58-4ba7b78fb338\") " pod="openstack/glance-default-external-api-0" Oct 06 10:17:35 crc kubenswrapper[4824]: I1006 10:17:35.739662 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-frbmb\" (UniqueName: \"kubernetes.io/projected/4b875c3d-d6b9-47c6-ae58-4ba7b78fb338-kube-api-access-frbmb\") pod \"glance-default-external-api-0\" (UID: \"4b875c3d-d6b9-47c6-ae58-4ba7b78fb338\") " pod="openstack/glance-default-external-api-0" Oct 06 10:17:35 crc kubenswrapper[4824]: I1006 10:17:35.739705 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4b875c3d-d6b9-47c6-ae58-4ba7b78fb338-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"4b875c3d-d6b9-47c6-ae58-4ba7b78fb338\") " pod="openstack/glance-default-external-api-0" Oct 06 10:17:35 crc kubenswrapper[4824]: I1006 10:17:35.740029 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4b875c3d-d6b9-47c6-ae58-4ba7b78fb338-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"4b875c3d-d6b9-47c6-ae58-4ba7b78fb338\") " pod="openstack/glance-default-external-api-0" Oct 06 10:17:35 crc kubenswrapper[4824]: I1006 10:17:35.740065 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4b875c3d-d6b9-47c6-ae58-4ba7b78fb338-logs\") pod \"glance-default-external-api-0\" (UID: \"4b875c3d-d6b9-47c6-ae58-4ba7b78fb338\") " pod="openstack/glance-default-external-api-0" Oct 06 10:17:35 crc kubenswrapper[4824]: I1006 10:17:35.740105 4824 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"4b875c3d-d6b9-47c6-ae58-4ba7b78fb338\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/glance-default-external-api-0" Oct 06 10:17:35 crc kubenswrapper[4824]: I1006 10:17:35.748036 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4b875c3d-d6b9-47c6-ae58-4ba7b78fb338-scripts\") pod \"glance-default-external-api-0\" (UID: \"4b875c3d-d6b9-47c6-ae58-4ba7b78fb338\") " pod="openstack/glance-default-external-api-0" Oct 06 10:17:35 crc kubenswrapper[4824]: I1006 10:17:35.748569 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4b875c3d-d6b9-47c6-ae58-4ba7b78fb338-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"4b875c3d-d6b9-47c6-ae58-4ba7b78fb338\") " pod="openstack/glance-default-external-api-0" Oct 06 10:17:35 crc kubenswrapper[4824]: I1006 10:17:35.760785 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b875c3d-d6b9-47c6-ae58-4ba7b78fb338-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"4b875c3d-d6b9-47c6-ae58-4ba7b78fb338\") " pod="openstack/glance-default-external-api-0" Oct 06 10:17:35 crc kubenswrapper[4824]: I1006 10:17:35.762765 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b875c3d-d6b9-47c6-ae58-4ba7b78fb338-config-data\") pod \"glance-default-external-api-0\" (UID: \"4b875c3d-d6b9-47c6-ae58-4ba7b78fb338\") " pod="openstack/glance-default-external-api-0" Oct 06 10:17:35 crc kubenswrapper[4824]: I1006 10:17:35.766138 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-frbmb\" (UniqueName: \"kubernetes.io/projected/4b875c3d-d6b9-47c6-ae58-4ba7b78fb338-kube-api-access-frbmb\") pod \"glance-default-external-api-0\" (UID: \"4b875c3d-d6b9-47c6-ae58-4ba7b78fb338\") " pod="openstack/glance-default-external-api-0" Oct 06 10:17:35 crc kubenswrapper[4824]: I1006 10:17:35.788272 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"4b875c3d-d6b9-47c6-ae58-4ba7b78fb338\") " pod="openstack/glance-default-external-api-0" Oct 06 10:17:35 crc kubenswrapper[4824]: I1006 10:17:35.882695 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 06 10:17:36 crc kubenswrapper[4824]: I1006 10:17:36.223617 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 06 10:17:36 crc kubenswrapper[4824]: I1006 10:17:36.260079 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-7txxj"] Oct 06 10:17:36 crc kubenswrapper[4824]: E1006 10:17:36.260727 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a85f7f99-c70b-40a6-8ea2-839d91a53407" containerName="glance-httpd" Oct 06 10:17:36 crc kubenswrapper[4824]: I1006 10:17:36.260750 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="a85f7f99-c70b-40a6-8ea2-839d91a53407" containerName="glance-httpd" Oct 06 10:17:36 crc kubenswrapper[4824]: E1006 10:17:36.260779 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a85f7f99-c70b-40a6-8ea2-839d91a53407" containerName="glance-log" Oct 06 10:17:36 crc kubenswrapper[4824]: I1006 10:17:36.260787 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="a85f7f99-c70b-40a6-8ea2-839d91a53407" containerName="glance-log" Oct 06 10:17:36 crc kubenswrapper[4824]: I1006 10:17:36.261095 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="a85f7f99-c70b-40a6-8ea2-839d91a53407" containerName="glance-log" Oct 06 10:17:36 crc kubenswrapper[4824]: I1006 10:17:36.261122 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="a85f7f99-c70b-40a6-8ea2-839d91a53407" containerName="glance-httpd" Oct 06 10:17:36 crc kubenswrapper[4824]: I1006 10:17:36.263234 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-7txxj" Oct 06 10:17:36 crc kubenswrapper[4824]: I1006 10:17:36.272378 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-wdpnq" Oct 06 10:17:36 crc kubenswrapper[4824]: I1006 10:17:36.272718 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Oct 06 10:17:36 crc kubenswrapper[4824]: I1006 10:17:36.273690 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 06 10:17:36 crc kubenswrapper[4824]: I1006 10:17:36.297204 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-7txxj"] Oct 06 10:17:36 crc kubenswrapper[4824]: I1006 10:17:36.354037 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a85f7f99-c70b-40a6-8ea2-839d91a53407-logs\") pod \"a85f7f99-c70b-40a6-8ea2-839d91a53407\" (UID: \"a85f7f99-c70b-40a6-8ea2-839d91a53407\") " Oct 06 10:17:36 crc kubenswrapper[4824]: I1006 10:17:36.354097 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"a85f7f99-c70b-40a6-8ea2-839d91a53407\" (UID: \"a85f7f99-c70b-40a6-8ea2-839d91a53407\") " Oct 06 10:17:36 crc kubenswrapper[4824]: I1006 10:17:36.354154 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a85f7f99-c70b-40a6-8ea2-839d91a53407-combined-ca-bundle\") pod \"a85f7f99-c70b-40a6-8ea2-839d91a53407\" (UID: \"a85f7f99-c70b-40a6-8ea2-839d91a53407\") " Oct 06 10:17:36 crc kubenswrapper[4824]: I1006 10:17:36.354775 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a85f7f99-c70b-40a6-8ea2-839d91a53407-logs" (OuterVolumeSpecName: "logs") pod "a85f7f99-c70b-40a6-8ea2-839d91a53407" (UID: "a85f7f99-c70b-40a6-8ea2-839d91a53407"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:17:36 crc kubenswrapper[4824]: I1006 10:17:36.355155 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a85f7f99-c70b-40a6-8ea2-839d91a53407-httpd-run\") pod \"a85f7f99-c70b-40a6-8ea2-839d91a53407\" (UID: \"a85f7f99-c70b-40a6-8ea2-839d91a53407\") " Oct 06 10:17:36 crc kubenswrapper[4824]: I1006 10:17:36.355251 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a85f7f99-c70b-40a6-8ea2-839d91a53407-scripts\") pod \"a85f7f99-c70b-40a6-8ea2-839d91a53407\" (UID: \"a85f7f99-c70b-40a6-8ea2-839d91a53407\") " Oct 06 10:17:36 crc kubenswrapper[4824]: I1006 10:17:36.355289 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n5zpc\" (UniqueName: \"kubernetes.io/projected/a85f7f99-c70b-40a6-8ea2-839d91a53407-kube-api-access-n5zpc\") pod \"a85f7f99-c70b-40a6-8ea2-839d91a53407\" (UID: \"a85f7f99-c70b-40a6-8ea2-839d91a53407\") " Oct 06 10:17:36 crc kubenswrapper[4824]: I1006 10:17:36.355431 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a85f7f99-c70b-40a6-8ea2-839d91a53407-config-data\") pod \"a85f7f99-c70b-40a6-8ea2-839d91a53407\" (UID: \"a85f7f99-c70b-40a6-8ea2-839d91a53407\") " Oct 06 10:17:36 crc kubenswrapper[4824]: I1006 10:17:36.355505 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a85f7f99-c70b-40a6-8ea2-839d91a53407-internal-tls-certs\") pod \"a85f7f99-c70b-40a6-8ea2-839d91a53407\" (UID: \"a85f7f99-c70b-40a6-8ea2-839d91a53407\") " Oct 06 10:17:36 crc kubenswrapper[4824]: I1006 10:17:36.356083 4824 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a85f7f99-c70b-40a6-8ea2-839d91a53407-logs\") on node \"crc\" DevicePath \"\"" Oct 06 10:17:36 crc kubenswrapper[4824]: I1006 10:17:36.357101 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a85f7f99-c70b-40a6-8ea2-839d91a53407-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "a85f7f99-c70b-40a6-8ea2-839d91a53407" (UID: "a85f7f99-c70b-40a6-8ea2-839d91a53407"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:17:36 crc kubenswrapper[4824]: I1006 10:17:36.359010 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "glance") pod "a85f7f99-c70b-40a6-8ea2-839d91a53407" (UID: "a85f7f99-c70b-40a6-8ea2-839d91a53407"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 06 10:17:36 crc kubenswrapper[4824]: I1006 10:17:36.365769 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a85f7f99-c70b-40a6-8ea2-839d91a53407-kube-api-access-n5zpc" (OuterVolumeSpecName: "kube-api-access-n5zpc") pod "a85f7f99-c70b-40a6-8ea2-839d91a53407" (UID: "a85f7f99-c70b-40a6-8ea2-839d91a53407"). InnerVolumeSpecName "kube-api-access-n5zpc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:17:36 crc kubenswrapper[4824]: I1006 10:17:36.374107 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a85f7f99-c70b-40a6-8ea2-839d91a53407-scripts" (OuterVolumeSpecName: "scripts") pod "a85f7f99-c70b-40a6-8ea2-839d91a53407" (UID: "a85f7f99-c70b-40a6-8ea2-839d91a53407"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:17:36 crc kubenswrapper[4824]: I1006 10:17:36.450117 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a85f7f99-c70b-40a6-8ea2-839d91a53407-config-data" (OuterVolumeSpecName: "config-data") pod "a85f7f99-c70b-40a6-8ea2-839d91a53407" (UID: "a85f7f99-c70b-40a6-8ea2-839d91a53407"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:17:36 crc kubenswrapper[4824]: I1006 10:17:36.456438 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a85f7f99-c70b-40a6-8ea2-839d91a53407-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a85f7f99-c70b-40a6-8ea2-839d91a53407" (UID: "a85f7f99-c70b-40a6-8ea2-839d91a53407"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:17:36 crc kubenswrapper[4824]: I1006 10:17:36.464955 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a08722d-580d-4dbc-80d5-f33caa5cae5b-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-7txxj\" (UID: \"8a08722d-580d-4dbc-80d5-f33caa5cae5b\") " pod="openstack/nova-cell0-conductor-db-sync-7txxj" Oct 06 10:17:36 crc kubenswrapper[4824]: I1006 10:17:36.465028 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-88t2l\" (UniqueName: \"kubernetes.io/projected/8a08722d-580d-4dbc-80d5-f33caa5cae5b-kube-api-access-88t2l\") pod \"nova-cell0-conductor-db-sync-7txxj\" (UID: \"8a08722d-580d-4dbc-80d5-f33caa5cae5b\") " pod="openstack/nova-cell0-conductor-db-sync-7txxj" Oct 06 10:17:36 crc kubenswrapper[4824]: I1006 10:17:36.465160 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a08722d-580d-4dbc-80d5-f33caa5cae5b-config-data\") pod \"nova-cell0-conductor-db-sync-7txxj\" (UID: \"8a08722d-580d-4dbc-80d5-f33caa5cae5b\") " pod="openstack/nova-cell0-conductor-db-sync-7txxj" Oct 06 10:17:36 crc kubenswrapper[4824]: I1006 10:17:36.465226 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8a08722d-580d-4dbc-80d5-f33caa5cae5b-scripts\") pod \"nova-cell0-conductor-db-sync-7txxj\" (UID: \"8a08722d-580d-4dbc-80d5-f33caa5cae5b\") " pod="openstack/nova-cell0-conductor-db-sync-7txxj" Oct 06 10:17:36 crc kubenswrapper[4824]: I1006 10:17:36.466366 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a85f7f99-c70b-40a6-8ea2-839d91a53407-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 10:17:36 crc kubenswrapper[4824]: I1006 10:17:36.466408 4824 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Oct 06 10:17:36 crc kubenswrapper[4824]: I1006 10:17:36.466909 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a85f7f99-c70b-40a6-8ea2-839d91a53407-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 10:17:36 crc kubenswrapper[4824]: I1006 10:17:36.466930 4824 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a85f7f99-c70b-40a6-8ea2-839d91a53407-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 06 10:17:36 crc kubenswrapper[4824]: I1006 10:17:36.466942 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a85f7f99-c70b-40a6-8ea2-839d91a53407-scripts\") on node \"crc\" DevicePath \"\"" Oct 06 10:17:36 crc kubenswrapper[4824]: I1006 10:17:36.466952 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n5zpc\" (UniqueName: \"kubernetes.io/projected/a85f7f99-c70b-40a6-8ea2-839d91a53407-kube-api-access-n5zpc\") on node \"crc\" DevicePath \"\"" Oct 06 10:17:36 crc kubenswrapper[4824]: I1006 10:17:36.499883 4824 generic.go:334] "Generic (PLEG): container finished" podID="a85f7f99-c70b-40a6-8ea2-839d91a53407" containerID="866e8b22170a7926f5516f9626a0711f498d48e51135ad184d51704db33d88a6" exitCode=0 Oct 06 10:17:36 crc kubenswrapper[4824]: I1006 10:17:36.499945 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"a85f7f99-c70b-40a6-8ea2-839d91a53407","Type":"ContainerDied","Data":"866e8b22170a7926f5516f9626a0711f498d48e51135ad184d51704db33d88a6"} Oct 06 10:17:36 crc kubenswrapper[4824]: I1006 10:17:36.499971 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"a85f7f99-c70b-40a6-8ea2-839d91a53407","Type":"ContainerDied","Data":"681b14ea91aa7b1bdb3c7c3c06994964c1bef09cf51181d7dca3cf76ae83bcd7"} Oct 06 10:17:36 crc kubenswrapper[4824]: I1006 10:17:36.502166 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 06 10:17:36 crc kubenswrapper[4824]: I1006 10:17:36.505495 4824 scope.go:117] "RemoveContainer" containerID="866e8b22170a7926f5516f9626a0711f498d48e51135ad184d51704db33d88a6" Oct 06 10:17:36 crc kubenswrapper[4824]: I1006 10:17:36.540101 4824 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Oct 06 10:17:36 crc kubenswrapper[4824]: I1006 10:17:36.546845 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a85f7f99-c70b-40a6-8ea2-839d91a53407-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "a85f7f99-c70b-40a6-8ea2-839d91a53407" (UID: "a85f7f99-c70b-40a6-8ea2-839d91a53407"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:17:36 crc kubenswrapper[4824]: I1006 10:17:36.568615 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8a08722d-580d-4dbc-80d5-f33caa5cae5b-scripts\") pod \"nova-cell0-conductor-db-sync-7txxj\" (UID: \"8a08722d-580d-4dbc-80d5-f33caa5cae5b\") " pod="openstack/nova-cell0-conductor-db-sync-7txxj" Oct 06 10:17:36 crc kubenswrapper[4824]: I1006 10:17:36.568708 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a08722d-580d-4dbc-80d5-f33caa5cae5b-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-7txxj\" (UID: \"8a08722d-580d-4dbc-80d5-f33caa5cae5b\") " pod="openstack/nova-cell0-conductor-db-sync-7txxj" Oct 06 10:17:36 crc kubenswrapper[4824]: I1006 10:17:36.568735 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-88t2l\" (UniqueName: \"kubernetes.io/projected/8a08722d-580d-4dbc-80d5-f33caa5cae5b-kube-api-access-88t2l\") pod \"nova-cell0-conductor-db-sync-7txxj\" (UID: \"8a08722d-580d-4dbc-80d5-f33caa5cae5b\") " pod="openstack/nova-cell0-conductor-db-sync-7txxj" Oct 06 10:17:36 crc kubenswrapper[4824]: I1006 10:17:36.568809 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a08722d-580d-4dbc-80d5-f33caa5cae5b-config-data\") pod \"nova-cell0-conductor-db-sync-7txxj\" (UID: \"8a08722d-580d-4dbc-80d5-f33caa5cae5b\") " pod="openstack/nova-cell0-conductor-db-sync-7txxj" Oct 06 10:17:36 crc kubenswrapper[4824]: I1006 10:17:36.568866 4824 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a85f7f99-c70b-40a6-8ea2-839d91a53407-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 06 10:17:36 crc kubenswrapper[4824]: I1006 10:17:36.568878 4824 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Oct 06 10:17:36 crc kubenswrapper[4824]: I1006 10:17:36.574751 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8a08722d-580d-4dbc-80d5-f33caa5cae5b-scripts\") pod \"nova-cell0-conductor-db-sync-7txxj\" (UID: \"8a08722d-580d-4dbc-80d5-f33caa5cae5b\") " pod="openstack/nova-cell0-conductor-db-sync-7txxj" Oct 06 10:17:36 crc kubenswrapper[4824]: I1006 10:17:36.587675 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a08722d-580d-4dbc-80d5-f33caa5cae5b-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-7txxj\" (UID: \"8a08722d-580d-4dbc-80d5-f33caa5cae5b\") " pod="openstack/nova-cell0-conductor-db-sync-7txxj" Oct 06 10:17:36 crc kubenswrapper[4824]: I1006 10:17:36.606671 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a08722d-580d-4dbc-80d5-f33caa5cae5b-config-data\") pod \"nova-cell0-conductor-db-sync-7txxj\" (UID: \"8a08722d-580d-4dbc-80d5-f33caa5cae5b\") " pod="openstack/nova-cell0-conductor-db-sync-7txxj" Oct 06 10:17:36 crc kubenswrapper[4824]: I1006 10:17:36.610838 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-88t2l\" (UniqueName: \"kubernetes.io/projected/8a08722d-580d-4dbc-80d5-f33caa5cae5b-kube-api-access-88t2l\") pod \"nova-cell0-conductor-db-sync-7txxj\" (UID: \"8a08722d-580d-4dbc-80d5-f33caa5cae5b\") " pod="openstack/nova-cell0-conductor-db-sync-7txxj" Oct 06 10:17:36 crc kubenswrapper[4824]: I1006 10:17:36.622200 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 06 10:17:36 crc kubenswrapper[4824]: W1006 10:17:36.634284 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4b875c3d_d6b9_47c6_ae58_4ba7b78fb338.slice/crio-e83f828c2f52fab6981059d74206fa043d0820d7885e84dde5ea70df08b7e433 WatchSource:0}: Error finding container e83f828c2f52fab6981059d74206fa043d0820d7885e84dde5ea70df08b7e433: Status 404 returned error can't find the container with id e83f828c2f52fab6981059d74206fa043d0820d7885e84dde5ea70df08b7e433 Oct 06 10:17:36 crc kubenswrapper[4824]: I1006 10:17:36.698261 4824 scope.go:117] "RemoveContainer" containerID="0a813108b412d32c6f3f573b57157b9dca2f4c7447b82953ab11535213306d0a" Oct 06 10:17:36 crc kubenswrapper[4824]: I1006 10:17:36.724199 4824 scope.go:117] "RemoveContainer" containerID="866e8b22170a7926f5516f9626a0711f498d48e51135ad184d51704db33d88a6" Oct 06 10:17:36 crc kubenswrapper[4824]: E1006 10:17:36.725133 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"866e8b22170a7926f5516f9626a0711f498d48e51135ad184d51704db33d88a6\": container with ID starting with 866e8b22170a7926f5516f9626a0711f498d48e51135ad184d51704db33d88a6 not found: ID does not exist" containerID="866e8b22170a7926f5516f9626a0711f498d48e51135ad184d51704db33d88a6" Oct 06 10:17:36 crc kubenswrapper[4824]: I1006 10:17:36.725169 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"866e8b22170a7926f5516f9626a0711f498d48e51135ad184d51704db33d88a6"} err="failed to get container status \"866e8b22170a7926f5516f9626a0711f498d48e51135ad184d51704db33d88a6\": rpc error: code = NotFound desc = could not find container \"866e8b22170a7926f5516f9626a0711f498d48e51135ad184d51704db33d88a6\": container with ID starting with 866e8b22170a7926f5516f9626a0711f498d48e51135ad184d51704db33d88a6 not found: ID does not exist" Oct 06 10:17:36 crc kubenswrapper[4824]: I1006 10:17:36.725192 4824 scope.go:117] "RemoveContainer" containerID="0a813108b412d32c6f3f573b57157b9dca2f4c7447b82953ab11535213306d0a" Oct 06 10:17:36 crc kubenswrapper[4824]: E1006 10:17:36.727312 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0a813108b412d32c6f3f573b57157b9dca2f4c7447b82953ab11535213306d0a\": container with ID starting with 0a813108b412d32c6f3f573b57157b9dca2f4c7447b82953ab11535213306d0a not found: ID does not exist" containerID="0a813108b412d32c6f3f573b57157b9dca2f4c7447b82953ab11535213306d0a" Oct 06 10:17:36 crc kubenswrapper[4824]: I1006 10:17:36.727403 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0a813108b412d32c6f3f573b57157b9dca2f4c7447b82953ab11535213306d0a"} err="failed to get container status \"0a813108b412d32c6f3f573b57157b9dca2f4c7447b82953ab11535213306d0a\": rpc error: code = NotFound desc = could not find container \"0a813108b412d32c6f3f573b57157b9dca2f4c7447b82953ab11535213306d0a\": container with ID starting with 0a813108b412d32c6f3f573b57157b9dca2f4c7447b82953ab11535213306d0a not found: ID does not exist" Oct 06 10:17:36 crc kubenswrapper[4824]: I1006 10:17:36.851631 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 06 10:17:36 crc kubenswrapper[4824]: I1006 10:17:36.860814 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 06 10:17:36 crc kubenswrapper[4824]: I1006 10:17:36.876037 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 06 10:17:36 crc kubenswrapper[4824]: I1006 10:17:36.878024 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 06 10:17:36 crc kubenswrapper[4824]: I1006 10:17:36.882865 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 06 10:17:36 crc kubenswrapper[4824]: I1006 10:17:36.883316 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-7txxj" Oct 06 10:17:36 crc kubenswrapper[4824]: I1006 10:17:36.887693 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Oct 06 10:17:36 crc kubenswrapper[4824]: I1006 10:17:36.893490 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 06 10:17:36 crc kubenswrapper[4824]: I1006 10:17:36.983398 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7d31f37b-142f-4f46-962a-e806e3d1269c-scripts\") pod \"glance-default-internal-api-0\" (UID: \"7d31f37b-142f-4f46-962a-e806e3d1269c\") " pod="openstack/glance-default-internal-api-0" Oct 06 10:17:36 crc kubenswrapper[4824]: I1006 10:17:36.983538 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d31f37b-142f-4f46-962a-e806e3d1269c-config-data\") pod \"glance-default-internal-api-0\" (UID: \"7d31f37b-142f-4f46-962a-e806e3d1269c\") " pod="openstack/glance-default-internal-api-0" Oct 06 10:17:36 crc kubenswrapper[4824]: I1006 10:17:36.983591 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7d31f37b-142f-4f46-962a-e806e3d1269c-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"7d31f37b-142f-4f46-962a-e806e3d1269c\") " pod="openstack/glance-default-internal-api-0" Oct 06 10:17:36 crc kubenswrapper[4824]: I1006 10:17:36.983622 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8lj7g\" (UniqueName: \"kubernetes.io/projected/7d31f37b-142f-4f46-962a-e806e3d1269c-kube-api-access-8lj7g\") pod \"glance-default-internal-api-0\" (UID: \"7d31f37b-142f-4f46-962a-e806e3d1269c\") " pod="openstack/glance-default-internal-api-0" Oct 06 10:17:36 crc kubenswrapper[4824]: I1006 10:17:36.983649 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7d31f37b-142f-4f46-962a-e806e3d1269c-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"7d31f37b-142f-4f46-962a-e806e3d1269c\") " pod="openstack/glance-default-internal-api-0" Oct 06 10:17:36 crc kubenswrapper[4824]: I1006 10:17:36.983671 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d31f37b-142f-4f46-962a-e806e3d1269c-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"7d31f37b-142f-4f46-962a-e806e3d1269c\") " pod="openstack/glance-default-internal-api-0" Oct 06 10:17:36 crc kubenswrapper[4824]: I1006 10:17:36.983707 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"7d31f37b-142f-4f46-962a-e806e3d1269c\") " pod="openstack/glance-default-internal-api-0" Oct 06 10:17:36 crc kubenswrapper[4824]: I1006 10:17:36.983750 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7d31f37b-142f-4f46-962a-e806e3d1269c-logs\") pod \"glance-default-internal-api-0\" (UID: \"7d31f37b-142f-4f46-962a-e806e3d1269c\") " pod="openstack/glance-default-internal-api-0" Oct 06 10:17:37 crc kubenswrapper[4824]: I1006 10:17:37.085385 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7d31f37b-142f-4f46-962a-e806e3d1269c-scripts\") pod \"glance-default-internal-api-0\" (UID: \"7d31f37b-142f-4f46-962a-e806e3d1269c\") " pod="openstack/glance-default-internal-api-0" Oct 06 10:17:37 crc kubenswrapper[4824]: I1006 10:17:37.085833 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d31f37b-142f-4f46-962a-e806e3d1269c-config-data\") pod \"glance-default-internal-api-0\" (UID: \"7d31f37b-142f-4f46-962a-e806e3d1269c\") " pod="openstack/glance-default-internal-api-0" Oct 06 10:17:37 crc kubenswrapper[4824]: I1006 10:17:37.085914 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7d31f37b-142f-4f46-962a-e806e3d1269c-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"7d31f37b-142f-4f46-962a-e806e3d1269c\") " pod="openstack/glance-default-internal-api-0" Oct 06 10:17:37 crc kubenswrapper[4824]: I1006 10:17:37.086038 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8lj7g\" (UniqueName: \"kubernetes.io/projected/7d31f37b-142f-4f46-962a-e806e3d1269c-kube-api-access-8lj7g\") pod \"glance-default-internal-api-0\" (UID: \"7d31f37b-142f-4f46-962a-e806e3d1269c\") " pod="openstack/glance-default-internal-api-0" Oct 06 10:17:37 crc kubenswrapper[4824]: I1006 10:17:37.086089 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7d31f37b-142f-4f46-962a-e806e3d1269c-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"7d31f37b-142f-4f46-962a-e806e3d1269c\") " pod="openstack/glance-default-internal-api-0" Oct 06 10:17:37 crc kubenswrapper[4824]: I1006 10:17:37.086118 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d31f37b-142f-4f46-962a-e806e3d1269c-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"7d31f37b-142f-4f46-962a-e806e3d1269c\") " pod="openstack/glance-default-internal-api-0" Oct 06 10:17:37 crc kubenswrapper[4824]: I1006 10:17:37.086160 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"7d31f37b-142f-4f46-962a-e806e3d1269c\") " pod="openstack/glance-default-internal-api-0" Oct 06 10:17:37 crc kubenswrapper[4824]: I1006 10:17:37.086219 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7d31f37b-142f-4f46-962a-e806e3d1269c-logs\") pod \"glance-default-internal-api-0\" (UID: \"7d31f37b-142f-4f46-962a-e806e3d1269c\") " pod="openstack/glance-default-internal-api-0" Oct 06 10:17:37 crc kubenswrapper[4824]: I1006 10:17:37.086725 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7d31f37b-142f-4f46-962a-e806e3d1269c-logs\") pod \"glance-default-internal-api-0\" (UID: \"7d31f37b-142f-4f46-962a-e806e3d1269c\") " pod="openstack/glance-default-internal-api-0" Oct 06 10:17:37 crc kubenswrapper[4824]: I1006 10:17:37.087867 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7d31f37b-142f-4f46-962a-e806e3d1269c-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"7d31f37b-142f-4f46-962a-e806e3d1269c\") " pod="openstack/glance-default-internal-api-0" Oct 06 10:17:37 crc kubenswrapper[4824]: I1006 10:17:37.088438 4824 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"7d31f37b-142f-4f46-962a-e806e3d1269c\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/glance-default-internal-api-0" Oct 06 10:17:37 crc kubenswrapper[4824]: I1006 10:17:37.095695 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7d31f37b-142f-4f46-962a-e806e3d1269c-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"7d31f37b-142f-4f46-962a-e806e3d1269c\") " pod="openstack/glance-default-internal-api-0" Oct 06 10:17:37 crc kubenswrapper[4824]: I1006 10:17:37.096165 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7d31f37b-142f-4f46-962a-e806e3d1269c-scripts\") pod \"glance-default-internal-api-0\" (UID: \"7d31f37b-142f-4f46-962a-e806e3d1269c\") " pod="openstack/glance-default-internal-api-0" Oct 06 10:17:37 crc kubenswrapper[4824]: I1006 10:17:37.104914 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d31f37b-142f-4f46-962a-e806e3d1269c-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"7d31f37b-142f-4f46-962a-e806e3d1269c\") " pod="openstack/glance-default-internal-api-0" Oct 06 10:17:37 crc kubenswrapper[4824]: I1006 10:17:37.105492 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d31f37b-142f-4f46-962a-e806e3d1269c-config-data\") pod \"glance-default-internal-api-0\" (UID: \"7d31f37b-142f-4f46-962a-e806e3d1269c\") " pod="openstack/glance-default-internal-api-0" Oct 06 10:17:37 crc kubenswrapper[4824]: I1006 10:17:37.111592 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8lj7g\" (UniqueName: \"kubernetes.io/projected/7d31f37b-142f-4f46-962a-e806e3d1269c-kube-api-access-8lj7g\") pod \"glance-default-internal-api-0\" (UID: \"7d31f37b-142f-4f46-962a-e806e3d1269c\") " pod="openstack/glance-default-internal-api-0" Oct 06 10:17:37 crc kubenswrapper[4824]: I1006 10:17:37.164430 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"7d31f37b-142f-4f46-962a-e806e3d1269c\") " pod="openstack/glance-default-internal-api-0" Oct 06 10:17:37 crc kubenswrapper[4824]: I1006 10:17:37.233630 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 06 10:17:37 crc kubenswrapper[4824]: I1006 10:17:37.293551 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="508d58be-fcaa-4360-ab57-fbaedb728761" path="/var/lib/kubelet/pods/508d58be-fcaa-4360-ab57-fbaedb728761/volumes" Oct 06 10:17:37 crc kubenswrapper[4824]: I1006 10:17:37.294526 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a85f7f99-c70b-40a6-8ea2-839d91a53407" path="/var/lib/kubelet/pods/a85f7f99-c70b-40a6-8ea2-839d91a53407/volumes" Oct 06 10:17:37 crc kubenswrapper[4824]: I1006 10:17:37.387405 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-7txxj"] Oct 06 10:17:37 crc kubenswrapper[4824]: I1006 10:17:37.581117 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"4b875c3d-d6b9-47c6-ae58-4ba7b78fb338","Type":"ContainerStarted","Data":"81ea1d1146fc194d6b86c8704769f755b78deac41d0fdbdcb7131656ddace535"} Oct 06 10:17:37 crc kubenswrapper[4824]: I1006 10:17:37.581679 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"4b875c3d-d6b9-47c6-ae58-4ba7b78fb338","Type":"ContainerStarted","Data":"e83f828c2f52fab6981059d74206fa043d0820d7885e84dde5ea70df08b7e433"} Oct 06 10:17:37 crc kubenswrapper[4824]: I1006 10:17:37.593732 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-7txxj" event={"ID":"8a08722d-580d-4dbc-80d5-f33caa5cae5b","Type":"ContainerStarted","Data":"5855e1d64bb16ed534d2c29a3d1a45535528de8f1c3a0ab71a0ac1200dad1176"} Oct 06 10:17:37 crc kubenswrapper[4824]: I1006 10:17:37.818033 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 06 10:17:38 crc kubenswrapper[4824]: I1006 10:17:38.610886 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"4b875c3d-d6b9-47c6-ae58-4ba7b78fb338","Type":"ContainerStarted","Data":"ea3ad38f9f36982350afcefddddc83fe7299464e5584d979260dcd939a76fd31"} Oct 06 10:17:38 crc kubenswrapper[4824]: I1006 10:17:38.617617 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7d31f37b-142f-4f46-962a-e806e3d1269c","Type":"ContainerStarted","Data":"2e176c8f297c52cbf0904f338584739b9ecfe050391b6c9fbf8cd1bebfe431cd"} Oct 06 10:17:38 crc kubenswrapper[4824]: I1006 10:17:38.617659 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7d31f37b-142f-4f46-962a-e806e3d1269c","Type":"ContainerStarted","Data":"f352444f55df2f9c82cd7f4368deb661d7e780a2dadfcefe389efa91f6cc1219"} Oct 06 10:17:38 crc kubenswrapper[4824]: I1006 10:17:38.645855 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.645833017 podStartE2EDuration="3.645833017s" podCreationTimestamp="2025-10-06 10:17:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:17:38.637752037 +0000 UTC m=+1168.002174898" watchObservedRunningTime="2025-10-06 10:17:38.645833017 +0000 UTC m=+1168.010255879" Oct 06 10:17:39 crc kubenswrapper[4824]: I1006 10:17:39.402795 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 06 10:17:39 crc kubenswrapper[4824]: I1006 10:17:39.584228 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/76e8c268-3c97-4c7f-9f91-a06880778160-sg-core-conf-yaml\") pod \"76e8c268-3c97-4c7f-9f91-a06880778160\" (UID: \"76e8c268-3c97-4c7f-9f91-a06880778160\") " Oct 06 10:17:39 crc kubenswrapper[4824]: I1006 10:17:39.584286 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9jhb9\" (UniqueName: \"kubernetes.io/projected/76e8c268-3c97-4c7f-9f91-a06880778160-kube-api-access-9jhb9\") pod \"76e8c268-3c97-4c7f-9f91-a06880778160\" (UID: \"76e8c268-3c97-4c7f-9f91-a06880778160\") " Oct 06 10:17:39 crc kubenswrapper[4824]: I1006 10:17:39.584317 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/76e8c268-3c97-4c7f-9f91-a06880778160-log-httpd\") pod \"76e8c268-3c97-4c7f-9f91-a06880778160\" (UID: \"76e8c268-3c97-4c7f-9f91-a06880778160\") " Oct 06 10:17:39 crc kubenswrapper[4824]: I1006 10:17:39.584451 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/76e8c268-3c97-4c7f-9f91-a06880778160-run-httpd\") pod \"76e8c268-3c97-4c7f-9f91-a06880778160\" (UID: \"76e8c268-3c97-4c7f-9f91-a06880778160\") " Oct 06 10:17:39 crc kubenswrapper[4824]: I1006 10:17:39.584606 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76e8c268-3c97-4c7f-9f91-a06880778160-config-data\") pod \"76e8c268-3c97-4c7f-9f91-a06880778160\" (UID: \"76e8c268-3c97-4c7f-9f91-a06880778160\") " Oct 06 10:17:39 crc kubenswrapper[4824]: I1006 10:17:39.584645 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/76e8c268-3c97-4c7f-9f91-a06880778160-scripts\") pod \"76e8c268-3c97-4c7f-9f91-a06880778160\" (UID: \"76e8c268-3c97-4c7f-9f91-a06880778160\") " Oct 06 10:17:39 crc kubenswrapper[4824]: I1006 10:17:39.584704 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76e8c268-3c97-4c7f-9f91-a06880778160-combined-ca-bundle\") pod \"76e8c268-3c97-4c7f-9f91-a06880778160\" (UID: \"76e8c268-3c97-4c7f-9f91-a06880778160\") " Oct 06 10:17:39 crc kubenswrapper[4824]: I1006 10:17:39.586694 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/76e8c268-3c97-4c7f-9f91-a06880778160-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "76e8c268-3c97-4c7f-9f91-a06880778160" (UID: "76e8c268-3c97-4c7f-9f91-a06880778160"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:17:39 crc kubenswrapper[4824]: I1006 10:17:39.586777 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/76e8c268-3c97-4c7f-9f91-a06880778160-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "76e8c268-3c97-4c7f-9f91-a06880778160" (UID: "76e8c268-3c97-4c7f-9f91-a06880778160"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:17:39 crc kubenswrapper[4824]: I1006 10:17:39.591501 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76e8c268-3c97-4c7f-9f91-a06880778160-scripts" (OuterVolumeSpecName: "scripts") pod "76e8c268-3c97-4c7f-9f91-a06880778160" (UID: "76e8c268-3c97-4c7f-9f91-a06880778160"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:17:39 crc kubenswrapper[4824]: I1006 10:17:39.591502 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/76e8c268-3c97-4c7f-9f91-a06880778160-kube-api-access-9jhb9" (OuterVolumeSpecName: "kube-api-access-9jhb9") pod "76e8c268-3c97-4c7f-9f91-a06880778160" (UID: "76e8c268-3c97-4c7f-9f91-a06880778160"). InnerVolumeSpecName "kube-api-access-9jhb9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:17:39 crc kubenswrapper[4824]: I1006 10:17:39.613420 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76e8c268-3c97-4c7f-9f91-a06880778160-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "76e8c268-3c97-4c7f-9f91-a06880778160" (UID: "76e8c268-3c97-4c7f-9f91-a06880778160"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:17:39 crc kubenswrapper[4824]: I1006 10:17:39.635155 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7d31f37b-142f-4f46-962a-e806e3d1269c","Type":"ContainerStarted","Data":"e9f474dbaf0e8425692cf4b3f1c4ac1eb18996b95f45d0a8b8b3f5a70c2ddfe6"} Oct 06 10:17:39 crc kubenswrapper[4824]: I1006 10:17:39.641757 4824 generic.go:334] "Generic (PLEG): container finished" podID="76e8c268-3c97-4c7f-9f91-a06880778160" containerID="144d9cca53691b88b9390df3fd5c1a2fccd54826c76cd61650cbaa3aca0bc2a2" exitCode=0 Oct 06 10:17:39 crc kubenswrapper[4824]: I1006 10:17:39.642415 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 06 10:17:39 crc kubenswrapper[4824]: I1006 10:17:39.642524 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"76e8c268-3c97-4c7f-9f91-a06880778160","Type":"ContainerDied","Data":"144d9cca53691b88b9390df3fd5c1a2fccd54826c76cd61650cbaa3aca0bc2a2"} Oct 06 10:17:39 crc kubenswrapper[4824]: I1006 10:17:39.642556 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"76e8c268-3c97-4c7f-9f91-a06880778160","Type":"ContainerDied","Data":"68b2610b541b3e3f0020d4469973472a80e816f403ec3ac561c0c9c851112a45"} Oct 06 10:17:39 crc kubenswrapper[4824]: I1006 10:17:39.642573 4824 scope.go:117] "RemoveContainer" containerID="2a5c65dc4113f85362f010990c9335a88b385f42366cef3a900ae692d8a15729" Oct 06 10:17:39 crc kubenswrapper[4824]: I1006 10:17:39.667486 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.667456859 podStartE2EDuration="3.667456859s" podCreationTimestamp="2025-10-06 10:17:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:17:39.657225636 +0000 UTC m=+1169.021648517" watchObservedRunningTime="2025-10-06 10:17:39.667456859 +0000 UTC m=+1169.031879720" Oct 06 10:17:39 crc kubenswrapper[4824]: I1006 10:17:39.669294 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76e8c268-3c97-4c7f-9f91-a06880778160-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "76e8c268-3c97-4c7f-9f91-a06880778160" (UID: "76e8c268-3c97-4c7f-9f91-a06880778160"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:17:39 crc kubenswrapper[4824]: I1006 10:17:39.686597 4824 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/76e8c268-3c97-4c7f-9f91-a06880778160-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 06 10:17:39 crc kubenswrapper[4824]: I1006 10:17:39.686627 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/76e8c268-3c97-4c7f-9f91-a06880778160-scripts\") on node \"crc\" DevicePath \"\"" Oct 06 10:17:39 crc kubenswrapper[4824]: I1006 10:17:39.686636 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76e8c268-3c97-4c7f-9f91-a06880778160-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 10:17:39 crc kubenswrapper[4824]: I1006 10:17:39.686645 4824 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/76e8c268-3c97-4c7f-9f91-a06880778160-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 06 10:17:39 crc kubenswrapper[4824]: I1006 10:17:39.686656 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9jhb9\" (UniqueName: \"kubernetes.io/projected/76e8c268-3c97-4c7f-9f91-a06880778160-kube-api-access-9jhb9\") on node \"crc\" DevicePath \"\"" Oct 06 10:17:39 crc kubenswrapper[4824]: I1006 10:17:39.686664 4824 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/76e8c268-3c97-4c7f-9f91-a06880778160-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 06 10:17:39 crc kubenswrapper[4824]: I1006 10:17:39.717258 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76e8c268-3c97-4c7f-9f91-a06880778160-config-data" (OuterVolumeSpecName: "config-data") pod "76e8c268-3c97-4c7f-9f91-a06880778160" (UID: "76e8c268-3c97-4c7f-9f91-a06880778160"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:17:39 crc kubenswrapper[4824]: I1006 10:17:39.788835 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76e8c268-3c97-4c7f-9f91-a06880778160-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 10:17:39 crc kubenswrapper[4824]: I1006 10:17:39.790508 4824 scope.go:117] "RemoveContainer" containerID="8a525b236722cb1e2fa01fb10a0cad64293948d71e21f9d1d07cc94ee0acd59a" Oct 06 10:17:39 crc kubenswrapper[4824]: I1006 10:17:39.823855 4824 scope.go:117] "RemoveContainer" containerID="1fd9058c49573e736f1913745838a146ff0a052b9f51b8079a33eb82aea18a30" Oct 06 10:17:39 crc kubenswrapper[4824]: I1006 10:17:39.846259 4824 scope.go:117] "RemoveContainer" containerID="144d9cca53691b88b9390df3fd5c1a2fccd54826c76cd61650cbaa3aca0bc2a2" Oct 06 10:17:39 crc kubenswrapper[4824]: I1006 10:17:39.863858 4824 scope.go:117] "RemoveContainer" containerID="2a5c65dc4113f85362f010990c9335a88b385f42366cef3a900ae692d8a15729" Oct 06 10:17:39 crc kubenswrapper[4824]: E1006 10:17:39.865055 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2a5c65dc4113f85362f010990c9335a88b385f42366cef3a900ae692d8a15729\": container with ID starting with 2a5c65dc4113f85362f010990c9335a88b385f42366cef3a900ae692d8a15729 not found: ID does not exist" containerID="2a5c65dc4113f85362f010990c9335a88b385f42366cef3a900ae692d8a15729" Oct 06 10:17:39 crc kubenswrapper[4824]: I1006 10:17:39.865090 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2a5c65dc4113f85362f010990c9335a88b385f42366cef3a900ae692d8a15729"} err="failed to get container status \"2a5c65dc4113f85362f010990c9335a88b385f42366cef3a900ae692d8a15729\": rpc error: code = NotFound desc = could not find container \"2a5c65dc4113f85362f010990c9335a88b385f42366cef3a900ae692d8a15729\": container with ID starting with 2a5c65dc4113f85362f010990c9335a88b385f42366cef3a900ae692d8a15729 not found: ID does not exist" Oct 06 10:17:39 crc kubenswrapper[4824]: I1006 10:17:39.865114 4824 scope.go:117] "RemoveContainer" containerID="8a525b236722cb1e2fa01fb10a0cad64293948d71e21f9d1d07cc94ee0acd59a" Oct 06 10:17:39 crc kubenswrapper[4824]: E1006 10:17:39.865387 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8a525b236722cb1e2fa01fb10a0cad64293948d71e21f9d1d07cc94ee0acd59a\": container with ID starting with 8a525b236722cb1e2fa01fb10a0cad64293948d71e21f9d1d07cc94ee0acd59a not found: ID does not exist" containerID="8a525b236722cb1e2fa01fb10a0cad64293948d71e21f9d1d07cc94ee0acd59a" Oct 06 10:17:39 crc kubenswrapper[4824]: I1006 10:17:39.865452 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8a525b236722cb1e2fa01fb10a0cad64293948d71e21f9d1d07cc94ee0acd59a"} err="failed to get container status \"8a525b236722cb1e2fa01fb10a0cad64293948d71e21f9d1d07cc94ee0acd59a\": rpc error: code = NotFound desc = could not find container \"8a525b236722cb1e2fa01fb10a0cad64293948d71e21f9d1d07cc94ee0acd59a\": container with ID starting with 8a525b236722cb1e2fa01fb10a0cad64293948d71e21f9d1d07cc94ee0acd59a not found: ID does not exist" Oct 06 10:17:39 crc kubenswrapper[4824]: I1006 10:17:39.865486 4824 scope.go:117] "RemoveContainer" containerID="1fd9058c49573e736f1913745838a146ff0a052b9f51b8079a33eb82aea18a30" Oct 06 10:17:39 crc kubenswrapper[4824]: E1006 10:17:39.865771 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1fd9058c49573e736f1913745838a146ff0a052b9f51b8079a33eb82aea18a30\": container with ID starting with 1fd9058c49573e736f1913745838a146ff0a052b9f51b8079a33eb82aea18a30 not found: ID does not exist" containerID="1fd9058c49573e736f1913745838a146ff0a052b9f51b8079a33eb82aea18a30" Oct 06 10:17:39 crc kubenswrapper[4824]: I1006 10:17:39.865799 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1fd9058c49573e736f1913745838a146ff0a052b9f51b8079a33eb82aea18a30"} err="failed to get container status \"1fd9058c49573e736f1913745838a146ff0a052b9f51b8079a33eb82aea18a30\": rpc error: code = NotFound desc = could not find container \"1fd9058c49573e736f1913745838a146ff0a052b9f51b8079a33eb82aea18a30\": container with ID starting with 1fd9058c49573e736f1913745838a146ff0a052b9f51b8079a33eb82aea18a30 not found: ID does not exist" Oct 06 10:17:39 crc kubenswrapper[4824]: I1006 10:17:39.865814 4824 scope.go:117] "RemoveContainer" containerID="144d9cca53691b88b9390df3fd5c1a2fccd54826c76cd61650cbaa3aca0bc2a2" Oct 06 10:17:39 crc kubenswrapper[4824]: E1006 10:17:39.866230 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"144d9cca53691b88b9390df3fd5c1a2fccd54826c76cd61650cbaa3aca0bc2a2\": container with ID starting with 144d9cca53691b88b9390df3fd5c1a2fccd54826c76cd61650cbaa3aca0bc2a2 not found: ID does not exist" containerID="144d9cca53691b88b9390df3fd5c1a2fccd54826c76cd61650cbaa3aca0bc2a2" Oct 06 10:17:39 crc kubenswrapper[4824]: I1006 10:17:39.866268 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"144d9cca53691b88b9390df3fd5c1a2fccd54826c76cd61650cbaa3aca0bc2a2"} err="failed to get container status \"144d9cca53691b88b9390df3fd5c1a2fccd54826c76cd61650cbaa3aca0bc2a2\": rpc error: code = NotFound desc = could not find container \"144d9cca53691b88b9390df3fd5c1a2fccd54826c76cd61650cbaa3aca0bc2a2\": container with ID starting with 144d9cca53691b88b9390df3fd5c1a2fccd54826c76cd61650cbaa3aca0bc2a2 not found: ID does not exist" Oct 06 10:17:39 crc kubenswrapper[4824]: I1006 10:17:39.977669 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 06 10:17:39 crc kubenswrapper[4824]: I1006 10:17:39.986839 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 06 10:17:40 crc kubenswrapper[4824]: I1006 10:17:40.003145 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 06 10:17:40 crc kubenswrapper[4824]: E1006 10:17:40.003573 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76e8c268-3c97-4c7f-9f91-a06880778160" containerName="ceilometer-notification-agent" Oct 06 10:17:40 crc kubenswrapper[4824]: I1006 10:17:40.003595 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="76e8c268-3c97-4c7f-9f91-a06880778160" containerName="ceilometer-notification-agent" Oct 06 10:17:40 crc kubenswrapper[4824]: E1006 10:17:40.003612 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76e8c268-3c97-4c7f-9f91-a06880778160" containerName="proxy-httpd" Oct 06 10:17:40 crc kubenswrapper[4824]: I1006 10:17:40.003619 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="76e8c268-3c97-4c7f-9f91-a06880778160" containerName="proxy-httpd" Oct 06 10:17:40 crc kubenswrapper[4824]: E1006 10:17:40.003627 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76e8c268-3c97-4c7f-9f91-a06880778160" containerName="ceilometer-central-agent" Oct 06 10:17:40 crc kubenswrapper[4824]: I1006 10:17:40.003634 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="76e8c268-3c97-4c7f-9f91-a06880778160" containerName="ceilometer-central-agent" Oct 06 10:17:40 crc kubenswrapper[4824]: E1006 10:17:40.003649 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76e8c268-3c97-4c7f-9f91-a06880778160" containerName="sg-core" Oct 06 10:17:40 crc kubenswrapper[4824]: I1006 10:17:40.003654 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="76e8c268-3c97-4c7f-9f91-a06880778160" containerName="sg-core" Oct 06 10:17:40 crc kubenswrapper[4824]: I1006 10:17:40.003834 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="76e8c268-3c97-4c7f-9f91-a06880778160" containerName="sg-core" Oct 06 10:17:40 crc kubenswrapper[4824]: I1006 10:17:40.003852 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="76e8c268-3c97-4c7f-9f91-a06880778160" containerName="ceilometer-notification-agent" Oct 06 10:17:40 crc kubenswrapper[4824]: I1006 10:17:40.003879 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="76e8c268-3c97-4c7f-9f91-a06880778160" containerName="ceilometer-central-agent" Oct 06 10:17:40 crc kubenswrapper[4824]: I1006 10:17:40.003893 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="76e8c268-3c97-4c7f-9f91-a06880778160" containerName="proxy-httpd" Oct 06 10:17:40 crc kubenswrapper[4824]: I1006 10:17:40.009013 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 06 10:17:40 crc kubenswrapper[4824]: I1006 10:17:40.011476 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 06 10:17:40 crc kubenswrapper[4824]: I1006 10:17:40.012779 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 06 10:17:40 crc kubenswrapper[4824]: I1006 10:17:40.021309 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 06 10:17:40 crc kubenswrapper[4824]: I1006 10:17:40.095394 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/78c95895-6099-43a1-a8d9-8aec0d0c1011-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"78c95895-6099-43a1-a8d9-8aec0d0c1011\") " pod="openstack/ceilometer-0" Oct 06 10:17:40 crc kubenswrapper[4824]: I1006 10:17:40.095470 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7pjcq\" (UniqueName: \"kubernetes.io/projected/78c95895-6099-43a1-a8d9-8aec0d0c1011-kube-api-access-7pjcq\") pod \"ceilometer-0\" (UID: \"78c95895-6099-43a1-a8d9-8aec0d0c1011\") " pod="openstack/ceilometer-0" Oct 06 10:17:40 crc kubenswrapper[4824]: I1006 10:17:40.095499 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/78c95895-6099-43a1-a8d9-8aec0d0c1011-scripts\") pod \"ceilometer-0\" (UID: \"78c95895-6099-43a1-a8d9-8aec0d0c1011\") " pod="openstack/ceilometer-0" Oct 06 10:17:40 crc kubenswrapper[4824]: I1006 10:17:40.095798 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/78c95895-6099-43a1-a8d9-8aec0d0c1011-log-httpd\") pod \"ceilometer-0\" (UID: \"78c95895-6099-43a1-a8d9-8aec0d0c1011\") " pod="openstack/ceilometer-0" Oct 06 10:17:40 crc kubenswrapper[4824]: I1006 10:17:40.096019 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/78c95895-6099-43a1-a8d9-8aec0d0c1011-run-httpd\") pod \"ceilometer-0\" (UID: \"78c95895-6099-43a1-a8d9-8aec0d0c1011\") " pod="openstack/ceilometer-0" Oct 06 10:17:40 crc kubenswrapper[4824]: I1006 10:17:40.096279 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78c95895-6099-43a1-a8d9-8aec0d0c1011-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"78c95895-6099-43a1-a8d9-8aec0d0c1011\") " pod="openstack/ceilometer-0" Oct 06 10:17:40 crc kubenswrapper[4824]: I1006 10:17:40.096358 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/78c95895-6099-43a1-a8d9-8aec0d0c1011-config-data\") pod \"ceilometer-0\" (UID: \"78c95895-6099-43a1-a8d9-8aec0d0c1011\") " pod="openstack/ceilometer-0" Oct 06 10:17:40 crc kubenswrapper[4824]: I1006 10:17:40.198054 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78c95895-6099-43a1-a8d9-8aec0d0c1011-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"78c95895-6099-43a1-a8d9-8aec0d0c1011\") " pod="openstack/ceilometer-0" Oct 06 10:17:40 crc kubenswrapper[4824]: I1006 10:17:40.198106 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/78c95895-6099-43a1-a8d9-8aec0d0c1011-config-data\") pod \"ceilometer-0\" (UID: \"78c95895-6099-43a1-a8d9-8aec0d0c1011\") " pod="openstack/ceilometer-0" Oct 06 10:17:40 crc kubenswrapper[4824]: I1006 10:17:40.198128 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/78c95895-6099-43a1-a8d9-8aec0d0c1011-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"78c95895-6099-43a1-a8d9-8aec0d0c1011\") " pod="openstack/ceilometer-0" Oct 06 10:17:40 crc kubenswrapper[4824]: I1006 10:17:40.198190 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7pjcq\" (UniqueName: \"kubernetes.io/projected/78c95895-6099-43a1-a8d9-8aec0d0c1011-kube-api-access-7pjcq\") pod \"ceilometer-0\" (UID: \"78c95895-6099-43a1-a8d9-8aec0d0c1011\") " pod="openstack/ceilometer-0" Oct 06 10:17:40 crc kubenswrapper[4824]: I1006 10:17:40.198207 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/78c95895-6099-43a1-a8d9-8aec0d0c1011-scripts\") pod \"ceilometer-0\" (UID: \"78c95895-6099-43a1-a8d9-8aec0d0c1011\") " pod="openstack/ceilometer-0" Oct 06 10:17:40 crc kubenswrapper[4824]: I1006 10:17:40.198258 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/78c95895-6099-43a1-a8d9-8aec0d0c1011-log-httpd\") pod \"ceilometer-0\" (UID: \"78c95895-6099-43a1-a8d9-8aec0d0c1011\") " pod="openstack/ceilometer-0" Oct 06 10:17:40 crc kubenswrapper[4824]: I1006 10:17:40.198311 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/78c95895-6099-43a1-a8d9-8aec0d0c1011-run-httpd\") pod \"ceilometer-0\" (UID: \"78c95895-6099-43a1-a8d9-8aec0d0c1011\") " pod="openstack/ceilometer-0" Oct 06 10:17:40 crc kubenswrapper[4824]: I1006 10:17:40.198850 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/78c95895-6099-43a1-a8d9-8aec0d0c1011-run-httpd\") pod \"ceilometer-0\" (UID: \"78c95895-6099-43a1-a8d9-8aec0d0c1011\") " pod="openstack/ceilometer-0" Oct 06 10:17:40 crc kubenswrapper[4824]: I1006 10:17:40.199179 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/78c95895-6099-43a1-a8d9-8aec0d0c1011-log-httpd\") pod \"ceilometer-0\" (UID: \"78c95895-6099-43a1-a8d9-8aec0d0c1011\") " pod="openstack/ceilometer-0" Oct 06 10:17:40 crc kubenswrapper[4824]: I1006 10:17:40.206680 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/78c95895-6099-43a1-a8d9-8aec0d0c1011-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"78c95895-6099-43a1-a8d9-8aec0d0c1011\") " pod="openstack/ceilometer-0" Oct 06 10:17:40 crc kubenswrapper[4824]: I1006 10:17:40.206725 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/78c95895-6099-43a1-a8d9-8aec0d0c1011-config-data\") pod \"ceilometer-0\" (UID: \"78c95895-6099-43a1-a8d9-8aec0d0c1011\") " pod="openstack/ceilometer-0" Oct 06 10:17:40 crc kubenswrapper[4824]: I1006 10:17:40.207317 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/78c95895-6099-43a1-a8d9-8aec0d0c1011-scripts\") pod \"ceilometer-0\" (UID: \"78c95895-6099-43a1-a8d9-8aec0d0c1011\") " pod="openstack/ceilometer-0" Oct 06 10:17:40 crc kubenswrapper[4824]: I1006 10:17:40.208141 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78c95895-6099-43a1-a8d9-8aec0d0c1011-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"78c95895-6099-43a1-a8d9-8aec0d0c1011\") " pod="openstack/ceilometer-0" Oct 06 10:17:40 crc kubenswrapper[4824]: I1006 10:17:40.218335 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7pjcq\" (UniqueName: \"kubernetes.io/projected/78c95895-6099-43a1-a8d9-8aec0d0c1011-kube-api-access-7pjcq\") pod \"ceilometer-0\" (UID: \"78c95895-6099-43a1-a8d9-8aec0d0c1011\") " pod="openstack/ceilometer-0" Oct 06 10:17:40 crc kubenswrapper[4824]: I1006 10:17:40.326794 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 06 10:17:40 crc kubenswrapper[4824]: I1006 10:17:40.327718 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 06 10:17:40 crc kubenswrapper[4824]: I1006 10:17:40.784798 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 06 10:17:41 crc kubenswrapper[4824]: I1006 10:17:41.285419 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="76e8c268-3c97-4c7f-9f91-a06880778160" path="/var/lib/kubelet/pods/76e8c268-3c97-4c7f-9f91-a06880778160/volumes" Oct 06 10:17:41 crc kubenswrapper[4824]: I1006 10:17:41.674307 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"78c95895-6099-43a1-a8d9-8aec0d0c1011","Type":"ContainerStarted","Data":"d818910a7991c1d8606e9b0d998a5c8de3fdf0bfc5cf878d41f2c73a24354569"} Oct 06 10:17:45 crc kubenswrapper[4824]: I1006 10:17:45.727784 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-7txxj" event={"ID":"8a08722d-580d-4dbc-80d5-f33caa5cae5b","Type":"ContainerStarted","Data":"e93f66db5683a4761c779328c5d70b916fcb98226c72e9ab3dc9a4e98af10ebd"} Oct 06 10:17:45 crc kubenswrapper[4824]: I1006 10:17:45.734083 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"78c95895-6099-43a1-a8d9-8aec0d0c1011","Type":"ContainerStarted","Data":"e04d4316e7bc61728eecb6ffdc75465ba879c8ca5d3e795f954c9d8e0adee5b0"} Oct 06 10:17:45 crc kubenswrapper[4824]: I1006 10:17:45.752943 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-7txxj" podStartSLOduration=2.27035811 podStartE2EDuration="9.752924363s" podCreationTimestamp="2025-10-06 10:17:36 +0000 UTC" firstStartedPulling="2025-10-06 10:17:37.400143363 +0000 UTC m=+1166.764566224" lastFinishedPulling="2025-10-06 10:17:44.882709606 +0000 UTC m=+1174.247132477" observedRunningTime="2025-10-06 10:17:45.743410817 +0000 UTC m=+1175.107833698" watchObservedRunningTime="2025-10-06 10:17:45.752924363 +0000 UTC m=+1175.117347244" Oct 06 10:17:45 crc kubenswrapper[4824]: I1006 10:17:45.883442 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 06 10:17:45 crc kubenswrapper[4824]: I1006 10:17:45.883504 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 06 10:17:45 crc kubenswrapper[4824]: I1006 10:17:45.920167 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 06 10:17:45 crc kubenswrapper[4824]: I1006 10:17:45.935215 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 06 10:17:46 crc kubenswrapper[4824]: I1006 10:17:46.748307 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 06 10:17:46 crc kubenswrapper[4824]: I1006 10:17:46.750428 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 06 10:17:47 crc kubenswrapper[4824]: I1006 10:17:47.236243 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 06 10:17:47 crc kubenswrapper[4824]: I1006 10:17:47.237210 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 06 10:17:47 crc kubenswrapper[4824]: I1006 10:17:47.271115 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 06 10:17:47 crc kubenswrapper[4824]: I1006 10:17:47.306658 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 06 10:17:47 crc kubenswrapper[4824]: I1006 10:17:47.763782 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"78c95895-6099-43a1-a8d9-8aec0d0c1011","Type":"ContainerStarted","Data":"a141c4f46628bcf2704823b1f797622cfff0896a3b75c2d6aad4ddce0d593343"} Oct 06 10:17:47 crc kubenswrapper[4824]: I1006 10:17:47.764412 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 06 10:17:47 crc kubenswrapper[4824]: I1006 10:17:47.764430 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"78c95895-6099-43a1-a8d9-8aec0d0c1011","Type":"ContainerStarted","Data":"514305e83facb91f5df3d37a4a60700d59550463507ec40e702b13ede6712483"} Oct 06 10:17:47 crc kubenswrapper[4824]: I1006 10:17:47.764445 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 06 10:17:48 crc kubenswrapper[4824]: I1006 10:17:48.771676 4824 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 06 10:17:48 crc kubenswrapper[4824]: I1006 10:17:48.772054 4824 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 06 10:17:49 crc kubenswrapper[4824]: I1006 10:17:49.264179 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 06 10:17:49 crc kubenswrapper[4824]: I1006 10:17:49.511265 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 06 10:17:49 crc kubenswrapper[4824]: I1006 10:17:49.783161 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"78c95895-6099-43a1-a8d9-8aec0d0c1011","Type":"ContainerStarted","Data":"2e3785442dbf3c1814ee90d1eac917b76c0b8a8d191cdb044959e05403a2e1c2"} Oct 06 10:17:49 crc kubenswrapper[4824]: I1006 10:17:49.783431 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="78c95895-6099-43a1-a8d9-8aec0d0c1011" containerName="ceilometer-central-agent" containerID="cri-o://e04d4316e7bc61728eecb6ffdc75465ba879c8ca5d3e795f954c9d8e0adee5b0" gracePeriod=30 Oct 06 10:17:49 crc kubenswrapper[4824]: I1006 10:17:49.783463 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="78c95895-6099-43a1-a8d9-8aec0d0c1011" containerName="proxy-httpd" containerID="cri-o://2e3785442dbf3c1814ee90d1eac917b76c0b8a8d191cdb044959e05403a2e1c2" gracePeriod=30 Oct 06 10:17:49 crc kubenswrapper[4824]: I1006 10:17:49.783564 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="78c95895-6099-43a1-a8d9-8aec0d0c1011" containerName="ceilometer-notification-agent" containerID="cri-o://514305e83facb91f5df3d37a4a60700d59550463507ec40e702b13ede6712483" gracePeriod=30 Oct 06 10:17:49 crc kubenswrapper[4824]: I1006 10:17:49.783631 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 06 10:17:49 crc kubenswrapper[4824]: I1006 10:17:49.783610 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="78c95895-6099-43a1-a8d9-8aec0d0c1011" containerName="sg-core" containerID="cri-o://a141c4f46628bcf2704823b1f797622cfff0896a3b75c2d6aad4ddce0d593343" gracePeriod=30 Oct 06 10:17:49 crc kubenswrapper[4824]: I1006 10:17:49.816532 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.689896967 podStartE2EDuration="10.816504792s" podCreationTimestamp="2025-10-06 10:17:39 +0000 UTC" firstStartedPulling="2025-10-06 10:17:40.80914939 +0000 UTC m=+1170.173572251" lastFinishedPulling="2025-10-06 10:17:48.935757215 +0000 UTC m=+1178.300180076" observedRunningTime="2025-10-06 10:17:49.811537829 +0000 UTC m=+1179.175960690" watchObservedRunningTime="2025-10-06 10:17:49.816504792 +0000 UTC m=+1179.180927653" Oct 06 10:17:50 crc kubenswrapper[4824]: I1006 10:17:50.257279 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 06 10:17:50 crc kubenswrapper[4824]: I1006 10:17:50.257714 4824 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 06 10:17:50 crc kubenswrapper[4824]: I1006 10:17:50.298382 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 06 10:17:50 crc kubenswrapper[4824]: I1006 10:17:50.806766 4824 generic.go:334] "Generic (PLEG): container finished" podID="78c95895-6099-43a1-a8d9-8aec0d0c1011" containerID="2e3785442dbf3c1814ee90d1eac917b76c0b8a8d191cdb044959e05403a2e1c2" exitCode=0 Oct 06 10:17:50 crc kubenswrapper[4824]: I1006 10:17:50.806807 4824 generic.go:334] "Generic (PLEG): container finished" podID="78c95895-6099-43a1-a8d9-8aec0d0c1011" containerID="a141c4f46628bcf2704823b1f797622cfff0896a3b75c2d6aad4ddce0d593343" exitCode=2 Oct 06 10:17:50 crc kubenswrapper[4824]: I1006 10:17:50.806815 4824 generic.go:334] "Generic (PLEG): container finished" podID="78c95895-6099-43a1-a8d9-8aec0d0c1011" containerID="514305e83facb91f5df3d37a4a60700d59550463507ec40e702b13ede6712483" exitCode=0 Oct 06 10:17:50 crc kubenswrapper[4824]: I1006 10:17:50.806842 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"78c95895-6099-43a1-a8d9-8aec0d0c1011","Type":"ContainerDied","Data":"2e3785442dbf3c1814ee90d1eac917b76c0b8a8d191cdb044959e05403a2e1c2"} Oct 06 10:17:50 crc kubenswrapper[4824]: I1006 10:17:50.806889 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"78c95895-6099-43a1-a8d9-8aec0d0c1011","Type":"ContainerDied","Data":"a141c4f46628bcf2704823b1f797622cfff0896a3b75c2d6aad4ddce0d593343"} Oct 06 10:17:50 crc kubenswrapper[4824]: I1006 10:17:50.806899 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"78c95895-6099-43a1-a8d9-8aec0d0c1011","Type":"ContainerDied","Data":"514305e83facb91f5df3d37a4a60700d59550463507ec40e702b13ede6712483"} Oct 06 10:17:51 crc kubenswrapper[4824]: I1006 10:17:51.800221 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 06 10:17:51 crc kubenswrapper[4824]: I1006 10:17:51.823647 4824 generic.go:334] "Generic (PLEG): container finished" podID="78c95895-6099-43a1-a8d9-8aec0d0c1011" containerID="e04d4316e7bc61728eecb6ffdc75465ba879c8ca5d3e795f954c9d8e0adee5b0" exitCode=0 Oct 06 10:17:51 crc kubenswrapper[4824]: I1006 10:17:51.823787 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 06 10:17:51 crc kubenswrapper[4824]: I1006 10:17:51.823805 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"78c95895-6099-43a1-a8d9-8aec0d0c1011","Type":"ContainerDied","Data":"e04d4316e7bc61728eecb6ffdc75465ba879c8ca5d3e795f954c9d8e0adee5b0"} Oct 06 10:17:51 crc kubenswrapper[4824]: I1006 10:17:51.824188 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"78c95895-6099-43a1-a8d9-8aec0d0c1011","Type":"ContainerDied","Data":"d818910a7991c1d8606e9b0d998a5c8de3fdf0bfc5cf878d41f2c73a24354569"} Oct 06 10:17:51 crc kubenswrapper[4824]: I1006 10:17:51.824205 4824 scope.go:117] "RemoveContainer" containerID="2e3785442dbf3c1814ee90d1eac917b76c0b8a8d191cdb044959e05403a2e1c2" Oct 06 10:17:51 crc kubenswrapper[4824]: I1006 10:17:51.852069 4824 scope.go:117] "RemoveContainer" containerID="a141c4f46628bcf2704823b1f797622cfff0896a3b75c2d6aad4ddce0d593343" Oct 06 10:17:51 crc kubenswrapper[4824]: I1006 10:17:51.884862 4824 scope.go:117] "RemoveContainer" containerID="514305e83facb91f5df3d37a4a60700d59550463507ec40e702b13ede6712483" Oct 06 10:17:51 crc kubenswrapper[4824]: I1006 10:17:51.921581 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/78c95895-6099-43a1-a8d9-8aec0d0c1011-config-data\") pod \"78c95895-6099-43a1-a8d9-8aec0d0c1011\" (UID: \"78c95895-6099-43a1-a8d9-8aec0d0c1011\") " Oct 06 10:17:51 crc kubenswrapper[4824]: I1006 10:17:51.921666 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/78c95895-6099-43a1-a8d9-8aec0d0c1011-sg-core-conf-yaml\") pod \"78c95895-6099-43a1-a8d9-8aec0d0c1011\" (UID: \"78c95895-6099-43a1-a8d9-8aec0d0c1011\") " Oct 06 10:17:51 crc kubenswrapper[4824]: I1006 10:17:51.921741 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/78c95895-6099-43a1-a8d9-8aec0d0c1011-log-httpd\") pod \"78c95895-6099-43a1-a8d9-8aec0d0c1011\" (UID: \"78c95895-6099-43a1-a8d9-8aec0d0c1011\") " Oct 06 10:17:51 crc kubenswrapper[4824]: I1006 10:17:51.921765 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/78c95895-6099-43a1-a8d9-8aec0d0c1011-run-httpd\") pod \"78c95895-6099-43a1-a8d9-8aec0d0c1011\" (UID: \"78c95895-6099-43a1-a8d9-8aec0d0c1011\") " Oct 06 10:17:51 crc kubenswrapper[4824]: I1006 10:17:51.921807 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78c95895-6099-43a1-a8d9-8aec0d0c1011-combined-ca-bundle\") pod \"78c95895-6099-43a1-a8d9-8aec0d0c1011\" (UID: \"78c95895-6099-43a1-a8d9-8aec0d0c1011\") " Oct 06 10:17:51 crc kubenswrapper[4824]: I1006 10:17:51.921959 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/78c95895-6099-43a1-a8d9-8aec0d0c1011-scripts\") pod \"78c95895-6099-43a1-a8d9-8aec0d0c1011\" (UID: \"78c95895-6099-43a1-a8d9-8aec0d0c1011\") " Oct 06 10:17:51 crc kubenswrapper[4824]: I1006 10:17:51.922041 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7pjcq\" (UniqueName: \"kubernetes.io/projected/78c95895-6099-43a1-a8d9-8aec0d0c1011-kube-api-access-7pjcq\") pod \"78c95895-6099-43a1-a8d9-8aec0d0c1011\" (UID: \"78c95895-6099-43a1-a8d9-8aec0d0c1011\") " Oct 06 10:17:51 crc kubenswrapper[4824]: I1006 10:17:51.922436 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/78c95895-6099-43a1-a8d9-8aec0d0c1011-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "78c95895-6099-43a1-a8d9-8aec0d0c1011" (UID: "78c95895-6099-43a1-a8d9-8aec0d0c1011"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:17:51 crc kubenswrapper[4824]: I1006 10:17:51.922556 4824 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/78c95895-6099-43a1-a8d9-8aec0d0c1011-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 06 10:17:51 crc kubenswrapper[4824]: I1006 10:17:51.922720 4824 scope.go:117] "RemoveContainer" containerID="e04d4316e7bc61728eecb6ffdc75465ba879c8ca5d3e795f954c9d8e0adee5b0" Oct 06 10:17:51 crc kubenswrapper[4824]: I1006 10:17:51.922967 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/78c95895-6099-43a1-a8d9-8aec0d0c1011-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "78c95895-6099-43a1-a8d9-8aec0d0c1011" (UID: "78c95895-6099-43a1-a8d9-8aec0d0c1011"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:17:51 crc kubenswrapper[4824]: I1006 10:17:51.929849 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/78c95895-6099-43a1-a8d9-8aec0d0c1011-scripts" (OuterVolumeSpecName: "scripts") pod "78c95895-6099-43a1-a8d9-8aec0d0c1011" (UID: "78c95895-6099-43a1-a8d9-8aec0d0c1011"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:17:51 crc kubenswrapper[4824]: I1006 10:17:51.930741 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/78c95895-6099-43a1-a8d9-8aec0d0c1011-kube-api-access-7pjcq" (OuterVolumeSpecName: "kube-api-access-7pjcq") pod "78c95895-6099-43a1-a8d9-8aec0d0c1011" (UID: "78c95895-6099-43a1-a8d9-8aec0d0c1011"). InnerVolumeSpecName "kube-api-access-7pjcq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:17:51 crc kubenswrapper[4824]: I1006 10:17:51.955551 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/78c95895-6099-43a1-a8d9-8aec0d0c1011-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "78c95895-6099-43a1-a8d9-8aec0d0c1011" (UID: "78c95895-6099-43a1-a8d9-8aec0d0c1011"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:17:52 crc kubenswrapper[4824]: I1006 10:17:52.026011 4824 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/78c95895-6099-43a1-a8d9-8aec0d0c1011-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 06 10:17:52 crc kubenswrapper[4824]: I1006 10:17:52.026052 4824 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/78c95895-6099-43a1-a8d9-8aec0d0c1011-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 06 10:17:52 crc kubenswrapper[4824]: I1006 10:17:52.026068 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/78c95895-6099-43a1-a8d9-8aec0d0c1011-scripts\") on node \"crc\" DevicePath \"\"" Oct 06 10:17:52 crc kubenswrapper[4824]: I1006 10:17:52.026084 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7pjcq\" (UniqueName: \"kubernetes.io/projected/78c95895-6099-43a1-a8d9-8aec0d0c1011-kube-api-access-7pjcq\") on node \"crc\" DevicePath \"\"" Oct 06 10:17:52 crc kubenswrapper[4824]: I1006 10:17:52.044104 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/78c95895-6099-43a1-a8d9-8aec0d0c1011-config-data" (OuterVolumeSpecName: "config-data") pod "78c95895-6099-43a1-a8d9-8aec0d0c1011" (UID: "78c95895-6099-43a1-a8d9-8aec0d0c1011"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:17:52 crc kubenswrapper[4824]: I1006 10:17:52.044184 4824 scope.go:117] "RemoveContainer" containerID="2e3785442dbf3c1814ee90d1eac917b76c0b8a8d191cdb044959e05403a2e1c2" Oct 06 10:17:52 crc kubenswrapper[4824]: E1006 10:17:52.044658 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2e3785442dbf3c1814ee90d1eac917b76c0b8a8d191cdb044959e05403a2e1c2\": container with ID starting with 2e3785442dbf3c1814ee90d1eac917b76c0b8a8d191cdb044959e05403a2e1c2 not found: ID does not exist" containerID="2e3785442dbf3c1814ee90d1eac917b76c0b8a8d191cdb044959e05403a2e1c2" Oct 06 10:17:52 crc kubenswrapper[4824]: I1006 10:17:52.044722 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2e3785442dbf3c1814ee90d1eac917b76c0b8a8d191cdb044959e05403a2e1c2"} err="failed to get container status \"2e3785442dbf3c1814ee90d1eac917b76c0b8a8d191cdb044959e05403a2e1c2\": rpc error: code = NotFound desc = could not find container \"2e3785442dbf3c1814ee90d1eac917b76c0b8a8d191cdb044959e05403a2e1c2\": container with ID starting with 2e3785442dbf3c1814ee90d1eac917b76c0b8a8d191cdb044959e05403a2e1c2 not found: ID does not exist" Oct 06 10:17:52 crc kubenswrapper[4824]: I1006 10:17:52.044766 4824 scope.go:117] "RemoveContainer" containerID="a141c4f46628bcf2704823b1f797622cfff0896a3b75c2d6aad4ddce0d593343" Oct 06 10:17:52 crc kubenswrapper[4824]: E1006 10:17:52.046171 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a141c4f46628bcf2704823b1f797622cfff0896a3b75c2d6aad4ddce0d593343\": container with ID starting with a141c4f46628bcf2704823b1f797622cfff0896a3b75c2d6aad4ddce0d593343 not found: ID does not exist" containerID="a141c4f46628bcf2704823b1f797622cfff0896a3b75c2d6aad4ddce0d593343" Oct 06 10:17:52 crc kubenswrapper[4824]: I1006 10:17:52.046230 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a141c4f46628bcf2704823b1f797622cfff0896a3b75c2d6aad4ddce0d593343"} err="failed to get container status \"a141c4f46628bcf2704823b1f797622cfff0896a3b75c2d6aad4ddce0d593343\": rpc error: code = NotFound desc = could not find container \"a141c4f46628bcf2704823b1f797622cfff0896a3b75c2d6aad4ddce0d593343\": container with ID starting with a141c4f46628bcf2704823b1f797622cfff0896a3b75c2d6aad4ddce0d593343 not found: ID does not exist" Oct 06 10:17:52 crc kubenswrapper[4824]: I1006 10:17:52.046260 4824 scope.go:117] "RemoveContainer" containerID="514305e83facb91f5df3d37a4a60700d59550463507ec40e702b13ede6712483" Oct 06 10:17:52 crc kubenswrapper[4824]: E1006 10:17:52.046548 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"514305e83facb91f5df3d37a4a60700d59550463507ec40e702b13ede6712483\": container with ID starting with 514305e83facb91f5df3d37a4a60700d59550463507ec40e702b13ede6712483 not found: ID does not exist" containerID="514305e83facb91f5df3d37a4a60700d59550463507ec40e702b13ede6712483" Oct 06 10:17:52 crc kubenswrapper[4824]: I1006 10:17:52.046586 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"514305e83facb91f5df3d37a4a60700d59550463507ec40e702b13ede6712483"} err="failed to get container status \"514305e83facb91f5df3d37a4a60700d59550463507ec40e702b13ede6712483\": rpc error: code = NotFound desc = could not find container \"514305e83facb91f5df3d37a4a60700d59550463507ec40e702b13ede6712483\": container with ID starting with 514305e83facb91f5df3d37a4a60700d59550463507ec40e702b13ede6712483 not found: ID does not exist" Oct 06 10:17:52 crc kubenswrapper[4824]: I1006 10:17:52.046610 4824 scope.go:117] "RemoveContainer" containerID="e04d4316e7bc61728eecb6ffdc75465ba879c8ca5d3e795f954c9d8e0adee5b0" Oct 06 10:17:52 crc kubenswrapper[4824]: E1006 10:17:52.047046 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e04d4316e7bc61728eecb6ffdc75465ba879c8ca5d3e795f954c9d8e0adee5b0\": container with ID starting with e04d4316e7bc61728eecb6ffdc75465ba879c8ca5d3e795f954c9d8e0adee5b0 not found: ID does not exist" containerID="e04d4316e7bc61728eecb6ffdc75465ba879c8ca5d3e795f954c9d8e0adee5b0" Oct 06 10:17:52 crc kubenswrapper[4824]: I1006 10:17:52.047077 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e04d4316e7bc61728eecb6ffdc75465ba879c8ca5d3e795f954c9d8e0adee5b0"} err="failed to get container status \"e04d4316e7bc61728eecb6ffdc75465ba879c8ca5d3e795f954c9d8e0adee5b0\": rpc error: code = NotFound desc = could not find container \"e04d4316e7bc61728eecb6ffdc75465ba879c8ca5d3e795f954c9d8e0adee5b0\": container with ID starting with e04d4316e7bc61728eecb6ffdc75465ba879c8ca5d3e795f954c9d8e0adee5b0 not found: ID does not exist" Oct 06 10:17:52 crc kubenswrapper[4824]: I1006 10:17:52.048942 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/78c95895-6099-43a1-a8d9-8aec0d0c1011-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "78c95895-6099-43a1-a8d9-8aec0d0c1011" (UID: "78c95895-6099-43a1-a8d9-8aec0d0c1011"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:17:52 crc kubenswrapper[4824]: I1006 10:17:52.128142 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/78c95895-6099-43a1-a8d9-8aec0d0c1011-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 10:17:52 crc kubenswrapper[4824]: I1006 10:17:52.128182 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78c95895-6099-43a1-a8d9-8aec0d0c1011-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 10:17:52 crc kubenswrapper[4824]: I1006 10:17:52.172266 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 06 10:17:52 crc kubenswrapper[4824]: I1006 10:17:52.187641 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 06 10:17:52 crc kubenswrapper[4824]: I1006 10:17:52.197400 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 06 10:17:52 crc kubenswrapper[4824]: E1006 10:17:52.197900 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78c95895-6099-43a1-a8d9-8aec0d0c1011" containerName="ceilometer-central-agent" Oct 06 10:17:52 crc kubenswrapper[4824]: I1006 10:17:52.197929 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="78c95895-6099-43a1-a8d9-8aec0d0c1011" containerName="ceilometer-central-agent" Oct 06 10:17:52 crc kubenswrapper[4824]: E1006 10:17:52.197957 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78c95895-6099-43a1-a8d9-8aec0d0c1011" containerName="proxy-httpd" Oct 06 10:17:52 crc kubenswrapper[4824]: I1006 10:17:52.197967 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="78c95895-6099-43a1-a8d9-8aec0d0c1011" containerName="proxy-httpd" Oct 06 10:17:52 crc kubenswrapper[4824]: E1006 10:17:52.198004 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78c95895-6099-43a1-a8d9-8aec0d0c1011" containerName="sg-core" Oct 06 10:17:52 crc kubenswrapper[4824]: I1006 10:17:52.198014 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="78c95895-6099-43a1-a8d9-8aec0d0c1011" containerName="sg-core" Oct 06 10:17:52 crc kubenswrapper[4824]: E1006 10:17:52.198048 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78c95895-6099-43a1-a8d9-8aec0d0c1011" containerName="ceilometer-notification-agent" Oct 06 10:17:52 crc kubenswrapper[4824]: I1006 10:17:52.198057 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="78c95895-6099-43a1-a8d9-8aec0d0c1011" containerName="ceilometer-notification-agent" Oct 06 10:17:52 crc kubenswrapper[4824]: I1006 10:17:52.198305 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="78c95895-6099-43a1-a8d9-8aec0d0c1011" containerName="ceilometer-notification-agent" Oct 06 10:17:52 crc kubenswrapper[4824]: I1006 10:17:52.198320 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="78c95895-6099-43a1-a8d9-8aec0d0c1011" containerName="proxy-httpd" Oct 06 10:17:52 crc kubenswrapper[4824]: I1006 10:17:52.198338 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="78c95895-6099-43a1-a8d9-8aec0d0c1011" containerName="ceilometer-central-agent" Oct 06 10:17:52 crc kubenswrapper[4824]: I1006 10:17:52.198352 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="78c95895-6099-43a1-a8d9-8aec0d0c1011" containerName="sg-core" Oct 06 10:17:52 crc kubenswrapper[4824]: I1006 10:17:52.200745 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 06 10:17:52 crc kubenswrapper[4824]: I1006 10:17:52.203925 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 06 10:17:52 crc kubenswrapper[4824]: I1006 10:17:52.204144 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 06 10:17:52 crc kubenswrapper[4824]: I1006 10:17:52.217651 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 06 10:17:52 crc kubenswrapper[4824]: I1006 10:17:52.335860 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/539425e3-fd6d-4f71-88b4-80d1a4cf498e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"539425e3-fd6d-4f71-88b4-80d1a4cf498e\") " pod="openstack/ceilometer-0" Oct 06 10:17:52 crc kubenswrapper[4824]: I1006 10:17:52.336033 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/539425e3-fd6d-4f71-88b4-80d1a4cf498e-scripts\") pod \"ceilometer-0\" (UID: \"539425e3-fd6d-4f71-88b4-80d1a4cf498e\") " pod="openstack/ceilometer-0" Oct 06 10:17:52 crc kubenswrapper[4824]: I1006 10:17:52.336099 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/539425e3-fd6d-4f71-88b4-80d1a4cf498e-log-httpd\") pod \"ceilometer-0\" (UID: \"539425e3-fd6d-4f71-88b4-80d1a4cf498e\") " pod="openstack/ceilometer-0" Oct 06 10:17:52 crc kubenswrapper[4824]: I1006 10:17:52.336209 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/539425e3-fd6d-4f71-88b4-80d1a4cf498e-config-data\") pod \"ceilometer-0\" (UID: \"539425e3-fd6d-4f71-88b4-80d1a4cf498e\") " pod="openstack/ceilometer-0" Oct 06 10:17:52 crc kubenswrapper[4824]: I1006 10:17:52.336280 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/539425e3-fd6d-4f71-88b4-80d1a4cf498e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"539425e3-fd6d-4f71-88b4-80d1a4cf498e\") " pod="openstack/ceilometer-0" Oct 06 10:17:52 crc kubenswrapper[4824]: I1006 10:17:52.336372 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/539425e3-fd6d-4f71-88b4-80d1a4cf498e-run-httpd\") pod \"ceilometer-0\" (UID: \"539425e3-fd6d-4f71-88b4-80d1a4cf498e\") " pod="openstack/ceilometer-0" Oct 06 10:17:52 crc kubenswrapper[4824]: I1006 10:17:52.336424 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6c62r\" (UniqueName: \"kubernetes.io/projected/539425e3-fd6d-4f71-88b4-80d1a4cf498e-kube-api-access-6c62r\") pod \"ceilometer-0\" (UID: \"539425e3-fd6d-4f71-88b4-80d1a4cf498e\") " pod="openstack/ceilometer-0" Oct 06 10:17:52 crc kubenswrapper[4824]: I1006 10:17:52.438354 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/539425e3-fd6d-4f71-88b4-80d1a4cf498e-run-httpd\") pod \"ceilometer-0\" (UID: \"539425e3-fd6d-4f71-88b4-80d1a4cf498e\") " pod="openstack/ceilometer-0" Oct 06 10:17:52 crc kubenswrapper[4824]: I1006 10:17:52.438419 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6c62r\" (UniqueName: \"kubernetes.io/projected/539425e3-fd6d-4f71-88b4-80d1a4cf498e-kube-api-access-6c62r\") pod \"ceilometer-0\" (UID: \"539425e3-fd6d-4f71-88b4-80d1a4cf498e\") " pod="openstack/ceilometer-0" Oct 06 10:17:52 crc kubenswrapper[4824]: I1006 10:17:52.438488 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/539425e3-fd6d-4f71-88b4-80d1a4cf498e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"539425e3-fd6d-4f71-88b4-80d1a4cf498e\") " pod="openstack/ceilometer-0" Oct 06 10:17:52 crc kubenswrapper[4824]: I1006 10:17:52.438510 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/539425e3-fd6d-4f71-88b4-80d1a4cf498e-scripts\") pod \"ceilometer-0\" (UID: \"539425e3-fd6d-4f71-88b4-80d1a4cf498e\") " pod="openstack/ceilometer-0" Oct 06 10:17:52 crc kubenswrapper[4824]: I1006 10:17:52.438542 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/539425e3-fd6d-4f71-88b4-80d1a4cf498e-log-httpd\") pod \"ceilometer-0\" (UID: \"539425e3-fd6d-4f71-88b4-80d1a4cf498e\") " pod="openstack/ceilometer-0" Oct 06 10:17:52 crc kubenswrapper[4824]: I1006 10:17:52.438605 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/539425e3-fd6d-4f71-88b4-80d1a4cf498e-config-data\") pod \"ceilometer-0\" (UID: \"539425e3-fd6d-4f71-88b4-80d1a4cf498e\") " pod="openstack/ceilometer-0" Oct 06 10:17:52 crc kubenswrapper[4824]: I1006 10:17:52.438640 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/539425e3-fd6d-4f71-88b4-80d1a4cf498e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"539425e3-fd6d-4f71-88b4-80d1a4cf498e\") " pod="openstack/ceilometer-0" Oct 06 10:17:52 crc kubenswrapper[4824]: I1006 10:17:52.439738 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/539425e3-fd6d-4f71-88b4-80d1a4cf498e-run-httpd\") pod \"ceilometer-0\" (UID: \"539425e3-fd6d-4f71-88b4-80d1a4cf498e\") " pod="openstack/ceilometer-0" Oct 06 10:17:52 crc kubenswrapper[4824]: I1006 10:17:52.439835 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/539425e3-fd6d-4f71-88b4-80d1a4cf498e-log-httpd\") pod \"ceilometer-0\" (UID: \"539425e3-fd6d-4f71-88b4-80d1a4cf498e\") " pod="openstack/ceilometer-0" Oct 06 10:17:52 crc kubenswrapper[4824]: I1006 10:17:52.443088 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/539425e3-fd6d-4f71-88b4-80d1a4cf498e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"539425e3-fd6d-4f71-88b4-80d1a4cf498e\") " pod="openstack/ceilometer-0" Oct 06 10:17:52 crc kubenswrapper[4824]: I1006 10:17:52.443447 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/539425e3-fd6d-4f71-88b4-80d1a4cf498e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"539425e3-fd6d-4f71-88b4-80d1a4cf498e\") " pod="openstack/ceilometer-0" Oct 06 10:17:52 crc kubenswrapper[4824]: I1006 10:17:52.445072 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/539425e3-fd6d-4f71-88b4-80d1a4cf498e-config-data\") pod \"ceilometer-0\" (UID: \"539425e3-fd6d-4f71-88b4-80d1a4cf498e\") " pod="openstack/ceilometer-0" Oct 06 10:17:52 crc kubenswrapper[4824]: I1006 10:17:52.445492 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/539425e3-fd6d-4f71-88b4-80d1a4cf498e-scripts\") pod \"ceilometer-0\" (UID: \"539425e3-fd6d-4f71-88b4-80d1a4cf498e\") " pod="openstack/ceilometer-0" Oct 06 10:17:52 crc kubenswrapper[4824]: I1006 10:17:52.457963 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6c62r\" (UniqueName: \"kubernetes.io/projected/539425e3-fd6d-4f71-88b4-80d1a4cf498e-kube-api-access-6c62r\") pod \"ceilometer-0\" (UID: \"539425e3-fd6d-4f71-88b4-80d1a4cf498e\") " pod="openstack/ceilometer-0" Oct 06 10:17:52 crc kubenswrapper[4824]: I1006 10:17:52.524494 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 06 10:17:53 crc kubenswrapper[4824]: I1006 10:17:53.036808 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 06 10:17:53 crc kubenswrapper[4824]: W1006 10:17:53.042383 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod539425e3_fd6d_4f71_88b4_80d1a4cf498e.slice/crio-e333ef4722e0b96c0abad57cd34ea3fa7c2674dcb5c6260623b43eb916bef206 WatchSource:0}: Error finding container e333ef4722e0b96c0abad57cd34ea3fa7c2674dcb5c6260623b43eb916bef206: Status 404 returned error can't find the container with id e333ef4722e0b96c0abad57cd34ea3fa7c2674dcb5c6260623b43eb916bef206 Oct 06 10:17:53 crc kubenswrapper[4824]: I1006 10:17:53.292830 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="78c95895-6099-43a1-a8d9-8aec0d0c1011" path="/var/lib/kubelet/pods/78c95895-6099-43a1-a8d9-8aec0d0c1011/volumes" Oct 06 10:17:53 crc kubenswrapper[4824]: I1006 10:17:53.848383 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"539425e3-fd6d-4f71-88b4-80d1a4cf498e","Type":"ContainerStarted","Data":"0f485ce06461c003f61fb2c1fb9a1f77664462c0c522974b881b19f5457827f7"} Oct 06 10:17:53 crc kubenswrapper[4824]: I1006 10:17:53.848810 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"539425e3-fd6d-4f71-88b4-80d1a4cf498e","Type":"ContainerStarted","Data":"e333ef4722e0b96c0abad57cd34ea3fa7c2674dcb5c6260623b43eb916bef206"} Oct 06 10:17:54 crc kubenswrapper[4824]: I1006 10:17:54.861099 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"539425e3-fd6d-4f71-88b4-80d1a4cf498e","Type":"ContainerStarted","Data":"3e8f7319f84c0db5b4c4b61cf8240040ec1a879aa1e37c178e94a242dff3714a"} Oct 06 10:17:55 crc kubenswrapper[4824]: I1006 10:17:55.871882 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"539425e3-fd6d-4f71-88b4-80d1a4cf498e","Type":"ContainerStarted","Data":"a8e659ed6196e181667d71808ddae81bad11d0ceeba3a2923ec383131f1b2928"} Oct 06 10:17:56 crc kubenswrapper[4824]: I1006 10:17:56.895337 4824 generic.go:334] "Generic (PLEG): container finished" podID="8a08722d-580d-4dbc-80d5-f33caa5cae5b" containerID="e93f66db5683a4761c779328c5d70b916fcb98226c72e9ab3dc9a4e98af10ebd" exitCode=0 Oct 06 10:17:56 crc kubenswrapper[4824]: I1006 10:17:56.895494 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-7txxj" event={"ID":"8a08722d-580d-4dbc-80d5-f33caa5cae5b","Type":"ContainerDied","Data":"e93f66db5683a4761c779328c5d70b916fcb98226c72e9ab3dc9a4e98af10ebd"} Oct 06 10:17:58 crc kubenswrapper[4824]: I1006 10:17:58.377971 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-7txxj" Oct 06 10:17:58 crc kubenswrapper[4824]: I1006 10:17:58.478613 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-88t2l\" (UniqueName: \"kubernetes.io/projected/8a08722d-580d-4dbc-80d5-f33caa5cae5b-kube-api-access-88t2l\") pod \"8a08722d-580d-4dbc-80d5-f33caa5cae5b\" (UID: \"8a08722d-580d-4dbc-80d5-f33caa5cae5b\") " Oct 06 10:17:58 crc kubenswrapper[4824]: I1006 10:17:58.478748 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a08722d-580d-4dbc-80d5-f33caa5cae5b-config-data\") pod \"8a08722d-580d-4dbc-80d5-f33caa5cae5b\" (UID: \"8a08722d-580d-4dbc-80d5-f33caa5cae5b\") " Oct 06 10:17:58 crc kubenswrapper[4824]: I1006 10:17:58.478862 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a08722d-580d-4dbc-80d5-f33caa5cae5b-combined-ca-bundle\") pod \"8a08722d-580d-4dbc-80d5-f33caa5cae5b\" (UID: \"8a08722d-580d-4dbc-80d5-f33caa5cae5b\") " Oct 06 10:17:58 crc kubenswrapper[4824]: I1006 10:17:58.478931 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8a08722d-580d-4dbc-80d5-f33caa5cae5b-scripts\") pod \"8a08722d-580d-4dbc-80d5-f33caa5cae5b\" (UID: \"8a08722d-580d-4dbc-80d5-f33caa5cae5b\") " Oct 06 10:17:58 crc kubenswrapper[4824]: I1006 10:17:58.486697 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a08722d-580d-4dbc-80d5-f33caa5cae5b-scripts" (OuterVolumeSpecName: "scripts") pod "8a08722d-580d-4dbc-80d5-f33caa5cae5b" (UID: "8a08722d-580d-4dbc-80d5-f33caa5cae5b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:17:58 crc kubenswrapper[4824]: I1006 10:17:58.487442 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8a08722d-580d-4dbc-80d5-f33caa5cae5b-kube-api-access-88t2l" (OuterVolumeSpecName: "kube-api-access-88t2l") pod "8a08722d-580d-4dbc-80d5-f33caa5cae5b" (UID: "8a08722d-580d-4dbc-80d5-f33caa5cae5b"). InnerVolumeSpecName "kube-api-access-88t2l". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:17:58 crc kubenswrapper[4824]: I1006 10:17:58.513163 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a08722d-580d-4dbc-80d5-f33caa5cae5b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8a08722d-580d-4dbc-80d5-f33caa5cae5b" (UID: "8a08722d-580d-4dbc-80d5-f33caa5cae5b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:17:58 crc kubenswrapper[4824]: I1006 10:17:58.514962 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a08722d-580d-4dbc-80d5-f33caa5cae5b-config-data" (OuterVolumeSpecName: "config-data") pod "8a08722d-580d-4dbc-80d5-f33caa5cae5b" (UID: "8a08722d-580d-4dbc-80d5-f33caa5cae5b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:17:58 crc kubenswrapper[4824]: I1006 10:17:58.581213 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-88t2l\" (UniqueName: \"kubernetes.io/projected/8a08722d-580d-4dbc-80d5-f33caa5cae5b-kube-api-access-88t2l\") on node \"crc\" DevicePath \"\"" Oct 06 10:17:58 crc kubenswrapper[4824]: I1006 10:17:58.581578 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a08722d-580d-4dbc-80d5-f33caa5cae5b-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 10:17:58 crc kubenswrapper[4824]: I1006 10:17:58.581592 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a08722d-580d-4dbc-80d5-f33caa5cae5b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 10:17:58 crc kubenswrapper[4824]: I1006 10:17:58.581604 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8a08722d-580d-4dbc-80d5-f33caa5cae5b-scripts\") on node \"crc\" DevicePath \"\"" Oct 06 10:17:58 crc kubenswrapper[4824]: I1006 10:17:58.945354 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-7txxj" Oct 06 10:17:58 crc kubenswrapper[4824]: I1006 10:17:58.945353 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-7txxj" event={"ID":"8a08722d-580d-4dbc-80d5-f33caa5cae5b","Type":"ContainerDied","Data":"5855e1d64bb16ed534d2c29a3d1a45535528de8f1c3a0ab71a0ac1200dad1176"} Oct 06 10:17:58 crc kubenswrapper[4824]: I1006 10:17:58.945565 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5855e1d64bb16ed534d2c29a3d1a45535528de8f1c3a0ab71a0ac1200dad1176" Oct 06 10:17:58 crc kubenswrapper[4824]: I1006 10:17:58.954419 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"539425e3-fd6d-4f71-88b4-80d1a4cf498e","Type":"ContainerStarted","Data":"6e1bfbb3d87cb6e39782a8b7402577417a374efcb86730702652bf68b748e638"} Oct 06 10:17:59 crc kubenswrapper[4824]: I1006 10:17:59.051056 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.205160743 podStartE2EDuration="7.051036454s" podCreationTimestamp="2025-10-06 10:17:52 +0000 UTC" firstStartedPulling="2025-10-06 10:17:53.045223939 +0000 UTC m=+1182.409646800" lastFinishedPulling="2025-10-06 10:17:57.89109961 +0000 UTC m=+1187.255522511" observedRunningTime="2025-10-06 10:17:59.001257132 +0000 UTC m=+1188.365680003" watchObservedRunningTime="2025-10-06 10:17:59.051036454 +0000 UTC m=+1188.415459315" Oct 06 10:17:59 crc kubenswrapper[4824]: I1006 10:17:59.053865 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 06 10:17:59 crc kubenswrapper[4824]: E1006 10:17:59.054231 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a08722d-580d-4dbc-80d5-f33caa5cae5b" containerName="nova-cell0-conductor-db-sync" Oct 06 10:17:59 crc kubenswrapper[4824]: I1006 10:17:59.054247 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a08722d-580d-4dbc-80d5-f33caa5cae5b" containerName="nova-cell0-conductor-db-sync" Oct 06 10:17:59 crc kubenswrapper[4824]: I1006 10:17:59.054429 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a08722d-580d-4dbc-80d5-f33caa5cae5b" containerName="nova-cell0-conductor-db-sync" Oct 06 10:17:59 crc kubenswrapper[4824]: I1006 10:17:59.055006 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 06 10:17:59 crc kubenswrapper[4824]: I1006 10:17:59.059615 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 06 10:17:59 crc kubenswrapper[4824]: I1006 10:17:59.060107 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-wdpnq" Oct 06 10:17:59 crc kubenswrapper[4824]: I1006 10:17:59.085549 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 06 10:17:59 crc kubenswrapper[4824]: I1006 10:17:59.089921 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e76a912-576d-41a3-86b0-f7fc6a6814a2-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"2e76a912-576d-41a3-86b0-f7fc6a6814a2\") " pod="openstack/nova-cell0-conductor-0" Oct 06 10:17:59 crc kubenswrapper[4824]: I1006 10:17:59.090210 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b852b\" (UniqueName: \"kubernetes.io/projected/2e76a912-576d-41a3-86b0-f7fc6a6814a2-kube-api-access-b852b\") pod \"nova-cell0-conductor-0\" (UID: \"2e76a912-576d-41a3-86b0-f7fc6a6814a2\") " pod="openstack/nova-cell0-conductor-0" Oct 06 10:17:59 crc kubenswrapper[4824]: I1006 10:17:59.090420 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e76a912-576d-41a3-86b0-f7fc6a6814a2-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"2e76a912-576d-41a3-86b0-f7fc6a6814a2\") " pod="openstack/nova-cell0-conductor-0" Oct 06 10:17:59 crc kubenswrapper[4824]: I1006 10:17:59.192402 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b852b\" (UniqueName: \"kubernetes.io/projected/2e76a912-576d-41a3-86b0-f7fc6a6814a2-kube-api-access-b852b\") pod \"nova-cell0-conductor-0\" (UID: \"2e76a912-576d-41a3-86b0-f7fc6a6814a2\") " pod="openstack/nova-cell0-conductor-0" Oct 06 10:17:59 crc kubenswrapper[4824]: I1006 10:17:59.192489 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e76a912-576d-41a3-86b0-f7fc6a6814a2-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"2e76a912-576d-41a3-86b0-f7fc6a6814a2\") " pod="openstack/nova-cell0-conductor-0" Oct 06 10:17:59 crc kubenswrapper[4824]: I1006 10:17:59.192557 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e76a912-576d-41a3-86b0-f7fc6a6814a2-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"2e76a912-576d-41a3-86b0-f7fc6a6814a2\") " pod="openstack/nova-cell0-conductor-0" Oct 06 10:17:59 crc kubenswrapper[4824]: I1006 10:17:59.196808 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e76a912-576d-41a3-86b0-f7fc6a6814a2-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"2e76a912-576d-41a3-86b0-f7fc6a6814a2\") " pod="openstack/nova-cell0-conductor-0" Oct 06 10:17:59 crc kubenswrapper[4824]: I1006 10:17:59.209423 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e76a912-576d-41a3-86b0-f7fc6a6814a2-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"2e76a912-576d-41a3-86b0-f7fc6a6814a2\") " pod="openstack/nova-cell0-conductor-0" Oct 06 10:17:59 crc kubenswrapper[4824]: I1006 10:17:59.209826 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b852b\" (UniqueName: \"kubernetes.io/projected/2e76a912-576d-41a3-86b0-f7fc6a6814a2-kube-api-access-b852b\") pod \"nova-cell0-conductor-0\" (UID: \"2e76a912-576d-41a3-86b0-f7fc6a6814a2\") " pod="openstack/nova-cell0-conductor-0" Oct 06 10:17:59 crc kubenswrapper[4824]: I1006 10:17:59.372690 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 06 10:17:59 crc kubenswrapper[4824]: I1006 10:17:59.873777 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 06 10:17:59 crc kubenswrapper[4824]: I1006 10:17:59.968529 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"2e76a912-576d-41a3-86b0-f7fc6a6814a2","Type":"ContainerStarted","Data":"285fae0678c81dcf8d8fec87d7436fcbdef0f6721a21bcb1a743fc8edca56d7c"} Oct 06 10:17:59 crc kubenswrapper[4824]: I1006 10:17:59.968649 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 06 10:18:00 crc kubenswrapper[4824]: I1006 10:18:00.990764 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"2e76a912-576d-41a3-86b0-f7fc6a6814a2","Type":"ContainerStarted","Data":"ef7f4ab8f0d75d7dbbf73383de3a8a2bd8235ba0be5d13cbd412702c381949a7"} Oct 06 10:18:00 crc kubenswrapper[4824]: I1006 10:18:00.991215 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Oct 06 10:18:01 crc kubenswrapper[4824]: I1006 10:18:01.022209 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.022181632 podStartE2EDuration="2.022181632s" podCreationTimestamp="2025-10-06 10:17:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:18:01.004321051 +0000 UTC m=+1190.368743952" watchObservedRunningTime="2025-10-06 10:18:01.022181632 +0000 UTC m=+1190.386604533" Oct 06 10:18:09 crc kubenswrapper[4824]: I1006 10:18:09.426660 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Oct 06 10:18:10 crc kubenswrapper[4824]: I1006 10:18:10.020697 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-zq5td"] Oct 06 10:18:10 crc kubenswrapper[4824]: I1006 10:18:10.023522 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-zq5td" Oct 06 10:18:10 crc kubenswrapper[4824]: I1006 10:18:10.038887 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-zq5td"] Oct 06 10:18:10 crc kubenswrapper[4824]: I1006 10:18:10.055343 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Oct 06 10:18:10 crc kubenswrapper[4824]: I1006 10:18:10.055763 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Oct 06 10:18:10 crc kubenswrapper[4824]: I1006 10:18:10.067477 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd95e609-bd54-4234-a449-66cd65e2558a-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-zq5td\" (UID: \"cd95e609-bd54-4234-a449-66cd65e2558a\") " pod="openstack/nova-cell0-cell-mapping-zq5td" Oct 06 10:18:10 crc kubenswrapper[4824]: I1006 10:18:10.067556 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd95e609-bd54-4234-a449-66cd65e2558a-config-data\") pod \"nova-cell0-cell-mapping-zq5td\" (UID: \"cd95e609-bd54-4234-a449-66cd65e2558a\") " pod="openstack/nova-cell0-cell-mapping-zq5td" Oct 06 10:18:10 crc kubenswrapper[4824]: I1006 10:18:10.067650 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cd95e609-bd54-4234-a449-66cd65e2558a-scripts\") pod \"nova-cell0-cell-mapping-zq5td\" (UID: \"cd95e609-bd54-4234-a449-66cd65e2558a\") " pod="openstack/nova-cell0-cell-mapping-zq5td" Oct 06 10:18:10 crc kubenswrapper[4824]: I1006 10:18:10.067939 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vxkv2\" (UniqueName: \"kubernetes.io/projected/cd95e609-bd54-4234-a449-66cd65e2558a-kube-api-access-vxkv2\") pod \"nova-cell0-cell-mapping-zq5td\" (UID: \"cd95e609-bd54-4234-a449-66cd65e2558a\") " pod="openstack/nova-cell0-cell-mapping-zq5td" Oct 06 10:18:10 crc kubenswrapper[4824]: I1006 10:18:10.173254 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vxkv2\" (UniqueName: \"kubernetes.io/projected/cd95e609-bd54-4234-a449-66cd65e2558a-kube-api-access-vxkv2\") pod \"nova-cell0-cell-mapping-zq5td\" (UID: \"cd95e609-bd54-4234-a449-66cd65e2558a\") " pod="openstack/nova-cell0-cell-mapping-zq5td" Oct 06 10:18:10 crc kubenswrapper[4824]: I1006 10:18:10.173365 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd95e609-bd54-4234-a449-66cd65e2558a-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-zq5td\" (UID: \"cd95e609-bd54-4234-a449-66cd65e2558a\") " pod="openstack/nova-cell0-cell-mapping-zq5td" Oct 06 10:18:10 crc kubenswrapper[4824]: I1006 10:18:10.173391 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd95e609-bd54-4234-a449-66cd65e2558a-config-data\") pod \"nova-cell0-cell-mapping-zq5td\" (UID: \"cd95e609-bd54-4234-a449-66cd65e2558a\") " pod="openstack/nova-cell0-cell-mapping-zq5td" Oct 06 10:18:10 crc kubenswrapper[4824]: I1006 10:18:10.173436 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cd95e609-bd54-4234-a449-66cd65e2558a-scripts\") pod \"nova-cell0-cell-mapping-zq5td\" (UID: \"cd95e609-bd54-4234-a449-66cd65e2558a\") " pod="openstack/nova-cell0-cell-mapping-zq5td" Oct 06 10:18:10 crc kubenswrapper[4824]: I1006 10:18:10.183809 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd95e609-bd54-4234-a449-66cd65e2558a-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-zq5td\" (UID: \"cd95e609-bd54-4234-a449-66cd65e2558a\") " pod="openstack/nova-cell0-cell-mapping-zq5td" Oct 06 10:18:10 crc kubenswrapper[4824]: I1006 10:18:10.201487 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cd95e609-bd54-4234-a449-66cd65e2558a-scripts\") pod \"nova-cell0-cell-mapping-zq5td\" (UID: \"cd95e609-bd54-4234-a449-66cd65e2558a\") " pod="openstack/nova-cell0-cell-mapping-zq5td" Oct 06 10:18:10 crc kubenswrapper[4824]: I1006 10:18:10.218035 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd95e609-bd54-4234-a449-66cd65e2558a-config-data\") pod \"nova-cell0-cell-mapping-zq5td\" (UID: \"cd95e609-bd54-4234-a449-66cd65e2558a\") " pod="openstack/nova-cell0-cell-mapping-zq5td" Oct 06 10:18:10 crc kubenswrapper[4824]: I1006 10:18:10.218589 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vxkv2\" (UniqueName: \"kubernetes.io/projected/cd95e609-bd54-4234-a449-66cd65e2558a-kube-api-access-vxkv2\") pod \"nova-cell0-cell-mapping-zq5td\" (UID: \"cd95e609-bd54-4234-a449-66cd65e2558a\") " pod="openstack/nova-cell0-cell-mapping-zq5td" Oct 06 10:18:10 crc kubenswrapper[4824]: I1006 10:18:10.282749 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 06 10:18:10 crc kubenswrapper[4824]: I1006 10:18:10.284423 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 06 10:18:10 crc kubenswrapper[4824]: I1006 10:18:10.290751 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 06 10:18:10 crc kubenswrapper[4824]: I1006 10:18:10.294428 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 06 10:18:10 crc kubenswrapper[4824]: I1006 10:18:10.295801 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 06 10:18:10 crc kubenswrapper[4824]: I1006 10:18:10.309263 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 06 10:18:10 crc kubenswrapper[4824]: I1006 10:18:10.355702 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 06 10:18:10 crc kubenswrapper[4824]: I1006 10:18:10.377247 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf250d98-9985-42e1-b5d8-d3cf5e660869-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"cf250d98-9985-42e1-b5d8-d3cf5e660869\") " pod="openstack/nova-metadata-0" Oct 06 10:18:10 crc kubenswrapper[4824]: I1006 10:18:10.377367 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hdzkz\" (UniqueName: \"kubernetes.io/projected/d56c5ee8-59b8-45d6-9078-29423490982f-kube-api-access-hdzkz\") pod \"nova-api-0\" (UID: \"d56c5ee8-59b8-45d6-9078-29423490982f\") " pod="openstack/nova-api-0" Oct 06 10:18:10 crc kubenswrapper[4824]: I1006 10:18:10.377429 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d56c5ee8-59b8-45d6-9078-29423490982f-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"d56c5ee8-59b8-45d6-9078-29423490982f\") " pod="openstack/nova-api-0" Oct 06 10:18:10 crc kubenswrapper[4824]: I1006 10:18:10.377498 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mxd2k\" (UniqueName: \"kubernetes.io/projected/cf250d98-9985-42e1-b5d8-d3cf5e660869-kube-api-access-mxd2k\") pod \"nova-metadata-0\" (UID: \"cf250d98-9985-42e1-b5d8-d3cf5e660869\") " pod="openstack/nova-metadata-0" Oct 06 10:18:10 crc kubenswrapper[4824]: I1006 10:18:10.377566 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d56c5ee8-59b8-45d6-9078-29423490982f-config-data\") pod \"nova-api-0\" (UID: \"d56c5ee8-59b8-45d6-9078-29423490982f\") " pod="openstack/nova-api-0" Oct 06 10:18:10 crc kubenswrapper[4824]: I1006 10:18:10.377677 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cf250d98-9985-42e1-b5d8-d3cf5e660869-logs\") pod \"nova-metadata-0\" (UID: \"cf250d98-9985-42e1-b5d8-d3cf5e660869\") " pod="openstack/nova-metadata-0" Oct 06 10:18:10 crc kubenswrapper[4824]: I1006 10:18:10.377717 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf250d98-9985-42e1-b5d8-d3cf5e660869-config-data\") pod \"nova-metadata-0\" (UID: \"cf250d98-9985-42e1-b5d8-d3cf5e660869\") " pod="openstack/nova-metadata-0" Oct 06 10:18:10 crc kubenswrapper[4824]: I1006 10:18:10.377756 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d56c5ee8-59b8-45d6-9078-29423490982f-logs\") pod \"nova-api-0\" (UID: \"d56c5ee8-59b8-45d6-9078-29423490982f\") " pod="openstack/nova-api-0" Oct 06 10:18:10 crc kubenswrapper[4824]: I1006 10:18:10.384073 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-zq5td" Oct 06 10:18:10 crc kubenswrapper[4824]: I1006 10:18:10.435826 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 06 10:18:10 crc kubenswrapper[4824]: I1006 10:18:10.478892 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mxd2k\" (UniqueName: \"kubernetes.io/projected/cf250d98-9985-42e1-b5d8-d3cf5e660869-kube-api-access-mxd2k\") pod \"nova-metadata-0\" (UID: \"cf250d98-9985-42e1-b5d8-d3cf5e660869\") " pod="openstack/nova-metadata-0" Oct 06 10:18:10 crc kubenswrapper[4824]: I1006 10:18:10.478959 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d56c5ee8-59b8-45d6-9078-29423490982f-config-data\") pod \"nova-api-0\" (UID: \"d56c5ee8-59b8-45d6-9078-29423490982f\") " pod="openstack/nova-api-0" Oct 06 10:18:10 crc kubenswrapper[4824]: I1006 10:18:10.479035 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cf250d98-9985-42e1-b5d8-d3cf5e660869-logs\") pod \"nova-metadata-0\" (UID: \"cf250d98-9985-42e1-b5d8-d3cf5e660869\") " pod="openstack/nova-metadata-0" Oct 06 10:18:10 crc kubenswrapper[4824]: I1006 10:18:10.479059 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf250d98-9985-42e1-b5d8-d3cf5e660869-config-data\") pod \"nova-metadata-0\" (UID: \"cf250d98-9985-42e1-b5d8-d3cf5e660869\") " pod="openstack/nova-metadata-0" Oct 06 10:18:10 crc kubenswrapper[4824]: I1006 10:18:10.479086 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d56c5ee8-59b8-45d6-9078-29423490982f-logs\") pod \"nova-api-0\" (UID: \"d56c5ee8-59b8-45d6-9078-29423490982f\") " pod="openstack/nova-api-0" Oct 06 10:18:10 crc kubenswrapper[4824]: I1006 10:18:10.479114 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf250d98-9985-42e1-b5d8-d3cf5e660869-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"cf250d98-9985-42e1-b5d8-d3cf5e660869\") " pod="openstack/nova-metadata-0" Oct 06 10:18:10 crc kubenswrapper[4824]: I1006 10:18:10.479158 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hdzkz\" (UniqueName: \"kubernetes.io/projected/d56c5ee8-59b8-45d6-9078-29423490982f-kube-api-access-hdzkz\") pod \"nova-api-0\" (UID: \"d56c5ee8-59b8-45d6-9078-29423490982f\") " pod="openstack/nova-api-0" Oct 06 10:18:10 crc kubenswrapper[4824]: I1006 10:18:10.479188 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d56c5ee8-59b8-45d6-9078-29423490982f-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"d56c5ee8-59b8-45d6-9078-29423490982f\") " pod="openstack/nova-api-0" Oct 06 10:18:10 crc kubenswrapper[4824]: I1006 10:18:10.490019 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d56c5ee8-59b8-45d6-9078-29423490982f-logs\") pod \"nova-api-0\" (UID: \"d56c5ee8-59b8-45d6-9078-29423490982f\") " pod="openstack/nova-api-0" Oct 06 10:18:10 crc kubenswrapper[4824]: I1006 10:18:10.490637 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cf250d98-9985-42e1-b5d8-d3cf5e660869-logs\") pod \"nova-metadata-0\" (UID: \"cf250d98-9985-42e1-b5d8-d3cf5e660869\") " pod="openstack/nova-metadata-0" Oct 06 10:18:10 crc kubenswrapper[4824]: I1006 10:18:10.511628 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf250d98-9985-42e1-b5d8-d3cf5e660869-config-data\") pod \"nova-metadata-0\" (UID: \"cf250d98-9985-42e1-b5d8-d3cf5e660869\") " pod="openstack/nova-metadata-0" Oct 06 10:18:10 crc kubenswrapper[4824]: I1006 10:18:10.512064 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d56c5ee8-59b8-45d6-9078-29423490982f-config-data\") pod \"nova-api-0\" (UID: \"d56c5ee8-59b8-45d6-9078-29423490982f\") " pod="openstack/nova-api-0" Oct 06 10:18:10 crc kubenswrapper[4824]: I1006 10:18:10.538084 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d56c5ee8-59b8-45d6-9078-29423490982f-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"d56c5ee8-59b8-45d6-9078-29423490982f\") " pod="openstack/nova-api-0" Oct 06 10:18:10 crc kubenswrapper[4824]: I1006 10:18:10.569545 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hdzkz\" (UniqueName: \"kubernetes.io/projected/d56c5ee8-59b8-45d6-9078-29423490982f-kube-api-access-hdzkz\") pod \"nova-api-0\" (UID: \"d56c5ee8-59b8-45d6-9078-29423490982f\") " pod="openstack/nova-api-0" Oct 06 10:18:10 crc kubenswrapper[4824]: I1006 10:18:10.570239 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf250d98-9985-42e1-b5d8-d3cf5e660869-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"cf250d98-9985-42e1-b5d8-d3cf5e660869\") " pod="openstack/nova-metadata-0" Oct 06 10:18:10 crc kubenswrapper[4824]: I1006 10:18:10.571446 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mxd2k\" (UniqueName: \"kubernetes.io/projected/cf250d98-9985-42e1-b5d8-d3cf5e660869-kube-api-access-mxd2k\") pod \"nova-metadata-0\" (UID: \"cf250d98-9985-42e1-b5d8-d3cf5e660869\") " pod="openstack/nova-metadata-0" Oct 06 10:18:10 crc kubenswrapper[4824]: I1006 10:18:10.591891 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-vg6jp"] Oct 06 10:18:10 crc kubenswrapper[4824]: I1006 10:18:10.593671 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-845d6d6f59-vg6jp" Oct 06 10:18:10 crc kubenswrapper[4824]: I1006 10:18:10.602574 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 06 10:18:10 crc kubenswrapper[4824]: I1006 10:18:10.603802 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 06 10:18:10 crc kubenswrapper[4824]: I1006 10:18:10.610283 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 06 10:18:10 crc kubenswrapper[4824]: I1006 10:18:10.617134 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 06 10:18:10 crc kubenswrapper[4824]: I1006 10:18:10.623315 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-vg6jp"] Oct 06 10:18:10 crc kubenswrapper[4824]: I1006 10:18:10.660635 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 06 10:18:10 crc kubenswrapper[4824]: I1006 10:18:10.668442 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 06 10:18:10 crc kubenswrapper[4824]: I1006 10:18:10.670220 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 06 10:18:10 crc kubenswrapper[4824]: I1006 10:18:10.674453 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Oct 06 10:18:10 crc kubenswrapper[4824]: I1006 10:18:10.715862 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 06 10:18:10 crc kubenswrapper[4824]: I1006 10:18:10.747483 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 06 10:18:10 crc kubenswrapper[4824]: I1006 10:18:10.786025 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/627f8bd8-ea33-486d-ac9e-2cc7c0b2ae15-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"627f8bd8-ea33-486d-ac9e-2cc7c0b2ae15\") " pod="openstack/nova-cell1-novncproxy-0" Oct 06 10:18:10 crc kubenswrapper[4824]: I1006 10:18:10.786099 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27cc4522-3bd4-41bd-9fb7-624c634a5288-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"27cc4522-3bd4-41bd-9fb7-624c634a5288\") " pod="openstack/nova-scheduler-0" Oct 06 10:18:10 crc kubenswrapper[4824]: I1006 10:18:10.786145 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/48fc2f94-426b-4056-88db-4445309b7e08-config\") pod \"dnsmasq-dns-845d6d6f59-vg6jp\" (UID: \"48fc2f94-426b-4056-88db-4445309b7e08\") " pod="openstack/dnsmasq-dns-845d6d6f59-vg6jp" Oct 06 10:18:10 crc kubenswrapper[4824]: I1006 10:18:10.786171 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4wvk8\" (UniqueName: \"kubernetes.io/projected/48fc2f94-426b-4056-88db-4445309b7e08-kube-api-access-4wvk8\") pod \"dnsmasq-dns-845d6d6f59-vg6jp\" (UID: \"48fc2f94-426b-4056-88db-4445309b7e08\") " pod="openstack/dnsmasq-dns-845d6d6f59-vg6jp" Oct 06 10:18:10 crc kubenswrapper[4824]: I1006 10:18:10.786215 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cxp5h\" (UniqueName: \"kubernetes.io/projected/627f8bd8-ea33-486d-ac9e-2cc7c0b2ae15-kube-api-access-cxp5h\") pod \"nova-cell1-novncproxy-0\" (UID: \"627f8bd8-ea33-486d-ac9e-2cc7c0b2ae15\") " pod="openstack/nova-cell1-novncproxy-0" Oct 06 10:18:10 crc kubenswrapper[4824]: I1006 10:18:10.786245 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5cbxw\" (UniqueName: \"kubernetes.io/projected/27cc4522-3bd4-41bd-9fb7-624c634a5288-kube-api-access-5cbxw\") pod \"nova-scheduler-0\" (UID: \"27cc4522-3bd4-41bd-9fb7-624c634a5288\") " pod="openstack/nova-scheduler-0" Oct 06 10:18:10 crc kubenswrapper[4824]: I1006 10:18:10.786309 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/627f8bd8-ea33-486d-ac9e-2cc7c0b2ae15-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"627f8bd8-ea33-486d-ac9e-2cc7c0b2ae15\") " pod="openstack/nova-cell1-novncproxy-0" Oct 06 10:18:10 crc kubenswrapper[4824]: I1006 10:18:10.786333 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/48fc2f94-426b-4056-88db-4445309b7e08-dns-svc\") pod \"dnsmasq-dns-845d6d6f59-vg6jp\" (UID: \"48fc2f94-426b-4056-88db-4445309b7e08\") " pod="openstack/dnsmasq-dns-845d6d6f59-vg6jp" Oct 06 10:18:10 crc kubenswrapper[4824]: I1006 10:18:10.786370 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/48fc2f94-426b-4056-88db-4445309b7e08-dns-swift-storage-0\") pod \"dnsmasq-dns-845d6d6f59-vg6jp\" (UID: \"48fc2f94-426b-4056-88db-4445309b7e08\") " pod="openstack/dnsmasq-dns-845d6d6f59-vg6jp" Oct 06 10:18:10 crc kubenswrapper[4824]: I1006 10:18:10.786398 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/48fc2f94-426b-4056-88db-4445309b7e08-ovsdbserver-sb\") pod \"dnsmasq-dns-845d6d6f59-vg6jp\" (UID: \"48fc2f94-426b-4056-88db-4445309b7e08\") " pod="openstack/dnsmasq-dns-845d6d6f59-vg6jp" Oct 06 10:18:10 crc kubenswrapper[4824]: I1006 10:18:10.786440 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/48fc2f94-426b-4056-88db-4445309b7e08-ovsdbserver-nb\") pod \"dnsmasq-dns-845d6d6f59-vg6jp\" (UID: \"48fc2f94-426b-4056-88db-4445309b7e08\") " pod="openstack/dnsmasq-dns-845d6d6f59-vg6jp" Oct 06 10:18:10 crc kubenswrapper[4824]: I1006 10:18:10.786486 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/27cc4522-3bd4-41bd-9fb7-624c634a5288-config-data\") pod \"nova-scheduler-0\" (UID: \"27cc4522-3bd4-41bd-9fb7-624c634a5288\") " pod="openstack/nova-scheduler-0" Oct 06 10:18:10 crc kubenswrapper[4824]: I1006 10:18:10.891134 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/627f8bd8-ea33-486d-ac9e-2cc7c0b2ae15-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"627f8bd8-ea33-486d-ac9e-2cc7c0b2ae15\") " pod="openstack/nova-cell1-novncproxy-0" Oct 06 10:18:10 crc kubenswrapper[4824]: I1006 10:18:10.891209 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/48fc2f94-426b-4056-88db-4445309b7e08-dns-svc\") pod \"dnsmasq-dns-845d6d6f59-vg6jp\" (UID: \"48fc2f94-426b-4056-88db-4445309b7e08\") " pod="openstack/dnsmasq-dns-845d6d6f59-vg6jp" Oct 06 10:18:10 crc kubenswrapper[4824]: I1006 10:18:10.891263 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/48fc2f94-426b-4056-88db-4445309b7e08-dns-swift-storage-0\") pod \"dnsmasq-dns-845d6d6f59-vg6jp\" (UID: \"48fc2f94-426b-4056-88db-4445309b7e08\") " pod="openstack/dnsmasq-dns-845d6d6f59-vg6jp" Oct 06 10:18:10 crc kubenswrapper[4824]: I1006 10:18:10.891312 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/48fc2f94-426b-4056-88db-4445309b7e08-ovsdbserver-sb\") pod \"dnsmasq-dns-845d6d6f59-vg6jp\" (UID: \"48fc2f94-426b-4056-88db-4445309b7e08\") " pod="openstack/dnsmasq-dns-845d6d6f59-vg6jp" Oct 06 10:18:10 crc kubenswrapper[4824]: I1006 10:18:10.891367 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/48fc2f94-426b-4056-88db-4445309b7e08-ovsdbserver-nb\") pod \"dnsmasq-dns-845d6d6f59-vg6jp\" (UID: \"48fc2f94-426b-4056-88db-4445309b7e08\") " pod="openstack/dnsmasq-dns-845d6d6f59-vg6jp" Oct 06 10:18:10 crc kubenswrapper[4824]: I1006 10:18:10.891429 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/27cc4522-3bd4-41bd-9fb7-624c634a5288-config-data\") pod \"nova-scheduler-0\" (UID: \"27cc4522-3bd4-41bd-9fb7-624c634a5288\") " pod="openstack/nova-scheduler-0" Oct 06 10:18:10 crc kubenswrapper[4824]: I1006 10:18:10.891471 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/627f8bd8-ea33-486d-ac9e-2cc7c0b2ae15-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"627f8bd8-ea33-486d-ac9e-2cc7c0b2ae15\") " pod="openstack/nova-cell1-novncproxy-0" Oct 06 10:18:10 crc kubenswrapper[4824]: I1006 10:18:10.891509 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27cc4522-3bd4-41bd-9fb7-624c634a5288-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"27cc4522-3bd4-41bd-9fb7-624c634a5288\") " pod="openstack/nova-scheduler-0" Oct 06 10:18:10 crc kubenswrapper[4824]: I1006 10:18:10.891535 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/48fc2f94-426b-4056-88db-4445309b7e08-config\") pod \"dnsmasq-dns-845d6d6f59-vg6jp\" (UID: \"48fc2f94-426b-4056-88db-4445309b7e08\") " pod="openstack/dnsmasq-dns-845d6d6f59-vg6jp" Oct 06 10:18:10 crc kubenswrapper[4824]: I1006 10:18:10.891565 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4wvk8\" (UniqueName: \"kubernetes.io/projected/48fc2f94-426b-4056-88db-4445309b7e08-kube-api-access-4wvk8\") pod \"dnsmasq-dns-845d6d6f59-vg6jp\" (UID: \"48fc2f94-426b-4056-88db-4445309b7e08\") " pod="openstack/dnsmasq-dns-845d6d6f59-vg6jp" Oct 06 10:18:10 crc kubenswrapper[4824]: I1006 10:18:10.891615 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cxp5h\" (UniqueName: \"kubernetes.io/projected/627f8bd8-ea33-486d-ac9e-2cc7c0b2ae15-kube-api-access-cxp5h\") pod \"nova-cell1-novncproxy-0\" (UID: \"627f8bd8-ea33-486d-ac9e-2cc7c0b2ae15\") " pod="openstack/nova-cell1-novncproxy-0" Oct 06 10:18:10 crc kubenswrapper[4824]: I1006 10:18:10.891679 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5cbxw\" (UniqueName: \"kubernetes.io/projected/27cc4522-3bd4-41bd-9fb7-624c634a5288-kube-api-access-5cbxw\") pod \"nova-scheduler-0\" (UID: \"27cc4522-3bd4-41bd-9fb7-624c634a5288\") " pod="openstack/nova-scheduler-0" Oct 06 10:18:10 crc kubenswrapper[4824]: I1006 10:18:10.894046 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/48fc2f94-426b-4056-88db-4445309b7e08-dns-swift-storage-0\") pod \"dnsmasq-dns-845d6d6f59-vg6jp\" (UID: \"48fc2f94-426b-4056-88db-4445309b7e08\") " pod="openstack/dnsmasq-dns-845d6d6f59-vg6jp" Oct 06 10:18:10 crc kubenswrapper[4824]: I1006 10:18:10.894331 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/48fc2f94-426b-4056-88db-4445309b7e08-dns-svc\") pod \"dnsmasq-dns-845d6d6f59-vg6jp\" (UID: \"48fc2f94-426b-4056-88db-4445309b7e08\") " pod="openstack/dnsmasq-dns-845d6d6f59-vg6jp" Oct 06 10:18:10 crc kubenswrapper[4824]: I1006 10:18:10.894625 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/48fc2f94-426b-4056-88db-4445309b7e08-ovsdbserver-sb\") pod \"dnsmasq-dns-845d6d6f59-vg6jp\" (UID: \"48fc2f94-426b-4056-88db-4445309b7e08\") " pod="openstack/dnsmasq-dns-845d6d6f59-vg6jp" Oct 06 10:18:10 crc kubenswrapper[4824]: I1006 10:18:10.895159 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/48fc2f94-426b-4056-88db-4445309b7e08-ovsdbserver-nb\") pod \"dnsmasq-dns-845d6d6f59-vg6jp\" (UID: \"48fc2f94-426b-4056-88db-4445309b7e08\") " pod="openstack/dnsmasq-dns-845d6d6f59-vg6jp" Oct 06 10:18:10 crc kubenswrapper[4824]: I1006 10:18:10.896804 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/48fc2f94-426b-4056-88db-4445309b7e08-config\") pod \"dnsmasq-dns-845d6d6f59-vg6jp\" (UID: \"48fc2f94-426b-4056-88db-4445309b7e08\") " pod="openstack/dnsmasq-dns-845d6d6f59-vg6jp" Oct 06 10:18:10 crc kubenswrapper[4824]: I1006 10:18:10.899631 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27cc4522-3bd4-41bd-9fb7-624c634a5288-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"27cc4522-3bd4-41bd-9fb7-624c634a5288\") " pod="openstack/nova-scheduler-0" Oct 06 10:18:10 crc kubenswrapper[4824]: I1006 10:18:10.899942 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/627f8bd8-ea33-486d-ac9e-2cc7c0b2ae15-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"627f8bd8-ea33-486d-ac9e-2cc7c0b2ae15\") " pod="openstack/nova-cell1-novncproxy-0" Oct 06 10:18:10 crc kubenswrapper[4824]: I1006 10:18:10.901607 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/27cc4522-3bd4-41bd-9fb7-624c634a5288-config-data\") pod \"nova-scheduler-0\" (UID: \"27cc4522-3bd4-41bd-9fb7-624c634a5288\") " pod="openstack/nova-scheduler-0" Oct 06 10:18:10 crc kubenswrapper[4824]: I1006 10:18:10.903505 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/627f8bd8-ea33-486d-ac9e-2cc7c0b2ae15-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"627f8bd8-ea33-486d-ac9e-2cc7c0b2ae15\") " pod="openstack/nova-cell1-novncproxy-0" Oct 06 10:18:10 crc kubenswrapper[4824]: I1006 10:18:10.914457 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cxp5h\" (UniqueName: \"kubernetes.io/projected/627f8bd8-ea33-486d-ac9e-2cc7c0b2ae15-kube-api-access-cxp5h\") pod \"nova-cell1-novncproxy-0\" (UID: \"627f8bd8-ea33-486d-ac9e-2cc7c0b2ae15\") " pod="openstack/nova-cell1-novncproxy-0" Oct 06 10:18:10 crc kubenswrapper[4824]: I1006 10:18:10.914727 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4wvk8\" (UniqueName: \"kubernetes.io/projected/48fc2f94-426b-4056-88db-4445309b7e08-kube-api-access-4wvk8\") pod \"dnsmasq-dns-845d6d6f59-vg6jp\" (UID: \"48fc2f94-426b-4056-88db-4445309b7e08\") " pod="openstack/dnsmasq-dns-845d6d6f59-vg6jp" Oct 06 10:18:10 crc kubenswrapper[4824]: I1006 10:18:10.915101 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5cbxw\" (UniqueName: \"kubernetes.io/projected/27cc4522-3bd4-41bd-9fb7-624c634a5288-kube-api-access-5cbxw\") pod \"nova-scheduler-0\" (UID: \"27cc4522-3bd4-41bd-9fb7-624c634a5288\") " pod="openstack/nova-scheduler-0" Oct 06 10:18:10 crc kubenswrapper[4824]: I1006 10:18:10.969091 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-845d6d6f59-vg6jp" Oct 06 10:18:11 crc kubenswrapper[4824]: I1006 10:18:11.006304 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 06 10:18:11 crc kubenswrapper[4824]: I1006 10:18:11.009168 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 06 10:18:11 crc kubenswrapper[4824]: I1006 10:18:11.115361 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-zq5td"] Oct 06 10:18:11 crc kubenswrapper[4824]: W1006 10:18:11.135211 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcd95e609_bd54_4234_a449_66cd65e2558a.slice/crio-da7f86acdc6cdf57487d522f158579fb5c3f5ff157072f1ac045c10b28947559 WatchSource:0}: Error finding container da7f86acdc6cdf57487d522f158579fb5c3f5ff157072f1ac045c10b28947559: Status 404 returned error can't find the container with id da7f86acdc6cdf57487d522f158579fb5c3f5ff157072f1ac045c10b28947559 Oct 06 10:18:11 crc kubenswrapper[4824]: I1006 10:18:11.152525 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-cbbc7"] Oct 06 10:18:11 crc kubenswrapper[4824]: I1006 10:18:11.154344 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-cbbc7" Oct 06 10:18:11 crc kubenswrapper[4824]: I1006 10:18:11.168278 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-cbbc7"] Oct 06 10:18:11 crc kubenswrapper[4824]: I1006 10:18:11.170277 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Oct 06 10:18:11 crc kubenswrapper[4824]: I1006 10:18:11.170625 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 06 10:18:11 crc kubenswrapper[4824]: I1006 10:18:11.197229 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/795af81e-370f-4845-8632-acfe7988e3cf-config-data\") pod \"nova-cell1-conductor-db-sync-cbbc7\" (UID: \"795af81e-370f-4845-8632-acfe7988e3cf\") " pod="openstack/nova-cell1-conductor-db-sync-cbbc7" Oct 06 10:18:11 crc kubenswrapper[4824]: I1006 10:18:11.197566 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/795af81e-370f-4845-8632-acfe7988e3cf-scripts\") pod \"nova-cell1-conductor-db-sync-cbbc7\" (UID: \"795af81e-370f-4845-8632-acfe7988e3cf\") " pod="openstack/nova-cell1-conductor-db-sync-cbbc7" Oct 06 10:18:11 crc kubenswrapper[4824]: I1006 10:18:11.197741 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wsrtp\" (UniqueName: \"kubernetes.io/projected/795af81e-370f-4845-8632-acfe7988e3cf-kube-api-access-wsrtp\") pod \"nova-cell1-conductor-db-sync-cbbc7\" (UID: \"795af81e-370f-4845-8632-acfe7988e3cf\") " pod="openstack/nova-cell1-conductor-db-sync-cbbc7" Oct 06 10:18:11 crc kubenswrapper[4824]: I1006 10:18:11.197893 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/795af81e-370f-4845-8632-acfe7988e3cf-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-cbbc7\" (UID: \"795af81e-370f-4845-8632-acfe7988e3cf\") " pod="openstack/nova-cell1-conductor-db-sync-cbbc7" Oct 06 10:18:11 crc kubenswrapper[4824]: I1006 10:18:11.305313 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/795af81e-370f-4845-8632-acfe7988e3cf-scripts\") pod \"nova-cell1-conductor-db-sync-cbbc7\" (UID: \"795af81e-370f-4845-8632-acfe7988e3cf\") " pod="openstack/nova-cell1-conductor-db-sync-cbbc7" Oct 06 10:18:11 crc kubenswrapper[4824]: I1006 10:18:11.305400 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wsrtp\" (UniqueName: \"kubernetes.io/projected/795af81e-370f-4845-8632-acfe7988e3cf-kube-api-access-wsrtp\") pod \"nova-cell1-conductor-db-sync-cbbc7\" (UID: \"795af81e-370f-4845-8632-acfe7988e3cf\") " pod="openstack/nova-cell1-conductor-db-sync-cbbc7" Oct 06 10:18:11 crc kubenswrapper[4824]: I1006 10:18:11.305453 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/795af81e-370f-4845-8632-acfe7988e3cf-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-cbbc7\" (UID: \"795af81e-370f-4845-8632-acfe7988e3cf\") " pod="openstack/nova-cell1-conductor-db-sync-cbbc7" Oct 06 10:18:11 crc kubenswrapper[4824]: I1006 10:18:11.305611 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/795af81e-370f-4845-8632-acfe7988e3cf-config-data\") pod \"nova-cell1-conductor-db-sync-cbbc7\" (UID: \"795af81e-370f-4845-8632-acfe7988e3cf\") " pod="openstack/nova-cell1-conductor-db-sync-cbbc7" Oct 06 10:18:11 crc kubenswrapper[4824]: I1006 10:18:11.310374 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Oct 06 10:18:11 crc kubenswrapper[4824]: I1006 10:18:11.310727 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 06 10:18:11 crc kubenswrapper[4824]: I1006 10:18:11.320825 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/795af81e-370f-4845-8632-acfe7988e3cf-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-cbbc7\" (UID: \"795af81e-370f-4845-8632-acfe7988e3cf\") " pod="openstack/nova-cell1-conductor-db-sync-cbbc7" Oct 06 10:18:11 crc kubenswrapper[4824]: I1006 10:18:11.330462 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/795af81e-370f-4845-8632-acfe7988e3cf-scripts\") pod \"nova-cell1-conductor-db-sync-cbbc7\" (UID: \"795af81e-370f-4845-8632-acfe7988e3cf\") " pod="openstack/nova-cell1-conductor-db-sync-cbbc7" Oct 06 10:18:11 crc kubenswrapper[4824]: I1006 10:18:11.338962 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/795af81e-370f-4845-8632-acfe7988e3cf-config-data\") pod \"nova-cell1-conductor-db-sync-cbbc7\" (UID: \"795af81e-370f-4845-8632-acfe7988e3cf\") " pod="openstack/nova-cell1-conductor-db-sync-cbbc7" Oct 06 10:18:11 crc kubenswrapper[4824]: I1006 10:18:11.340112 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wsrtp\" (UniqueName: \"kubernetes.io/projected/795af81e-370f-4845-8632-acfe7988e3cf-kube-api-access-wsrtp\") pod \"nova-cell1-conductor-db-sync-cbbc7\" (UID: \"795af81e-370f-4845-8632-acfe7988e3cf\") " pod="openstack/nova-cell1-conductor-db-sync-cbbc7" Oct 06 10:18:11 crc kubenswrapper[4824]: I1006 10:18:11.349313 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 06 10:18:11 crc kubenswrapper[4824]: I1006 10:18:11.411246 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 06 10:18:11 crc kubenswrapper[4824]: I1006 10:18:11.433496 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-cbbc7" Oct 06 10:18:11 crc kubenswrapper[4824]: I1006 10:18:11.644420 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-vg6jp"] Oct 06 10:18:11 crc kubenswrapper[4824]: I1006 10:18:11.706075 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-cbbc7"] Oct 06 10:18:11 crc kubenswrapper[4824]: I1006 10:18:11.757102 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 06 10:18:11 crc kubenswrapper[4824]: I1006 10:18:11.771074 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 06 10:18:12 crc kubenswrapper[4824]: I1006 10:18:12.145165 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d56c5ee8-59b8-45d6-9078-29423490982f","Type":"ContainerStarted","Data":"41a9a04e7c022b18eaf8b6513683d463b433430bb44bb3e467f2eebc574f3c89"} Oct 06 10:18:12 crc kubenswrapper[4824]: I1006 10:18:12.152395 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-zq5td" event={"ID":"cd95e609-bd54-4234-a449-66cd65e2558a","Type":"ContainerStarted","Data":"d7a21bc36e001b5d47e59fe8dd2712c72c5714c4632964f952a86eead6ad34bc"} Oct 06 10:18:12 crc kubenswrapper[4824]: I1006 10:18:12.152486 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-zq5td" event={"ID":"cd95e609-bd54-4234-a449-66cd65e2558a","Type":"ContainerStarted","Data":"da7f86acdc6cdf57487d522f158579fb5c3f5ff157072f1ac045c10b28947559"} Oct 06 10:18:12 crc kubenswrapper[4824]: I1006 10:18:12.160189 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"27cc4522-3bd4-41bd-9fb7-624c634a5288","Type":"ContainerStarted","Data":"13d5cd70391e4730d3bf4f7643e877628d5fd4fa929172f4c435fe8c9b55c929"} Oct 06 10:18:12 crc kubenswrapper[4824]: I1006 10:18:12.174707 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-zq5td" podStartSLOduration=3.174687658 podStartE2EDuration="3.174687658s" podCreationTimestamp="2025-10-06 10:18:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:18:12.167210642 +0000 UTC m=+1201.531633513" watchObservedRunningTime="2025-10-06 10:18:12.174687658 +0000 UTC m=+1201.539110519" Oct 06 10:18:12 crc kubenswrapper[4824]: I1006 10:18:12.175935 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-cbbc7" event={"ID":"795af81e-370f-4845-8632-acfe7988e3cf","Type":"ContainerStarted","Data":"7ebd18e643bfc8279abcc47f6aeb5952741001d0ec2376eecfaf1520463b43e1"} Oct 06 10:18:12 crc kubenswrapper[4824]: I1006 10:18:12.176005 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-cbbc7" event={"ID":"795af81e-370f-4845-8632-acfe7988e3cf","Type":"ContainerStarted","Data":"969203d24cb1cd8e340b8662626d1b94b8a1955de432722bd801bc9599b8eab5"} Oct 06 10:18:12 crc kubenswrapper[4824]: I1006 10:18:12.184129 4824 generic.go:334] "Generic (PLEG): container finished" podID="48fc2f94-426b-4056-88db-4445309b7e08" containerID="fd2ae08b0d41823968cf87f77265c104dcc97a748c00a0eeaf761cf405f8abe5" exitCode=0 Oct 06 10:18:12 crc kubenswrapper[4824]: I1006 10:18:12.184237 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845d6d6f59-vg6jp" event={"ID":"48fc2f94-426b-4056-88db-4445309b7e08","Type":"ContainerDied","Data":"fd2ae08b0d41823968cf87f77265c104dcc97a748c00a0eeaf761cf405f8abe5"} Oct 06 10:18:12 crc kubenswrapper[4824]: I1006 10:18:12.184267 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845d6d6f59-vg6jp" event={"ID":"48fc2f94-426b-4056-88db-4445309b7e08","Type":"ContainerStarted","Data":"01a56efed5bc48d24af8943161b071a81b6bab26718103cafb5fa94368de1122"} Oct 06 10:18:12 crc kubenswrapper[4824]: I1006 10:18:12.199557 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"cf250d98-9985-42e1-b5d8-d3cf5e660869","Type":"ContainerStarted","Data":"dc34c0b0a04dedef55a5975099258a7f56afe072ea3747b1a8609eae95978460"} Oct 06 10:18:12 crc kubenswrapper[4824]: I1006 10:18:12.201748 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-cbbc7" podStartSLOduration=1.201724846 podStartE2EDuration="1.201724846s" podCreationTimestamp="2025-10-06 10:18:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:18:12.192818746 +0000 UTC m=+1201.557241607" watchObservedRunningTime="2025-10-06 10:18:12.201724846 +0000 UTC m=+1201.566147697" Oct 06 10:18:12 crc kubenswrapper[4824]: I1006 10:18:12.215527 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"627f8bd8-ea33-486d-ac9e-2cc7c0b2ae15","Type":"ContainerStarted","Data":"543492853f249ef04194d5b4be5598b2417659e4845101dc492848e8fc0ae67c"} Oct 06 10:18:13 crc kubenswrapper[4824]: I1006 10:18:13.240255 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845d6d6f59-vg6jp" event={"ID":"48fc2f94-426b-4056-88db-4445309b7e08","Type":"ContainerStarted","Data":"311bea70f6af6e61a096848b0d42449d4d2c5caf0e2671a1a2dcef94afab3072"} Oct 06 10:18:13 crc kubenswrapper[4824]: I1006 10:18:13.241408 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-845d6d6f59-vg6jp" Oct 06 10:18:13 crc kubenswrapper[4824]: I1006 10:18:13.270483 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-845d6d6f59-vg6jp" podStartSLOduration=3.270458493 podStartE2EDuration="3.270458493s" podCreationTimestamp="2025-10-06 10:18:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:18:13.260055256 +0000 UTC m=+1202.624478107" watchObservedRunningTime="2025-10-06 10:18:13.270458493 +0000 UTC m=+1202.634881354" Oct 06 10:18:13 crc kubenswrapper[4824]: I1006 10:18:13.916945 4824 patch_prober.go:28] interesting pod/machine-config-daemon-khgzw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 10:18:13 crc kubenswrapper[4824]: I1006 10:18:13.917363 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 10:18:14 crc kubenswrapper[4824]: I1006 10:18:14.093135 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 06 10:18:14 crc kubenswrapper[4824]: I1006 10:18:14.104264 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 06 10:18:16 crc kubenswrapper[4824]: I1006 10:18:16.278769 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"cf250d98-9985-42e1-b5d8-d3cf5e660869","Type":"ContainerStarted","Data":"6174287384be068eff6a0dd697225c001efdcb58cc4b85fc52f00814454a71db"} Oct 06 10:18:16 crc kubenswrapper[4824]: I1006 10:18:16.279736 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"cf250d98-9985-42e1-b5d8-d3cf5e660869","Type":"ContainerStarted","Data":"6d78bfd67d22f7068df792acf9212d183e0c32ed376654c72921f1abb1301adc"} Oct 06 10:18:16 crc kubenswrapper[4824]: I1006 10:18:16.279064 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="cf250d98-9985-42e1-b5d8-d3cf5e660869" containerName="nova-metadata-metadata" containerID="cri-o://6174287384be068eff6a0dd697225c001efdcb58cc4b85fc52f00814454a71db" gracePeriod=30 Oct 06 10:18:16 crc kubenswrapper[4824]: I1006 10:18:16.278951 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="cf250d98-9985-42e1-b5d8-d3cf5e660869" containerName="nova-metadata-log" containerID="cri-o://6d78bfd67d22f7068df792acf9212d183e0c32ed376654c72921f1abb1301adc" gracePeriod=30 Oct 06 10:18:16 crc kubenswrapper[4824]: I1006 10:18:16.282256 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"627f8bd8-ea33-486d-ac9e-2cc7c0b2ae15","Type":"ContainerStarted","Data":"64652f98eb9c1402f8d72a1b0d7d5a5e717088663152bb7f3cf9cce181f45f05"} Oct 06 10:18:16 crc kubenswrapper[4824]: I1006 10:18:16.282352 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="627f8bd8-ea33-486d-ac9e-2cc7c0b2ae15" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://64652f98eb9c1402f8d72a1b0d7d5a5e717088663152bb7f3cf9cce181f45f05" gracePeriod=30 Oct 06 10:18:16 crc kubenswrapper[4824]: I1006 10:18:16.289218 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d56c5ee8-59b8-45d6-9078-29423490982f","Type":"ContainerStarted","Data":"488627b76866421d5a277cabe307c6c6ab3d23a1bca9671b420a9ca1953b960e"} Oct 06 10:18:16 crc kubenswrapper[4824]: I1006 10:18:16.289289 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d56c5ee8-59b8-45d6-9078-29423490982f","Type":"ContainerStarted","Data":"11e6da3b70f0831659aade67abc845e1bb7b3a14433d49fd0924bdbb9518e1ad"} Oct 06 10:18:16 crc kubenswrapper[4824]: I1006 10:18:16.294278 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"27cc4522-3bd4-41bd-9fb7-624c634a5288","Type":"ContainerStarted","Data":"5aa36cdff7d909fd98c3a8a65be66bbf3f6a6b660ad12183d0da7f5a94b443b3"} Oct 06 10:18:16 crc kubenswrapper[4824]: I1006 10:18:16.311875 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.344680041 podStartE2EDuration="6.311846356s" podCreationTimestamp="2025-10-06 10:18:10 +0000 UTC" firstStartedPulling="2025-10-06 10:18:11.37882617 +0000 UTC m=+1200.743249031" lastFinishedPulling="2025-10-06 10:18:15.345992485 +0000 UTC m=+1204.710415346" observedRunningTime="2025-10-06 10:18:16.304805012 +0000 UTC m=+1205.669227883" watchObservedRunningTime="2025-10-06 10:18:16.311846356 +0000 UTC m=+1205.676269217" Oct 06 10:18:16 crc kubenswrapper[4824]: I1006 10:18:16.332152 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.746133874 podStartE2EDuration="6.332122818s" podCreationTimestamp="2025-10-06 10:18:10 +0000 UTC" firstStartedPulling="2025-10-06 10:18:11.760184795 +0000 UTC m=+1201.124607656" lastFinishedPulling="2025-10-06 10:18:15.346173739 +0000 UTC m=+1204.710596600" observedRunningTime="2025-10-06 10:18:16.326208902 +0000 UTC m=+1205.690631773" watchObservedRunningTime="2025-10-06 10:18:16.332122818 +0000 UTC m=+1205.696545679" Oct 06 10:18:16 crc kubenswrapper[4824]: I1006 10:18:16.350998 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.421334109 podStartE2EDuration="6.350955744s" podCreationTimestamp="2025-10-06 10:18:10 +0000 UTC" firstStartedPulling="2025-10-06 10:18:11.414837202 +0000 UTC m=+1200.779260063" lastFinishedPulling="2025-10-06 10:18:15.344458837 +0000 UTC m=+1204.708881698" observedRunningTime="2025-10-06 10:18:16.344094954 +0000 UTC m=+1205.708517815" watchObservedRunningTime="2025-10-06 10:18:16.350955744 +0000 UTC m=+1205.715378605" Oct 06 10:18:16 crc kubenswrapper[4824]: I1006 10:18:16.372846 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.7800604030000002 podStartE2EDuration="6.372826915s" podCreationTimestamp="2025-10-06 10:18:10 +0000 UTC" firstStartedPulling="2025-10-06 10:18:11.756491303 +0000 UTC m=+1201.120914164" lastFinishedPulling="2025-10-06 10:18:15.349257815 +0000 UTC m=+1204.713680676" observedRunningTime="2025-10-06 10:18:16.364241003 +0000 UTC m=+1205.728663864" watchObservedRunningTime="2025-10-06 10:18:16.372826915 +0000 UTC m=+1205.737249776" Oct 06 10:18:16 crc kubenswrapper[4824]: I1006 10:18:16.941686 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 06 10:18:17 crc kubenswrapper[4824]: I1006 10:18:17.072302 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cf250d98-9985-42e1-b5d8-d3cf5e660869-logs\") pod \"cf250d98-9985-42e1-b5d8-d3cf5e660869\" (UID: \"cf250d98-9985-42e1-b5d8-d3cf5e660869\") " Oct 06 10:18:17 crc kubenswrapper[4824]: I1006 10:18:17.072409 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf250d98-9985-42e1-b5d8-d3cf5e660869-combined-ca-bundle\") pod \"cf250d98-9985-42e1-b5d8-d3cf5e660869\" (UID: \"cf250d98-9985-42e1-b5d8-d3cf5e660869\") " Oct 06 10:18:17 crc kubenswrapper[4824]: I1006 10:18:17.072442 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf250d98-9985-42e1-b5d8-d3cf5e660869-config-data\") pod \"cf250d98-9985-42e1-b5d8-d3cf5e660869\" (UID: \"cf250d98-9985-42e1-b5d8-d3cf5e660869\") " Oct 06 10:18:17 crc kubenswrapper[4824]: I1006 10:18:17.072910 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mxd2k\" (UniqueName: \"kubernetes.io/projected/cf250d98-9985-42e1-b5d8-d3cf5e660869-kube-api-access-mxd2k\") pod \"cf250d98-9985-42e1-b5d8-d3cf5e660869\" (UID: \"cf250d98-9985-42e1-b5d8-d3cf5e660869\") " Oct 06 10:18:17 crc kubenswrapper[4824]: I1006 10:18:17.074910 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cf250d98-9985-42e1-b5d8-d3cf5e660869-logs" (OuterVolumeSpecName: "logs") pod "cf250d98-9985-42e1-b5d8-d3cf5e660869" (UID: "cf250d98-9985-42e1-b5d8-d3cf5e660869"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:18:17 crc kubenswrapper[4824]: I1006 10:18:17.084490 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cf250d98-9985-42e1-b5d8-d3cf5e660869-kube-api-access-mxd2k" (OuterVolumeSpecName: "kube-api-access-mxd2k") pod "cf250d98-9985-42e1-b5d8-d3cf5e660869" (UID: "cf250d98-9985-42e1-b5d8-d3cf5e660869"). InnerVolumeSpecName "kube-api-access-mxd2k". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:18:17 crc kubenswrapper[4824]: I1006 10:18:17.115278 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cf250d98-9985-42e1-b5d8-d3cf5e660869-config-data" (OuterVolumeSpecName: "config-data") pod "cf250d98-9985-42e1-b5d8-d3cf5e660869" (UID: "cf250d98-9985-42e1-b5d8-d3cf5e660869"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:18:17 crc kubenswrapper[4824]: I1006 10:18:17.120557 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cf250d98-9985-42e1-b5d8-d3cf5e660869-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cf250d98-9985-42e1-b5d8-d3cf5e660869" (UID: "cf250d98-9985-42e1-b5d8-d3cf5e660869"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:18:17 crc kubenswrapper[4824]: I1006 10:18:17.175838 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mxd2k\" (UniqueName: \"kubernetes.io/projected/cf250d98-9985-42e1-b5d8-d3cf5e660869-kube-api-access-mxd2k\") on node \"crc\" DevicePath \"\"" Oct 06 10:18:17 crc kubenswrapper[4824]: I1006 10:18:17.176342 4824 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cf250d98-9985-42e1-b5d8-d3cf5e660869-logs\") on node \"crc\" DevicePath \"\"" Oct 06 10:18:17 crc kubenswrapper[4824]: I1006 10:18:17.176443 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf250d98-9985-42e1-b5d8-d3cf5e660869-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 10:18:17 crc kubenswrapper[4824]: I1006 10:18:17.176539 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf250d98-9985-42e1-b5d8-d3cf5e660869-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 10:18:17 crc kubenswrapper[4824]: I1006 10:18:17.307547 4824 generic.go:334] "Generic (PLEG): container finished" podID="cf250d98-9985-42e1-b5d8-d3cf5e660869" containerID="6174287384be068eff6a0dd697225c001efdcb58cc4b85fc52f00814454a71db" exitCode=0 Oct 06 10:18:17 crc kubenswrapper[4824]: I1006 10:18:17.307606 4824 generic.go:334] "Generic (PLEG): container finished" podID="cf250d98-9985-42e1-b5d8-d3cf5e660869" containerID="6d78bfd67d22f7068df792acf9212d183e0c32ed376654c72921f1abb1301adc" exitCode=143 Oct 06 10:18:17 crc kubenswrapper[4824]: I1006 10:18:17.307614 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"cf250d98-9985-42e1-b5d8-d3cf5e660869","Type":"ContainerDied","Data":"6174287384be068eff6a0dd697225c001efdcb58cc4b85fc52f00814454a71db"} Oct 06 10:18:17 crc kubenswrapper[4824]: I1006 10:18:17.307662 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 06 10:18:17 crc kubenswrapper[4824]: I1006 10:18:17.307703 4824 scope.go:117] "RemoveContainer" containerID="6174287384be068eff6a0dd697225c001efdcb58cc4b85fc52f00814454a71db" Oct 06 10:18:17 crc kubenswrapper[4824]: I1006 10:18:17.307684 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"cf250d98-9985-42e1-b5d8-d3cf5e660869","Type":"ContainerDied","Data":"6d78bfd67d22f7068df792acf9212d183e0c32ed376654c72921f1abb1301adc"} Oct 06 10:18:17 crc kubenswrapper[4824]: I1006 10:18:17.307892 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"cf250d98-9985-42e1-b5d8-d3cf5e660869","Type":"ContainerDied","Data":"dc34c0b0a04dedef55a5975099258a7f56afe072ea3747b1a8609eae95978460"} Oct 06 10:18:17 crc kubenswrapper[4824]: I1006 10:18:17.357066 4824 scope.go:117] "RemoveContainer" containerID="6d78bfd67d22f7068df792acf9212d183e0c32ed376654c72921f1abb1301adc" Oct 06 10:18:17 crc kubenswrapper[4824]: I1006 10:18:17.419486 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 06 10:18:17 crc kubenswrapper[4824]: I1006 10:18:17.459081 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 06 10:18:17 crc kubenswrapper[4824]: I1006 10:18:17.466835 4824 scope.go:117] "RemoveContainer" containerID="6174287384be068eff6a0dd697225c001efdcb58cc4b85fc52f00814454a71db" Oct 06 10:18:17 crc kubenswrapper[4824]: E1006 10:18:17.467400 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6174287384be068eff6a0dd697225c001efdcb58cc4b85fc52f00814454a71db\": container with ID starting with 6174287384be068eff6a0dd697225c001efdcb58cc4b85fc52f00814454a71db not found: ID does not exist" containerID="6174287384be068eff6a0dd697225c001efdcb58cc4b85fc52f00814454a71db" Oct 06 10:18:17 crc kubenswrapper[4824]: I1006 10:18:17.467566 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6174287384be068eff6a0dd697225c001efdcb58cc4b85fc52f00814454a71db"} err="failed to get container status \"6174287384be068eff6a0dd697225c001efdcb58cc4b85fc52f00814454a71db\": rpc error: code = NotFound desc = could not find container \"6174287384be068eff6a0dd697225c001efdcb58cc4b85fc52f00814454a71db\": container with ID starting with 6174287384be068eff6a0dd697225c001efdcb58cc4b85fc52f00814454a71db not found: ID does not exist" Oct 06 10:18:17 crc kubenswrapper[4824]: I1006 10:18:17.467682 4824 scope.go:117] "RemoveContainer" containerID="6d78bfd67d22f7068df792acf9212d183e0c32ed376654c72921f1abb1301adc" Oct 06 10:18:17 crc kubenswrapper[4824]: E1006 10:18:17.468072 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6d78bfd67d22f7068df792acf9212d183e0c32ed376654c72921f1abb1301adc\": container with ID starting with 6d78bfd67d22f7068df792acf9212d183e0c32ed376654c72921f1abb1301adc not found: ID does not exist" containerID="6d78bfd67d22f7068df792acf9212d183e0c32ed376654c72921f1abb1301adc" Oct 06 10:18:17 crc kubenswrapper[4824]: I1006 10:18:17.468099 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6d78bfd67d22f7068df792acf9212d183e0c32ed376654c72921f1abb1301adc"} err="failed to get container status \"6d78bfd67d22f7068df792acf9212d183e0c32ed376654c72921f1abb1301adc\": rpc error: code = NotFound desc = could not find container \"6d78bfd67d22f7068df792acf9212d183e0c32ed376654c72921f1abb1301adc\": container with ID starting with 6d78bfd67d22f7068df792acf9212d183e0c32ed376654c72921f1abb1301adc not found: ID does not exist" Oct 06 10:18:17 crc kubenswrapper[4824]: I1006 10:18:17.468117 4824 scope.go:117] "RemoveContainer" containerID="6174287384be068eff6a0dd697225c001efdcb58cc4b85fc52f00814454a71db" Oct 06 10:18:17 crc kubenswrapper[4824]: I1006 10:18:17.468623 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6174287384be068eff6a0dd697225c001efdcb58cc4b85fc52f00814454a71db"} err="failed to get container status \"6174287384be068eff6a0dd697225c001efdcb58cc4b85fc52f00814454a71db\": rpc error: code = NotFound desc = could not find container \"6174287384be068eff6a0dd697225c001efdcb58cc4b85fc52f00814454a71db\": container with ID starting with 6174287384be068eff6a0dd697225c001efdcb58cc4b85fc52f00814454a71db not found: ID does not exist" Oct 06 10:18:17 crc kubenswrapper[4824]: I1006 10:18:17.468646 4824 scope.go:117] "RemoveContainer" containerID="6d78bfd67d22f7068df792acf9212d183e0c32ed376654c72921f1abb1301adc" Oct 06 10:18:17 crc kubenswrapper[4824]: I1006 10:18:17.468919 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6d78bfd67d22f7068df792acf9212d183e0c32ed376654c72921f1abb1301adc"} err="failed to get container status \"6d78bfd67d22f7068df792acf9212d183e0c32ed376654c72921f1abb1301adc\": rpc error: code = NotFound desc = could not find container \"6d78bfd67d22f7068df792acf9212d183e0c32ed376654c72921f1abb1301adc\": container with ID starting with 6d78bfd67d22f7068df792acf9212d183e0c32ed376654c72921f1abb1301adc not found: ID does not exist" Oct 06 10:18:17 crc kubenswrapper[4824]: I1006 10:18:17.470180 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 06 10:18:17 crc kubenswrapper[4824]: E1006 10:18:17.470791 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf250d98-9985-42e1-b5d8-d3cf5e660869" containerName="nova-metadata-log" Oct 06 10:18:17 crc kubenswrapper[4824]: I1006 10:18:17.470808 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf250d98-9985-42e1-b5d8-d3cf5e660869" containerName="nova-metadata-log" Oct 06 10:18:17 crc kubenswrapper[4824]: E1006 10:18:17.470848 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf250d98-9985-42e1-b5d8-d3cf5e660869" containerName="nova-metadata-metadata" Oct 06 10:18:17 crc kubenswrapper[4824]: I1006 10:18:17.470856 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf250d98-9985-42e1-b5d8-d3cf5e660869" containerName="nova-metadata-metadata" Oct 06 10:18:17 crc kubenswrapper[4824]: I1006 10:18:17.471115 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="cf250d98-9985-42e1-b5d8-d3cf5e660869" containerName="nova-metadata-log" Oct 06 10:18:17 crc kubenswrapper[4824]: I1006 10:18:17.471139 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="cf250d98-9985-42e1-b5d8-d3cf5e660869" containerName="nova-metadata-metadata" Oct 06 10:18:17 crc kubenswrapper[4824]: I1006 10:18:17.472407 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 06 10:18:17 crc kubenswrapper[4824]: I1006 10:18:17.475372 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 06 10:18:17 crc kubenswrapper[4824]: I1006 10:18:17.475844 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 06 10:18:17 crc kubenswrapper[4824]: I1006 10:18:17.482119 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 06 10:18:17 crc kubenswrapper[4824]: I1006 10:18:17.603740 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d722707-3130-4de6-ac0c-40d63a0fdff3-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"7d722707-3130-4de6-ac0c-40d63a0fdff3\") " pod="openstack/nova-metadata-0" Oct 06 10:18:17 crc kubenswrapper[4824]: I1006 10:18:17.603864 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rzfdn\" (UniqueName: \"kubernetes.io/projected/7d722707-3130-4de6-ac0c-40d63a0fdff3-kube-api-access-rzfdn\") pod \"nova-metadata-0\" (UID: \"7d722707-3130-4de6-ac0c-40d63a0fdff3\") " pod="openstack/nova-metadata-0" Oct 06 10:18:17 crc kubenswrapper[4824]: I1006 10:18:17.603917 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7d722707-3130-4de6-ac0c-40d63a0fdff3-logs\") pod \"nova-metadata-0\" (UID: \"7d722707-3130-4de6-ac0c-40d63a0fdff3\") " pod="openstack/nova-metadata-0" Oct 06 10:18:17 crc kubenswrapper[4824]: I1006 10:18:17.603938 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d722707-3130-4de6-ac0c-40d63a0fdff3-config-data\") pod \"nova-metadata-0\" (UID: \"7d722707-3130-4de6-ac0c-40d63a0fdff3\") " pod="openstack/nova-metadata-0" Oct 06 10:18:17 crc kubenswrapper[4824]: I1006 10:18:17.604016 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/7d722707-3130-4de6-ac0c-40d63a0fdff3-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"7d722707-3130-4de6-ac0c-40d63a0fdff3\") " pod="openstack/nova-metadata-0" Oct 06 10:18:17 crc kubenswrapper[4824]: I1006 10:18:17.706134 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7d722707-3130-4de6-ac0c-40d63a0fdff3-logs\") pod \"nova-metadata-0\" (UID: \"7d722707-3130-4de6-ac0c-40d63a0fdff3\") " pod="openstack/nova-metadata-0" Oct 06 10:18:17 crc kubenswrapper[4824]: I1006 10:18:17.706202 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d722707-3130-4de6-ac0c-40d63a0fdff3-config-data\") pod \"nova-metadata-0\" (UID: \"7d722707-3130-4de6-ac0c-40d63a0fdff3\") " pod="openstack/nova-metadata-0" Oct 06 10:18:17 crc kubenswrapper[4824]: I1006 10:18:17.706302 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/7d722707-3130-4de6-ac0c-40d63a0fdff3-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"7d722707-3130-4de6-ac0c-40d63a0fdff3\") " pod="openstack/nova-metadata-0" Oct 06 10:18:17 crc kubenswrapper[4824]: I1006 10:18:17.706376 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d722707-3130-4de6-ac0c-40d63a0fdff3-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"7d722707-3130-4de6-ac0c-40d63a0fdff3\") " pod="openstack/nova-metadata-0" Oct 06 10:18:17 crc kubenswrapper[4824]: I1006 10:18:17.706446 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rzfdn\" (UniqueName: \"kubernetes.io/projected/7d722707-3130-4de6-ac0c-40d63a0fdff3-kube-api-access-rzfdn\") pod \"nova-metadata-0\" (UID: \"7d722707-3130-4de6-ac0c-40d63a0fdff3\") " pod="openstack/nova-metadata-0" Oct 06 10:18:17 crc kubenswrapper[4824]: I1006 10:18:17.707852 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7d722707-3130-4de6-ac0c-40d63a0fdff3-logs\") pod \"nova-metadata-0\" (UID: \"7d722707-3130-4de6-ac0c-40d63a0fdff3\") " pod="openstack/nova-metadata-0" Oct 06 10:18:17 crc kubenswrapper[4824]: I1006 10:18:17.713374 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d722707-3130-4de6-ac0c-40d63a0fdff3-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"7d722707-3130-4de6-ac0c-40d63a0fdff3\") " pod="openstack/nova-metadata-0" Oct 06 10:18:17 crc kubenswrapper[4824]: I1006 10:18:17.714942 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d722707-3130-4de6-ac0c-40d63a0fdff3-config-data\") pod \"nova-metadata-0\" (UID: \"7d722707-3130-4de6-ac0c-40d63a0fdff3\") " pod="openstack/nova-metadata-0" Oct 06 10:18:17 crc kubenswrapper[4824]: I1006 10:18:17.731542 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/7d722707-3130-4de6-ac0c-40d63a0fdff3-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"7d722707-3130-4de6-ac0c-40d63a0fdff3\") " pod="openstack/nova-metadata-0" Oct 06 10:18:17 crc kubenswrapper[4824]: I1006 10:18:17.734429 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rzfdn\" (UniqueName: \"kubernetes.io/projected/7d722707-3130-4de6-ac0c-40d63a0fdff3-kube-api-access-rzfdn\") pod \"nova-metadata-0\" (UID: \"7d722707-3130-4de6-ac0c-40d63a0fdff3\") " pod="openstack/nova-metadata-0" Oct 06 10:18:17 crc kubenswrapper[4824]: I1006 10:18:17.790220 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 06 10:18:18 crc kubenswrapper[4824]: I1006 10:18:18.347886 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 06 10:18:18 crc kubenswrapper[4824]: W1006 10:18:18.348566 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7d722707_3130_4de6_ac0c_40d63a0fdff3.slice/crio-4ee59599d2fbe9519798888d0c521f76b4f4642bb1fa3dba8614fffdbf28fc1c WatchSource:0}: Error finding container 4ee59599d2fbe9519798888d0c521f76b4f4642bb1fa3dba8614fffdbf28fc1c: Status 404 returned error can't find the container with id 4ee59599d2fbe9519798888d0c521f76b4f4642bb1fa3dba8614fffdbf28fc1c Oct 06 10:18:19 crc kubenswrapper[4824]: I1006 10:18:19.288811 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cf250d98-9985-42e1-b5d8-d3cf5e660869" path="/var/lib/kubelet/pods/cf250d98-9985-42e1-b5d8-d3cf5e660869/volumes" Oct 06 10:18:19 crc kubenswrapper[4824]: I1006 10:18:19.341771 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"7d722707-3130-4de6-ac0c-40d63a0fdff3","Type":"ContainerStarted","Data":"7041e6ab4a92c1e02ee6f0ec262c66d7249ed1296d27f2ba4bedb3b75d530003"} Oct 06 10:18:19 crc kubenswrapper[4824]: I1006 10:18:19.341844 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"7d722707-3130-4de6-ac0c-40d63a0fdff3","Type":"ContainerStarted","Data":"2578847baa70b643e2068c5f76c873450f4ab5b34ac01e713b1b85b4be7f1ecb"} Oct 06 10:18:19 crc kubenswrapper[4824]: I1006 10:18:19.341864 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"7d722707-3130-4de6-ac0c-40d63a0fdff3","Type":"ContainerStarted","Data":"4ee59599d2fbe9519798888d0c521f76b4f4642bb1fa3dba8614fffdbf28fc1c"} Oct 06 10:18:19 crc kubenswrapper[4824]: I1006 10:18:19.385334 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.385312944 podStartE2EDuration="2.385312944s" podCreationTimestamp="2025-10-06 10:18:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:18:19.372150998 +0000 UTC m=+1208.736573879" watchObservedRunningTime="2025-10-06 10:18:19.385312944 +0000 UTC m=+1208.749735805" Oct 06 10:18:20 crc kubenswrapper[4824]: I1006 10:18:20.669806 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 06 10:18:20 crc kubenswrapper[4824]: I1006 10:18:20.671560 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 06 10:18:20 crc kubenswrapper[4824]: I1006 10:18:20.971179 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-845d6d6f59-vg6jp" Oct 06 10:18:21 crc kubenswrapper[4824]: I1006 10:18:21.007104 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 06 10:18:21 crc kubenswrapper[4824]: I1006 10:18:21.007637 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 06 10:18:21 crc kubenswrapper[4824]: I1006 10:18:21.009473 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Oct 06 10:18:21 crc kubenswrapper[4824]: I1006 10:18:21.070650 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 06 10:18:21 crc kubenswrapper[4824]: I1006 10:18:21.117815 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-k2mqf"] Oct 06 10:18:21 crc kubenswrapper[4824]: I1006 10:18:21.118315 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5784cf869f-k2mqf" podUID="8de9ec83-5bc3-47a8-9a63-84acb7d6b774" containerName="dnsmasq-dns" containerID="cri-o://dbc55ac9aa7e345cc7e9f4cc96d3aeba3693c9c1e166bdf53a0fa3d1eb241ab1" gracePeriod=10 Oct 06 10:18:21 crc kubenswrapper[4824]: I1006 10:18:21.379251 4824 generic.go:334] "Generic (PLEG): container finished" podID="795af81e-370f-4845-8632-acfe7988e3cf" containerID="7ebd18e643bfc8279abcc47f6aeb5952741001d0ec2376eecfaf1520463b43e1" exitCode=0 Oct 06 10:18:21 crc kubenswrapper[4824]: I1006 10:18:21.379315 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-cbbc7" event={"ID":"795af81e-370f-4845-8632-acfe7988e3cf","Type":"ContainerDied","Data":"7ebd18e643bfc8279abcc47f6aeb5952741001d0ec2376eecfaf1520463b43e1"} Oct 06 10:18:21 crc kubenswrapper[4824]: I1006 10:18:21.382544 4824 generic.go:334] "Generic (PLEG): container finished" podID="8de9ec83-5bc3-47a8-9a63-84acb7d6b774" containerID="dbc55ac9aa7e345cc7e9f4cc96d3aeba3693c9c1e166bdf53a0fa3d1eb241ab1" exitCode=0 Oct 06 10:18:21 crc kubenswrapper[4824]: I1006 10:18:21.382864 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-k2mqf" event={"ID":"8de9ec83-5bc3-47a8-9a63-84acb7d6b774","Type":"ContainerDied","Data":"dbc55ac9aa7e345cc7e9f4cc96d3aeba3693c9c1e166bdf53a0fa3d1eb241ab1"} Oct 06 10:18:21 crc kubenswrapper[4824]: I1006 10:18:21.430690 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 06 10:18:21 crc kubenswrapper[4824]: I1006 10:18:21.752172 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="d56c5ee8-59b8-45d6-9078-29423490982f" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.189:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 06 10:18:21 crc kubenswrapper[4824]: I1006 10:18:21.752225 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="d56c5ee8-59b8-45d6-9078-29423490982f" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.189:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 06 10:18:21 crc kubenswrapper[4824]: I1006 10:18:21.767169 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5784cf869f-k2mqf" Oct 06 10:18:21 crc kubenswrapper[4824]: I1006 10:18:21.846578 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8de9ec83-5bc3-47a8-9a63-84acb7d6b774-ovsdbserver-nb\") pod \"8de9ec83-5bc3-47a8-9a63-84acb7d6b774\" (UID: \"8de9ec83-5bc3-47a8-9a63-84acb7d6b774\") " Oct 06 10:18:21 crc kubenswrapper[4824]: I1006 10:18:21.846678 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8de9ec83-5bc3-47a8-9a63-84acb7d6b774-ovsdbserver-sb\") pod \"8de9ec83-5bc3-47a8-9a63-84acb7d6b774\" (UID: \"8de9ec83-5bc3-47a8-9a63-84acb7d6b774\") " Oct 06 10:18:21 crc kubenswrapper[4824]: I1006 10:18:21.846731 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wbdxh\" (UniqueName: \"kubernetes.io/projected/8de9ec83-5bc3-47a8-9a63-84acb7d6b774-kube-api-access-wbdxh\") pod \"8de9ec83-5bc3-47a8-9a63-84acb7d6b774\" (UID: \"8de9ec83-5bc3-47a8-9a63-84acb7d6b774\") " Oct 06 10:18:21 crc kubenswrapper[4824]: I1006 10:18:21.846800 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8de9ec83-5bc3-47a8-9a63-84acb7d6b774-config\") pod \"8de9ec83-5bc3-47a8-9a63-84acb7d6b774\" (UID: \"8de9ec83-5bc3-47a8-9a63-84acb7d6b774\") " Oct 06 10:18:21 crc kubenswrapper[4824]: I1006 10:18:21.846852 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8de9ec83-5bc3-47a8-9a63-84acb7d6b774-dns-svc\") pod \"8de9ec83-5bc3-47a8-9a63-84acb7d6b774\" (UID: \"8de9ec83-5bc3-47a8-9a63-84acb7d6b774\") " Oct 06 10:18:21 crc kubenswrapper[4824]: I1006 10:18:21.846896 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8de9ec83-5bc3-47a8-9a63-84acb7d6b774-dns-swift-storage-0\") pod \"8de9ec83-5bc3-47a8-9a63-84acb7d6b774\" (UID: \"8de9ec83-5bc3-47a8-9a63-84acb7d6b774\") " Oct 06 10:18:21 crc kubenswrapper[4824]: I1006 10:18:21.853876 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8de9ec83-5bc3-47a8-9a63-84acb7d6b774-kube-api-access-wbdxh" (OuterVolumeSpecName: "kube-api-access-wbdxh") pod "8de9ec83-5bc3-47a8-9a63-84acb7d6b774" (UID: "8de9ec83-5bc3-47a8-9a63-84acb7d6b774"). InnerVolumeSpecName "kube-api-access-wbdxh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:18:21 crc kubenswrapper[4824]: I1006 10:18:21.909845 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8de9ec83-5bc3-47a8-9a63-84acb7d6b774-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "8de9ec83-5bc3-47a8-9a63-84acb7d6b774" (UID: "8de9ec83-5bc3-47a8-9a63-84acb7d6b774"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:18:21 crc kubenswrapper[4824]: I1006 10:18:21.922715 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8de9ec83-5bc3-47a8-9a63-84acb7d6b774-config" (OuterVolumeSpecName: "config") pod "8de9ec83-5bc3-47a8-9a63-84acb7d6b774" (UID: "8de9ec83-5bc3-47a8-9a63-84acb7d6b774"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:18:21 crc kubenswrapper[4824]: I1006 10:18:21.923633 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8de9ec83-5bc3-47a8-9a63-84acb7d6b774-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "8de9ec83-5bc3-47a8-9a63-84acb7d6b774" (UID: "8de9ec83-5bc3-47a8-9a63-84acb7d6b774"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:18:21 crc kubenswrapper[4824]: I1006 10:18:21.932094 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8de9ec83-5bc3-47a8-9a63-84acb7d6b774-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "8de9ec83-5bc3-47a8-9a63-84acb7d6b774" (UID: "8de9ec83-5bc3-47a8-9a63-84acb7d6b774"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:18:21 crc kubenswrapper[4824]: I1006 10:18:21.933269 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8de9ec83-5bc3-47a8-9a63-84acb7d6b774-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "8de9ec83-5bc3-47a8-9a63-84acb7d6b774" (UID: "8de9ec83-5bc3-47a8-9a63-84acb7d6b774"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:18:21 crc kubenswrapper[4824]: I1006 10:18:21.948602 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8de9ec83-5bc3-47a8-9a63-84acb7d6b774-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 06 10:18:21 crc kubenswrapper[4824]: I1006 10:18:21.948631 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8de9ec83-5bc3-47a8-9a63-84acb7d6b774-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 06 10:18:21 crc kubenswrapper[4824]: I1006 10:18:21.948644 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wbdxh\" (UniqueName: \"kubernetes.io/projected/8de9ec83-5bc3-47a8-9a63-84acb7d6b774-kube-api-access-wbdxh\") on node \"crc\" DevicePath \"\"" Oct 06 10:18:21 crc kubenswrapper[4824]: I1006 10:18:21.948656 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8de9ec83-5bc3-47a8-9a63-84acb7d6b774-config\") on node \"crc\" DevicePath \"\"" Oct 06 10:18:21 crc kubenswrapper[4824]: I1006 10:18:21.948669 4824 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8de9ec83-5bc3-47a8-9a63-84acb7d6b774-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 06 10:18:21 crc kubenswrapper[4824]: I1006 10:18:21.948678 4824 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8de9ec83-5bc3-47a8-9a63-84acb7d6b774-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 06 10:18:22 crc kubenswrapper[4824]: I1006 10:18:22.395710 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-k2mqf" event={"ID":"8de9ec83-5bc3-47a8-9a63-84acb7d6b774","Type":"ContainerDied","Data":"eeb1879cd0e57c70a9013aa1b3b8d54c760330a1e66cc57dd26dd82535dfb51d"} Oct 06 10:18:22 crc kubenswrapper[4824]: I1006 10:18:22.395772 4824 scope.go:117] "RemoveContainer" containerID="dbc55ac9aa7e345cc7e9f4cc96d3aeba3693c9c1e166bdf53a0fa3d1eb241ab1" Oct 06 10:18:22 crc kubenswrapper[4824]: I1006 10:18:22.395937 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5784cf869f-k2mqf" Oct 06 10:18:22 crc kubenswrapper[4824]: I1006 10:18:22.404875 4824 generic.go:334] "Generic (PLEG): container finished" podID="cd95e609-bd54-4234-a449-66cd65e2558a" containerID="d7a21bc36e001b5d47e59fe8dd2712c72c5714c4632964f952a86eead6ad34bc" exitCode=0 Oct 06 10:18:22 crc kubenswrapper[4824]: I1006 10:18:22.404944 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-zq5td" event={"ID":"cd95e609-bd54-4234-a449-66cd65e2558a","Type":"ContainerDied","Data":"d7a21bc36e001b5d47e59fe8dd2712c72c5714c4632964f952a86eead6ad34bc"} Oct 06 10:18:22 crc kubenswrapper[4824]: I1006 10:18:22.438272 4824 scope.go:117] "RemoveContainer" containerID="63dd1ad8bb50bab14c438eb74051e3b85ef276dcae7b29ca87493769fd16c862" Oct 06 10:18:22 crc kubenswrapper[4824]: I1006 10:18:22.458414 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-k2mqf"] Oct 06 10:18:22 crc kubenswrapper[4824]: I1006 10:18:22.467664 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-k2mqf"] Oct 06 10:18:22 crc kubenswrapper[4824]: I1006 10:18:22.544055 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 06 10:18:22 crc kubenswrapper[4824]: I1006 10:18:22.790435 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 06 10:18:22 crc kubenswrapper[4824]: I1006 10:18:22.790535 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 06 10:18:22 crc kubenswrapper[4824]: I1006 10:18:22.877271 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-cbbc7" Oct 06 10:18:23 crc kubenswrapper[4824]: I1006 10:18:23.075807 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/795af81e-370f-4845-8632-acfe7988e3cf-combined-ca-bundle\") pod \"795af81e-370f-4845-8632-acfe7988e3cf\" (UID: \"795af81e-370f-4845-8632-acfe7988e3cf\") " Oct 06 10:18:23 crc kubenswrapper[4824]: I1006 10:18:23.075947 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/795af81e-370f-4845-8632-acfe7988e3cf-scripts\") pod \"795af81e-370f-4845-8632-acfe7988e3cf\" (UID: \"795af81e-370f-4845-8632-acfe7988e3cf\") " Oct 06 10:18:23 crc kubenswrapper[4824]: I1006 10:18:23.076173 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/795af81e-370f-4845-8632-acfe7988e3cf-config-data\") pod \"795af81e-370f-4845-8632-acfe7988e3cf\" (UID: \"795af81e-370f-4845-8632-acfe7988e3cf\") " Oct 06 10:18:23 crc kubenswrapper[4824]: I1006 10:18:23.076209 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wsrtp\" (UniqueName: \"kubernetes.io/projected/795af81e-370f-4845-8632-acfe7988e3cf-kube-api-access-wsrtp\") pod \"795af81e-370f-4845-8632-acfe7988e3cf\" (UID: \"795af81e-370f-4845-8632-acfe7988e3cf\") " Oct 06 10:18:23 crc kubenswrapper[4824]: I1006 10:18:23.083861 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/795af81e-370f-4845-8632-acfe7988e3cf-scripts" (OuterVolumeSpecName: "scripts") pod "795af81e-370f-4845-8632-acfe7988e3cf" (UID: "795af81e-370f-4845-8632-acfe7988e3cf"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:18:23 crc kubenswrapper[4824]: I1006 10:18:23.088262 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/795af81e-370f-4845-8632-acfe7988e3cf-kube-api-access-wsrtp" (OuterVolumeSpecName: "kube-api-access-wsrtp") pod "795af81e-370f-4845-8632-acfe7988e3cf" (UID: "795af81e-370f-4845-8632-acfe7988e3cf"). InnerVolumeSpecName "kube-api-access-wsrtp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:18:23 crc kubenswrapper[4824]: I1006 10:18:23.116533 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/795af81e-370f-4845-8632-acfe7988e3cf-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "795af81e-370f-4845-8632-acfe7988e3cf" (UID: "795af81e-370f-4845-8632-acfe7988e3cf"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:18:23 crc kubenswrapper[4824]: I1006 10:18:23.124304 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/795af81e-370f-4845-8632-acfe7988e3cf-config-data" (OuterVolumeSpecName: "config-data") pod "795af81e-370f-4845-8632-acfe7988e3cf" (UID: "795af81e-370f-4845-8632-acfe7988e3cf"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:18:23 crc kubenswrapper[4824]: I1006 10:18:23.180450 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/795af81e-370f-4845-8632-acfe7988e3cf-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 10:18:23 crc kubenswrapper[4824]: I1006 10:18:23.180499 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/795af81e-370f-4845-8632-acfe7988e3cf-scripts\") on node \"crc\" DevicePath \"\"" Oct 06 10:18:23 crc kubenswrapper[4824]: I1006 10:18:23.180514 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/795af81e-370f-4845-8632-acfe7988e3cf-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 10:18:23 crc kubenswrapper[4824]: I1006 10:18:23.180530 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wsrtp\" (UniqueName: \"kubernetes.io/projected/795af81e-370f-4845-8632-acfe7988e3cf-kube-api-access-wsrtp\") on node \"crc\" DevicePath \"\"" Oct 06 10:18:23 crc kubenswrapper[4824]: I1006 10:18:23.298334 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8de9ec83-5bc3-47a8-9a63-84acb7d6b774" path="/var/lib/kubelet/pods/8de9ec83-5bc3-47a8-9a63-84acb7d6b774/volumes" Oct 06 10:18:23 crc kubenswrapper[4824]: I1006 10:18:23.456879 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-cbbc7" event={"ID":"795af81e-370f-4845-8632-acfe7988e3cf","Type":"ContainerDied","Data":"969203d24cb1cd8e340b8662626d1b94b8a1955de432722bd801bc9599b8eab5"} Oct 06 10:18:23 crc kubenswrapper[4824]: I1006 10:18:23.456944 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="969203d24cb1cd8e340b8662626d1b94b8a1955de432722bd801bc9599b8eab5" Oct 06 10:18:23 crc kubenswrapper[4824]: I1006 10:18:23.456996 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-cbbc7" Oct 06 10:18:23 crc kubenswrapper[4824]: I1006 10:18:23.543225 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 06 10:18:23 crc kubenswrapper[4824]: E1006 10:18:23.543774 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8de9ec83-5bc3-47a8-9a63-84acb7d6b774" containerName="init" Oct 06 10:18:23 crc kubenswrapper[4824]: I1006 10:18:23.543793 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="8de9ec83-5bc3-47a8-9a63-84acb7d6b774" containerName="init" Oct 06 10:18:23 crc kubenswrapper[4824]: E1006 10:18:23.543816 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8de9ec83-5bc3-47a8-9a63-84acb7d6b774" containerName="dnsmasq-dns" Oct 06 10:18:23 crc kubenswrapper[4824]: I1006 10:18:23.543823 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="8de9ec83-5bc3-47a8-9a63-84acb7d6b774" containerName="dnsmasq-dns" Oct 06 10:18:23 crc kubenswrapper[4824]: E1006 10:18:23.543830 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="795af81e-370f-4845-8632-acfe7988e3cf" containerName="nova-cell1-conductor-db-sync" Oct 06 10:18:23 crc kubenswrapper[4824]: I1006 10:18:23.543837 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="795af81e-370f-4845-8632-acfe7988e3cf" containerName="nova-cell1-conductor-db-sync" Oct 06 10:18:23 crc kubenswrapper[4824]: I1006 10:18:23.544024 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="8de9ec83-5bc3-47a8-9a63-84acb7d6b774" containerName="dnsmasq-dns" Oct 06 10:18:23 crc kubenswrapper[4824]: I1006 10:18:23.544037 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="795af81e-370f-4845-8632-acfe7988e3cf" containerName="nova-cell1-conductor-db-sync" Oct 06 10:18:23 crc kubenswrapper[4824]: I1006 10:18:23.544745 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 06 10:18:23 crc kubenswrapper[4824]: I1006 10:18:23.547418 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 06 10:18:23 crc kubenswrapper[4824]: I1006 10:18:23.553590 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 06 10:18:23 crc kubenswrapper[4824]: E1006 10:18:23.600431 4824 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod795af81e_370f_4845_8632_acfe7988e3cf.slice/crio-969203d24cb1cd8e340b8662626d1b94b8a1955de432722bd801bc9599b8eab5\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod795af81e_370f_4845_8632_acfe7988e3cf.slice\": RecentStats: unable to find data in memory cache]" Oct 06 10:18:23 crc kubenswrapper[4824]: I1006 10:18:23.693282 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-st9lf\" (UniqueName: \"kubernetes.io/projected/e57703dd-1748-4d3d-8f2f-e4edc0901a76-kube-api-access-st9lf\") pod \"nova-cell1-conductor-0\" (UID: \"e57703dd-1748-4d3d-8f2f-e4edc0901a76\") " pod="openstack/nova-cell1-conductor-0" Oct 06 10:18:23 crc kubenswrapper[4824]: I1006 10:18:23.693472 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e57703dd-1748-4d3d-8f2f-e4edc0901a76-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"e57703dd-1748-4d3d-8f2f-e4edc0901a76\") " pod="openstack/nova-cell1-conductor-0" Oct 06 10:18:23 crc kubenswrapper[4824]: I1006 10:18:23.693551 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e57703dd-1748-4d3d-8f2f-e4edc0901a76-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"e57703dd-1748-4d3d-8f2f-e4edc0901a76\") " pod="openstack/nova-cell1-conductor-0" Oct 06 10:18:23 crc kubenswrapper[4824]: I1006 10:18:23.796633 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-st9lf\" (UniqueName: \"kubernetes.io/projected/e57703dd-1748-4d3d-8f2f-e4edc0901a76-kube-api-access-st9lf\") pod \"nova-cell1-conductor-0\" (UID: \"e57703dd-1748-4d3d-8f2f-e4edc0901a76\") " pod="openstack/nova-cell1-conductor-0" Oct 06 10:18:23 crc kubenswrapper[4824]: I1006 10:18:23.796709 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e57703dd-1748-4d3d-8f2f-e4edc0901a76-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"e57703dd-1748-4d3d-8f2f-e4edc0901a76\") " pod="openstack/nova-cell1-conductor-0" Oct 06 10:18:23 crc kubenswrapper[4824]: I1006 10:18:23.796739 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e57703dd-1748-4d3d-8f2f-e4edc0901a76-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"e57703dd-1748-4d3d-8f2f-e4edc0901a76\") " pod="openstack/nova-cell1-conductor-0" Oct 06 10:18:23 crc kubenswrapper[4824]: I1006 10:18:23.806097 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e57703dd-1748-4d3d-8f2f-e4edc0901a76-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"e57703dd-1748-4d3d-8f2f-e4edc0901a76\") " pod="openstack/nova-cell1-conductor-0" Oct 06 10:18:23 crc kubenswrapper[4824]: I1006 10:18:23.811944 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e57703dd-1748-4d3d-8f2f-e4edc0901a76-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"e57703dd-1748-4d3d-8f2f-e4edc0901a76\") " pod="openstack/nova-cell1-conductor-0" Oct 06 10:18:23 crc kubenswrapper[4824]: I1006 10:18:23.828194 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-st9lf\" (UniqueName: \"kubernetes.io/projected/e57703dd-1748-4d3d-8f2f-e4edc0901a76-kube-api-access-st9lf\") pod \"nova-cell1-conductor-0\" (UID: \"e57703dd-1748-4d3d-8f2f-e4edc0901a76\") " pod="openstack/nova-cell1-conductor-0" Oct 06 10:18:23 crc kubenswrapper[4824]: I1006 10:18:23.871324 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 06 10:18:23 crc kubenswrapper[4824]: I1006 10:18:23.994615 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-zq5td" Oct 06 10:18:24 crc kubenswrapper[4824]: I1006 10:18:24.109660 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd95e609-bd54-4234-a449-66cd65e2558a-combined-ca-bundle\") pod \"cd95e609-bd54-4234-a449-66cd65e2558a\" (UID: \"cd95e609-bd54-4234-a449-66cd65e2558a\") " Oct 06 10:18:24 crc kubenswrapper[4824]: I1006 10:18:24.110357 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd95e609-bd54-4234-a449-66cd65e2558a-config-data\") pod \"cd95e609-bd54-4234-a449-66cd65e2558a\" (UID: \"cd95e609-bd54-4234-a449-66cd65e2558a\") " Oct 06 10:18:24 crc kubenswrapper[4824]: I1006 10:18:24.110484 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vxkv2\" (UniqueName: \"kubernetes.io/projected/cd95e609-bd54-4234-a449-66cd65e2558a-kube-api-access-vxkv2\") pod \"cd95e609-bd54-4234-a449-66cd65e2558a\" (UID: \"cd95e609-bd54-4234-a449-66cd65e2558a\") " Oct 06 10:18:24 crc kubenswrapper[4824]: I1006 10:18:24.110531 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cd95e609-bd54-4234-a449-66cd65e2558a-scripts\") pod \"cd95e609-bd54-4234-a449-66cd65e2558a\" (UID: \"cd95e609-bd54-4234-a449-66cd65e2558a\") " Oct 06 10:18:24 crc kubenswrapper[4824]: I1006 10:18:24.116966 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd95e609-bd54-4234-a449-66cd65e2558a-scripts" (OuterVolumeSpecName: "scripts") pod "cd95e609-bd54-4234-a449-66cd65e2558a" (UID: "cd95e609-bd54-4234-a449-66cd65e2558a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:18:24 crc kubenswrapper[4824]: I1006 10:18:24.122766 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd95e609-bd54-4234-a449-66cd65e2558a-kube-api-access-vxkv2" (OuterVolumeSpecName: "kube-api-access-vxkv2") pod "cd95e609-bd54-4234-a449-66cd65e2558a" (UID: "cd95e609-bd54-4234-a449-66cd65e2558a"). InnerVolumeSpecName "kube-api-access-vxkv2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:18:24 crc kubenswrapper[4824]: I1006 10:18:24.154589 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd95e609-bd54-4234-a449-66cd65e2558a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cd95e609-bd54-4234-a449-66cd65e2558a" (UID: "cd95e609-bd54-4234-a449-66cd65e2558a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:18:24 crc kubenswrapper[4824]: I1006 10:18:24.156334 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd95e609-bd54-4234-a449-66cd65e2558a-config-data" (OuterVolumeSpecName: "config-data") pod "cd95e609-bd54-4234-a449-66cd65e2558a" (UID: "cd95e609-bd54-4234-a449-66cd65e2558a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:18:24 crc kubenswrapper[4824]: I1006 10:18:24.202578 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 06 10:18:24 crc kubenswrapper[4824]: I1006 10:18:24.214616 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vxkv2\" (UniqueName: \"kubernetes.io/projected/cd95e609-bd54-4234-a449-66cd65e2558a-kube-api-access-vxkv2\") on node \"crc\" DevicePath \"\"" Oct 06 10:18:24 crc kubenswrapper[4824]: I1006 10:18:24.214667 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cd95e609-bd54-4234-a449-66cd65e2558a-scripts\") on node \"crc\" DevicePath \"\"" Oct 06 10:18:24 crc kubenswrapper[4824]: I1006 10:18:24.214679 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd95e609-bd54-4234-a449-66cd65e2558a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 10:18:24 crc kubenswrapper[4824]: I1006 10:18:24.214689 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd95e609-bd54-4234-a449-66cd65e2558a-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 10:18:24 crc kubenswrapper[4824]: I1006 10:18:24.472335 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-zq5td" Oct 06 10:18:24 crc kubenswrapper[4824]: I1006 10:18:24.472351 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-zq5td" event={"ID":"cd95e609-bd54-4234-a449-66cd65e2558a","Type":"ContainerDied","Data":"da7f86acdc6cdf57487d522f158579fb5c3f5ff157072f1ac045c10b28947559"} Oct 06 10:18:24 crc kubenswrapper[4824]: I1006 10:18:24.472794 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="da7f86acdc6cdf57487d522f158579fb5c3f5ff157072f1ac045c10b28947559" Oct 06 10:18:24 crc kubenswrapper[4824]: I1006 10:18:24.473920 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"e57703dd-1748-4d3d-8f2f-e4edc0901a76","Type":"ContainerStarted","Data":"4b9efd476698ffe087683640b19a407ae0785fa27ba41dce2ada079969faf4ea"} Oct 06 10:18:24 crc kubenswrapper[4824]: I1006 10:18:24.473969 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"e57703dd-1748-4d3d-8f2f-e4edc0901a76","Type":"ContainerStarted","Data":"9ec9300ef0e3a91a8064776bd858e5b34e6cf37729331694b08435f595c29536"} Oct 06 10:18:24 crc kubenswrapper[4824]: I1006 10:18:24.474197 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Oct 06 10:18:24 crc kubenswrapper[4824]: I1006 10:18:24.500795 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=1.5007694520000001 podStartE2EDuration="1.500769452s" podCreationTimestamp="2025-10-06 10:18:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:18:24.496837625 +0000 UTC m=+1213.861260496" watchObservedRunningTime="2025-10-06 10:18:24.500769452 +0000 UTC m=+1213.865192313" Oct 06 10:18:24 crc kubenswrapper[4824]: I1006 10:18:24.611892 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 06 10:18:24 crc kubenswrapper[4824]: I1006 10:18:24.612154 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="d56c5ee8-59b8-45d6-9078-29423490982f" containerName="nova-api-log" containerID="cri-o://11e6da3b70f0831659aade67abc845e1bb7b3a14433d49fd0924bdbb9518e1ad" gracePeriod=30 Oct 06 10:18:24 crc kubenswrapper[4824]: I1006 10:18:24.612290 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="d56c5ee8-59b8-45d6-9078-29423490982f" containerName="nova-api-api" containerID="cri-o://488627b76866421d5a277cabe307c6c6ab3d23a1bca9671b420a9ca1953b960e" gracePeriod=30 Oct 06 10:18:24 crc kubenswrapper[4824]: I1006 10:18:24.638272 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 06 10:18:24 crc kubenswrapper[4824]: I1006 10:18:24.638950 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="27cc4522-3bd4-41bd-9fb7-624c634a5288" containerName="nova-scheduler-scheduler" containerID="cri-o://5aa36cdff7d909fd98c3a8a65be66bbf3f6a6b660ad12183d0da7f5a94b443b3" gracePeriod=30 Oct 06 10:18:24 crc kubenswrapper[4824]: I1006 10:18:24.658091 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 06 10:18:24 crc kubenswrapper[4824]: I1006 10:18:24.658376 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="7d722707-3130-4de6-ac0c-40d63a0fdff3" containerName="nova-metadata-log" containerID="cri-o://2578847baa70b643e2068c5f76c873450f4ab5b34ac01e713b1b85b4be7f1ecb" gracePeriod=30 Oct 06 10:18:24 crc kubenswrapper[4824]: I1006 10:18:24.658475 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="7d722707-3130-4de6-ac0c-40d63a0fdff3" containerName="nova-metadata-metadata" containerID="cri-o://7041e6ab4a92c1e02ee6f0ec262c66d7249ed1296d27f2ba4bedb3b75d530003" gracePeriod=30 Oct 06 10:18:25 crc kubenswrapper[4824]: I1006 10:18:25.199012 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 06 10:18:25 crc kubenswrapper[4824]: I1006 10:18:25.340257 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d722707-3130-4de6-ac0c-40d63a0fdff3-config-data\") pod \"7d722707-3130-4de6-ac0c-40d63a0fdff3\" (UID: \"7d722707-3130-4de6-ac0c-40d63a0fdff3\") " Oct 06 10:18:25 crc kubenswrapper[4824]: I1006 10:18:25.340318 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d722707-3130-4de6-ac0c-40d63a0fdff3-combined-ca-bundle\") pod \"7d722707-3130-4de6-ac0c-40d63a0fdff3\" (UID: \"7d722707-3130-4de6-ac0c-40d63a0fdff3\") " Oct 06 10:18:25 crc kubenswrapper[4824]: I1006 10:18:25.340504 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rzfdn\" (UniqueName: \"kubernetes.io/projected/7d722707-3130-4de6-ac0c-40d63a0fdff3-kube-api-access-rzfdn\") pod \"7d722707-3130-4de6-ac0c-40d63a0fdff3\" (UID: \"7d722707-3130-4de6-ac0c-40d63a0fdff3\") " Oct 06 10:18:25 crc kubenswrapper[4824]: I1006 10:18:25.340556 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/7d722707-3130-4de6-ac0c-40d63a0fdff3-nova-metadata-tls-certs\") pod \"7d722707-3130-4de6-ac0c-40d63a0fdff3\" (UID: \"7d722707-3130-4de6-ac0c-40d63a0fdff3\") " Oct 06 10:18:25 crc kubenswrapper[4824]: I1006 10:18:25.340677 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7d722707-3130-4de6-ac0c-40d63a0fdff3-logs\") pod \"7d722707-3130-4de6-ac0c-40d63a0fdff3\" (UID: \"7d722707-3130-4de6-ac0c-40d63a0fdff3\") " Oct 06 10:18:25 crc kubenswrapper[4824]: I1006 10:18:25.341300 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7d722707-3130-4de6-ac0c-40d63a0fdff3-logs" (OuterVolumeSpecName: "logs") pod "7d722707-3130-4de6-ac0c-40d63a0fdff3" (UID: "7d722707-3130-4de6-ac0c-40d63a0fdff3"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:18:25 crc kubenswrapper[4824]: I1006 10:18:25.358220 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7d722707-3130-4de6-ac0c-40d63a0fdff3-kube-api-access-rzfdn" (OuterVolumeSpecName: "kube-api-access-rzfdn") pod "7d722707-3130-4de6-ac0c-40d63a0fdff3" (UID: "7d722707-3130-4de6-ac0c-40d63a0fdff3"). InnerVolumeSpecName "kube-api-access-rzfdn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:18:25 crc kubenswrapper[4824]: I1006 10:18:25.419056 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d722707-3130-4de6-ac0c-40d63a0fdff3-config-data" (OuterVolumeSpecName: "config-data") pod "7d722707-3130-4de6-ac0c-40d63a0fdff3" (UID: "7d722707-3130-4de6-ac0c-40d63a0fdff3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:18:25 crc kubenswrapper[4824]: I1006 10:18:25.434512 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d722707-3130-4de6-ac0c-40d63a0fdff3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7d722707-3130-4de6-ac0c-40d63a0fdff3" (UID: "7d722707-3130-4de6-ac0c-40d63a0fdff3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:18:25 crc kubenswrapper[4824]: I1006 10:18:25.446301 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d722707-3130-4de6-ac0c-40d63a0fdff3-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 10:18:25 crc kubenswrapper[4824]: I1006 10:18:25.446341 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d722707-3130-4de6-ac0c-40d63a0fdff3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 10:18:25 crc kubenswrapper[4824]: I1006 10:18:25.446357 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rzfdn\" (UniqueName: \"kubernetes.io/projected/7d722707-3130-4de6-ac0c-40d63a0fdff3-kube-api-access-rzfdn\") on node \"crc\" DevicePath \"\"" Oct 06 10:18:25 crc kubenswrapper[4824]: I1006 10:18:25.446368 4824 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7d722707-3130-4de6-ac0c-40d63a0fdff3-logs\") on node \"crc\" DevicePath \"\"" Oct 06 10:18:25 crc kubenswrapper[4824]: I1006 10:18:25.457377 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d722707-3130-4de6-ac0c-40d63a0fdff3-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "7d722707-3130-4de6-ac0c-40d63a0fdff3" (UID: "7d722707-3130-4de6-ac0c-40d63a0fdff3"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:18:25 crc kubenswrapper[4824]: I1006 10:18:25.500548 4824 generic.go:334] "Generic (PLEG): container finished" podID="7d722707-3130-4de6-ac0c-40d63a0fdff3" containerID="7041e6ab4a92c1e02ee6f0ec262c66d7249ed1296d27f2ba4bedb3b75d530003" exitCode=0 Oct 06 10:18:25 crc kubenswrapper[4824]: I1006 10:18:25.500583 4824 generic.go:334] "Generic (PLEG): container finished" podID="7d722707-3130-4de6-ac0c-40d63a0fdff3" containerID="2578847baa70b643e2068c5f76c873450f4ab5b34ac01e713b1b85b4be7f1ecb" exitCode=143 Oct 06 10:18:25 crc kubenswrapper[4824]: I1006 10:18:25.500630 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"7d722707-3130-4de6-ac0c-40d63a0fdff3","Type":"ContainerDied","Data":"7041e6ab4a92c1e02ee6f0ec262c66d7249ed1296d27f2ba4bedb3b75d530003"} Oct 06 10:18:25 crc kubenswrapper[4824]: I1006 10:18:25.500660 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"7d722707-3130-4de6-ac0c-40d63a0fdff3","Type":"ContainerDied","Data":"2578847baa70b643e2068c5f76c873450f4ab5b34ac01e713b1b85b4be7f1ecb"} Oct 06 10:18:25 crc kubenswrapper[4824]: I1006 10:18:25.500670 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"7d722707-3130-4de6-ac0c-40d63a0fdff3","Type":"ContainerDied","Data":"4ee59599d2fbe9519798888d0c521f76b4f4642bb1fa3dba8614fffdbf28fc1c"} Oct 06 10:18:25 crc kubenswrapper[4824]: I1006 10:18:25.500689 4824 scope.go:117] "RemoveContainer" containerID="7041e6ab4a92c1e02ee6f0ec262c66d7249ed1296d27f2ba4bedb3b75d530003" Oct 06 10:18:25 crc kubenswrapper[4824]: I1006 10:18:25.500830 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 06 10:18:25 crc kubenswrapper[4824]: I1006 10:18:25.513807 4824 generic.go:334] "Generic (PLEG): container finished" podID="d56c5ee8-59b8-45d6-9078-29423490982f" containerID="11e6da3b70f0831659aade67abc845e1bb7b3a14433d49fd0924bdbb9518e1ad" exitCode=143 Oct 06 10:18:25 crc kubenswrapper[4824]: I1006 10:18:25.514259 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d56c5ee8-59b8-45d6-9078-29423490982f","Type":"ContainerDied","Data":"11e6da3b70f0831659aade67abc845e1bb7b3a14433d49fd0924bdbb9518e1ad"} Oct 06 10:18:25 crc kubenswrapper[4824]: I1006 10:18:25.546268 4824 scope.go:117] "RemoveContainer" containerID="2578847baa70b643e2068c5f76c873450f4ab5b34ac01e713b1b85b4be7f1ecb" Oct 06 10:18:25 crc kubenswrapper[4824]: I1006 10:18:25.548133 4824 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/7d722707-3130-4de6-ac0c-40d63a0fdff3-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 06 10:18:25 crc kubenswrapper[4824]: I1006 10:18:25.593046 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 06 10:18:25 crc kubenswrapper[4824]: I1006 10:18:25.621225 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 06 10:18:25 crc kubenswrapper[4824]: I1006 10:18:25.637400 4824 scope.go:117] "RemoveContainer" containerID="7041e6ab4a92c1e02ee6f0ec262c66d7249ed1296d27f2ba4bedb3b75d530003" Oct 06 10:18:25 crc kubenswrapper[4824]: E1006 10:18:25.638814 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7041e6ab4a92c1e02ee6f0ec262c66d7249ed1296d27f2ba4bedb3b75d530003\": container with ID starting with 7041e6ab4a92c1e02ee6f0ec262c66d7249ed1296d27f2ba4bedb3b75d530003 not found: ID does not exist" containerID="7041e6ab4a92c1e02ee6f0ec262c66d7249ed1296d27f2ba4bedb3b75d530003" Oct 06 10:18:25 crc kubenswrapper[4824]: I1006 10:18:25.638843 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7041e6ab4a92c1e02ee6f0ec262c66d7249ed1296d27f2ba4bedb3b75d530003"} err="failed to get container status \"7041e6ab4a92c1e02ee6f0ec262c66d7249ed1296d27f2ba4bedb3b75d530003\": rpc error: code = NotFound desc = could not find container \"7041e6ab4a92c1e02ee6f0ec262c66d7249ed1296d27f2ba4bedb3b75d530003\": container with ID starting with 7041e6ab4a92c1e02ee6f0ec262c66d7249ed1296d27f2ba4bedb3b75d530003 not found: ID does not exist" Oct 06 10:18:25 crc kubenswrapper[4824]: I1006 10:18:25.638863 4824 scope.go:117] "RemoveContainer" containerID="2578847baa70b643e2068c5f76c873450f4ab5b34ac01e713b1b85b4be7f1ecb" Oct 06 10:18:25 crc kubenswrapper[4824]: E1006 10:18:25.642051 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2578847baa70b643e2068c5f76c873450f4ab5b34ac01e713b1b85b4be7f1ecb\": container with ID starting with 2578847baa70b643e2068c5f76c873450f4ab5b34ac01e713b1b85b4be7f1ecb not found: ID does not exist" containerID="2578847baa70b643e2068c5f76c873450f4ab5b34ac01e713b1b85b4be7f1ecb" Oct 06 10:18:25 crc kubenswrapper[4824]: I1006 10:18:25.642078 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2578847baa70b643e2068c5f76c873450f4ab5b34ac01e713b1b85b4be7f1ecb"} err="failed to get container status \"2578847baa70b643e2068c5f76c873450f4ab5b34ac01e713b1b85b4be7f1ecb\": rpc error: code = NotFound desc = could not find container \"2578847baa70b643e2068c5f76c873450f4ab5b34ac01e713b1b85b4be7f1ecb\": container with ID starting with 2578847baa70b643e2068c5f76c873450f4ab5b34ac01e713b1b85b4be7f1ecb not found: ID does not exist" Oct 06 10:18:25 crc kubenswrapper[4824]: I1006 10:18:25.642107 4824 scope.go:117] "RemoveContainer" containerID="7041e6ab4a92c1e02ee6f0ec262c66d7249ed1296d27f2ba4bedb3b75d530003" Oct 06 10:18:25 crc kubenswrapper[4824]: I1006 10:18:25.642158 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 06 10:18:25 crc kubenswrapper[4824]: E1006 10:18:25.642594 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd95e609-bd54-4234-a449-66cd65e2558a" containerName="nova-manage" Oct 06 10:18:25 crc kubenswrapper[4824]: I1006 10:18:25.642611 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd95e609-bd54-4234-a449-66cd65e2558a" containerName="nova-manage" Oct 06 10:18:25 crc kubenswrapper[4824]: E1006 10:18:25.642622 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d722707-3130-4de6-ac0c-40d63a0fdff3" containerName="nova-metadata-log" Oct 06 10:18:25 crc kubenswrapper[4824]: I1006 10:18:25.642631 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d722707-3130-4de6-ac0c-40d63a0fdff3" containerName="nova-metadata-log" Oct 06 10:18:25 crc kubenswrapper[4824]: E1006 10:18:25.642658 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d722707-3130-4de6-ac0c-40d63a0fdff3" containerName="nova-metadata-metadata" Oct 06 10:18:25 crc kubenswrapper[4824]: I1006 10:18:25.642664 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d722707-3130-4de6-ac0c-40d63a0fdff3" containerName="nova-metadata-metadata" Oct 06 10:18:25 crc kubenswrapper[4824]: I1006 10:18:25.642857 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="cd95e609-bd54-4234-a449-66cd65e2558a" containerName="nova-manage" Oct 06 10:18:25 crc kubenswrapper[4824]: I1006 10:18:25.642884 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="7d722707-3130-4de6-ac0c-40d63a0fdff3" containerName="nova-metadata-metadata" Oct 06 10:18:25 crc kubenswrapper[4824]: I1006 10:18:25.642903 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="7d722707-3130-4de6-ac0c-40d63a0fdff3" containerName="nova-metadata-log" Oct 06 10:18:25 crc kubenswrapper[4824]: I1006 10:18:25.643880 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 06 10:18:25 crc kubenswrapper[4824]: I1006 10:18:25.665581 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7041e6ab4a92c1e02ee6f0ec262c66d7249ed1296d27f2ba4bedb3b75d530003"} err="failed to get container status \"7041e6ab4a92c1e02ee6f0ec262c66d7249ed1296d27f2ba4bedb3b75d530003\": rpc error: code = NotFound desc = could not find container \"7041e6ab4a92c1e02ee6f0ec262c66d7249ed1296d27f2ba4bedb3b75d530003\": container with ID starting with 7041e6ab4a92c1e02ee6f0ec262c66d7249ed1296d27f2ba4bedb3b75d530003 not found: ID does not exist" Oct 06 10:18:25 crc kubenswrapper[4824]: I1006 10:18:25.665624 4824 scope.go:117] "RemoveContainer" containerID="2578847baa70b643e2068c5f76c873450f4ab5b34ac01e713b1b85b4be7f1ecb" Oct 06 10:18:25 crc kubenswrapper[4824]: I1006 10:18:25.665949 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 06 10:18:25 crc kubenswrapper[4824]: I1006 10:18:25.666120 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 06 10:18:25 crc kubenswrapper[4824]: I1006 10:18:25.669313 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 06 10:18:25 crc kubenswrapper[4824]: I1006 10:18:25.675220 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2578847baa70b643e2068c5f76c873450f4ab5b34ac01e713b1b85b4be7f1ecb"} err="failed to get container status \"2578847baa70b643e2068c5f76c873450f4ab5b34ac01e713b1b85b4be7f1ecb\": rpc error: code = NotFound desc = could not find container \"2578847baa70b643e2068c5f76c873450f4ab5b34ac01e713b1b85b4be7f1ecb\": container with ID starting with 2578847baa70b643e2068c5f76c873450f4ab5b34ac01e713b1b85b4be7f1ecb not found: ID does not exist" Oct 06 10:18:25 crc kubenswrapper[4824]: I1006 10:18:25.755397 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/38f30d16-bb2b-4a0b-aac6-64078e032eeb-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"38f30d16-bb2b-4a0b-aac6-64078e032eeb\") " pod="openstack/nova-metadata-0" Oct 06 10:18:25 crc kubenswrapper[4824]: I1006 10:18:25.755501 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/38f30d16-bb2b-4a0b-aac6-64078e032eeb-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"38f30d16-bb2b-4a0b-aac6-64078e032eeb\") " pod="openstack/nova-metadata-0" Oct 06 10:18:25 crc kubenswrapper[4824]: I1006 10:18:25.755523 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vh8r5\" (UniqueName: \"kubernetes.io/projected/38f30d16-bb2b-4a0b-aac6-64078e032eeb-kube-api-access-vh8r5\") pod \"nova-metadata-0\" (UID: \"38f30d16-bb2b-4a0b-aac6-64078e032eeb\") " pod="openstack/nova-metadata-0" Oct 06 10:18:25 crc kubenswrapper[4824]: I1006 10:18:25.755574 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/38f30d16-bb2b-4a0b-aac6-64078e032eeb-logs\") pod \"nova-metadata-0\" (UID: \"38f30d16-bb2b-4a0b-aac6-64078e032eeb\") " pod="openstack/nova-metadata-0" Oct 06 10:18:25 crc kubenswrapper[4824]: I1006 10:18:25.755644 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/38f30d16-bb2b-4a0b-aac6-64078e032eeb-config-data\") pod \"nova-metadata-0\" (UID: \"38f30d16-bb2b-4a0b-aac6-64078e032eeb\") " pod="openstack/nova-metadata-0" Oct 06 10:18:25 crc kubenswrapper[4824]: I1006 10:18:25.858290 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/38f30d16-bb2b-4a0b-aac6-64078e032eeb-logs\") pod \"nova-metadata-0\" (UID: \"38f30d16-bb2b-4a0b-aac6-64078e032eeb\") " pod="openstack/nova-metadata-0" Oct 06 10:18:25 crc kubenswrapper[4824]: I1006 10:18:25.858476 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/38f30d16-bb2b-4a0b-aac6-64078e032eeb-config-data\") pod \"nova-metadata-0\" (UID: \"38f30d16-bb2b-4a0b-aac6-64078e032eeb\") " pod="openstack/nova-metadata-0" Oct 06 10:18:25 crc kubenswrapper[4824]: I1006 10:18:25.858507 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/38f30d16-bb2b-4a0b-aac6-64078e032eeb-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"38f30d16-bb2b-4a0b-aac6-64078e032eeb\") " pod="openstack/nova-metadata-0" Oct 06 10:18:25 crc kubenswrapper[4824]: I1006 10:18:25.858578 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/38f30d16-bb2b-4a0b-aac6-64078e032eeb-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"38f30d16-bb2b-4a0b-aac6-64078e032eeb\") " pod="openstack/nova-metadata-0" Oct 06 10:18:25 crc kubenswrapper[4824]: I1006 10:18:25.858600 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vh8r5\" (UniqueName: \"kubernetes.io/projected/38f30d16-bb2b-4a0b-aac6-64078e032eeb-kube-api-access-vh8r5\") pod \"nova-metadata-0\" (UID: \"38f30d16-bb2b-4a0b-aac6-64078e032eeb\") " pod="openstack/nova-metadata-0" Oct 06 10:18:25 crc kubenswrapper[4824]: I1006 10:18:25.858924 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/38f30d16-bb2b-4a0b-aac6-64078e032eeb-logs\") pod \"nova-metadata-0\" (UID: \"38f30d16-bb2b-4a0b-aac6-64078e032eeb\") " pod="openstack/nova-metadata-0" Oct 06 10:18:25 crc kubenswrapper[4824]: I1006 10:18:25.866567 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/38f30d16-bb2b-4a0b-aac6-64078e032eeb-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"38f30d16-bb2b-4a0b-aac6-64078e032eeb\") " pod="openstack/nova-metadata-0" Oct 06 10:18:25 crc kubenswrapper[4824]: I1006 10:18:25.867497 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/38f30d16-bb2b-4a0b-aac6-64078e032eeb-config-data\") pod \"nova-metadata-0\" (UID: \"38f30d16-bb2b-4a0b-aac6-64078e032eeb\") " pod="openstack/nova-metadata-0" Oct 06 10:18:25 crc kubenswrapper[4824]: I1006 10:18:25.875764 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/38f30d16-bb2b-4a0b-aac6-64078e032eeb-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"38f30d16-bb2b-4a0b-aac6-64078e032eeb\") " pod="openstack/nova-metadata-0" Oct 06 10:18:25 crc kubenswrapper[4824]: I1006 10:18:25.877557 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vh8r5\" (UniqueName: \"kubernetes.io/projected/38f30d16-bb2b-4a0b-aac6-64078e032eeb-kube-api-access-vh8r5\") pod \"nova-metadata-0\" (UID: \"38f30d16-bb2b-4a0b-aac6-64078e032eeb\") " pod="openstack/nova-metadata-0" Oct 06 10:18:25 crc kubenswrapper[4824]: I1006 10:18:25.988114 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 06 10:18:26 crc kubenswrapper[4824]: E1006 10:18:26.010426 4824 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5aa36cdff7d909fd98c3a8a65be66bbf3f6a6b660ad12183d0da7f5a94b443b3" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 06 10:18:26 crc kubenswrapper[4824]: E1006 10:18:26.012153 4824 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5aa36cdff7d909fd98c3a8a65be66bbf3f6a6b660ad12183d0da7f5a94b443b3" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 06 10:18:26 crc kubenswrapper[4824]: E1006 10:18:26.017095 4824 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5aa36cdff7d909fd98c3a8a65be66bbf3f6a6b660ad12183d0da7f5a94b443b3" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 06 10:18:26 crc kubenswrapper[4824]: E1006 10:18:26.017199 4824 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="27cc4522-3bd4-41bd-9fb7-624c634a5288" containerName="nova-scheduler-scheduler" Oct 06 10:18:26 crc kubenswrapper[4824]: I1006 10:18:26.511306 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 06 10:18:26 crc kubenswrapper[4824]: I1006 10:18:26.530560 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"38f30d16-bb2b-4a0b-aac6-64078e032eeb","Type":"ContainerStarted","Data":"bd41727b70698dc94ba3e3fb0c13a2822d04f71771ae3b6ad6d7626d71132a91"} Oct 06 10:18:27 crc kubenswrapper[4824]: I1006 10:18:27.048604 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 06 10:18:27 crc kubenswrapper[4824]: I1006 10:18:27.049296 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="1d365fcf-17fb-411e-ab2c-3c5dc8384b21" containerName="kube-state-metrics" containerID="cri-o://5bbaaf09ee506a794b2f4f3000b75a5e8afaa034fab8e585936a61b6d460e76d" gracePeriod=30 Oct 06 10:18:27 crc kubenswrapper[4824]: I1006 10:18:27.295062 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7d722707-3130-4de6-ac0c-40d63a0fdff3" path="/var/lib/kubelet/pods/7d722707-3130-4de6-ac0c-40d63a0fdff3/volumes" Oct 06 10:18:27 crc kubenswrapper[4824]: I1006 10:18:27.548418 4824 generic.go:334] "Generic (PLEG): container finished" podID="1d365fcf-17fb-411e-ab2c-3c5dc8384b21" containerID="5bbaaf09ee506a794b2f4f3000b75a5e8afaa034fab8e585936a61b6d460e76d" exitCode=2 Oct 06 10:18:27 crc kubenswrapper[4824]: I1006 10:18:27.548496 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"1d365fcf-17fb-411e-ab2c-3c5dc8384b21","Type":"ContainerDied","Data":"5bbaaf09ee506a794b2f4f3000b75a5e8afaa034fab8e585936a61b6d460e76d"} Oct 06 10:18:27 crc kubenswrapper[4824]: I1006 10:18:27.552789 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"38f30d16-bb2b-4a0b-aac6-64078e032eeb","Type":"ContainerStarted","Data":"aa67c97d929e74e959591f108ea9503ddd0ba9c3e4d21ac0322b719670baaf66"} Oct 06 10:18:27 crc kubenswrapper[4824]: I1006 10:18:27.553291 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"38f30d16-bb2b-4a0b-aac6-64078e032eeb","Type":"ContainerStarted","Data":"e582782b73e98e3ec11c149376d44e09b4ebf2e98421da7607dfbc12b3b526a9"} Oct 06 10:18:27 crc kubenswrapper[4824]: I1006 10:18:27.582674 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.582621666 podStartE2EDuration="2.582621666s" podCreationTimestamp="2025-10-06 10:18:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:18:27.57788512 +0000 UTC m=+1216.942307981" watchObservedRunningTime="2025-10-06 10:18:27.582621666 +0000 UTC m=+1216.947044537" Oct 06 10:18:27 crc kubenswrapper[4824]: I1006 10:18:27.656154 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 06 10:18:27 crc kubenswrapper[4824]: I1006 10:18:27.708385 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mplf5\" (UniqueName: \"kubernetes.io/projected/1d365fcf-17fb-411e-ab2c-3c5dc8384b21-kube-api-access-mplf5\") pod \"1d365fcf-17fb-411e-ab2c-3c5dc8384b21\" (UID: \"1d365fcf-17fb-411e-ab2c-3c5dc8384b21\") " Oct 06 10:18:27 crc kubenswrapper[4824]: I1006 10:18:27.731855 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d365fcf-17fb-411e-ab2c-3c5dc8384b21-kube-api-access-mplf5" (OuterVolumeSpecName: "kube-api-access-mplf5") pod "1d365fcf-17fb-411e-ab2c-3c5dc8384b21" (UID: "1d365fcf-17fb-411e-ab2c-3c5dc8384b21"). InnerVolumeSpecName "kube-api-access-mplf5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:18:27 crc kubenswrapper[4824]: I1006 10:18:27.810657 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mplf5\" (UniqueName: \"kubernetes.io/projected/1d365fcf-17fb-411e-ab2c-3c5dc8384b21-kube-api-access-mplf5\") on node \"crc\" DevicePath \"\"" Oct 06 10:18:28 crc kubenswrapper[4824]: I1006 10:18:28.481135 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 06 10:18:28 crc kubenswrapper[4824]: I1006 10:18:28.566544 4824 generic.go:334] "Generic (PLEG): container finished" podID="d56c5ee8-59b8-45d6-9078-29423490982f" containerID="488627b76866421d5a277cabe307c6c6ab3d23a1bca9671b420a9ca1953b960e" exitCode=0 Oct 06 10:18:28 crc kubenswrapper[4824]: I1006 10:18:28.566603 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d56c5ee8-59b8-45d6-9078-29423490982f","Type":"ContainerDied","Data":"488627b76866421d5a277cabe307c6c6ab3d23a1bca9671b420a9ca1953b960e"} Oct 06 10:18:28 crc kubenswrapper[4824]: I1006 10:18:28.566631 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d56c5ee8-59b8-45d6-9078-29423490982f","Type":"ContainerDied","Data":"41a9a04e7c022b18eaf8b6513683d463b433430bb44bb3e467f2eebc574f3c89"} Oct 06 10:18:28 crc kubenswrapper[4824]: I1006 10:18:28.566651 4824 scope.go:117] "RemoveContainer" containerID="488627b76866421d5a277cabe307c6c6ab3d23a1bca9671b420a9ca1953b960e" Oct 06 10:18:28 crc kubenswrapper[4824]: I1006 10:18:28.566769 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 06 10:18:28 crc kubenswrapper[4824]: I1006 10:18:28.571322 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 06 10:18:28 crc kubenswrapper[4824]: I1006 10:18:28.571354 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"1d365fcf-17fb-411e-ab2c-3c5dc8384b21","Type":"ContainerDied","Data":"72614a4f22524423b9852d1c0767b06c5bb22df6e098392a586f481e71f95774"} Oct 06 10:18:28 crc kubenswrapper[4824]: I1006 10:18:28.575244 4824 generic.go:334] "Generic (PLEG): container finished" podID="27cc4522-3bd4-41bd-9fb7-624c634a5288" containerID="5aa36cdff7d909fd98c3a8a65be66bbf3f6a6b660ad12183d0da7f5a94b443b3" exitCode=0 Oct 06 10:18:28 crc kubenswrapper[4824]: I1006 10:18:28.575316 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"27cc4522-3bd4-41bd-9fb7-624c634a5288","Type":"ContainerDied","Data":"5aa36cdff7d909fd98c3a8a65be66bbf3f6a6b660ad12183d0da7f5a94b443b3"} Oct 06 10:18:28 crc kubenswrapper[4824]: I1006 10:18:28.575367 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"27cc4522-3bd4-41bd-9fb7-624c634a5288","Type":"ContainerDied","Data":"13d5cd70391e4730d3bf4f7643e877628d5fd4fa929172f4c435fe8c9b55c929"} Oct 06 10:18:28 crc kubenswrapper[4824]: I1006 10:18:28.575378 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="13d5cd70391e4730d3bf4f7643e877628d5fd4fa929172f4c435fe8c9b55c929" Oct 06 10:18:28 crc kubenswrapper[4824]: I1006 10:18:28.638841 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d56c5ee8-59b8-45d6-9078-29423490982f-combined-ca-bundle\") pod \"d56c5ee8-59b8-45d6-9078-29423490982f\" (UID: \"d56c5ee8-59b8-45d6-9078-29423490982f\") " Oct 06 10:18:28 crc kubenswrapper[4824]: I1006 10:18:28.639050 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d56c5ee8-59b8-45d6-9078-29423490982f-config-data\") pod \"d56c5ee8-59b8-45d6-9078-29423490982f\" (UID: \"d56c5ee8-59b8-45d6-9078-29423490982f\") " Oct 06 10:18:28 crc kubenswrapper[4824]: I1006 10:18:28.639326 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d56c5ee8-59b8-45d6-9078-29423490982f-logs\") pod \"d56c5ee8-59b8-45d6-9078-29423490982f\" (UID: \"d56c5ee8-59b8-45d6-9078-29423490982f\") " Oct 06 10:18:28 crc kubenswrapper[4824]: I1006 10:18:28.639376 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hdzkz\" (UniqueName: \"kubernetes.io/projected/d56c5ee8-59b8-45d6-9078-29423490982f-kube-api-access-hdzkz\") pod \"d56c5ee8-59b8-45d6-9078-29423490982f\" (UID: \"d56c5ee8-59b8-45d6-9078-29423490982f\") " Oct 06 10:18:28 crc kubenswrapper[4824]: I1006 10:18:28.639971 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d56c5ee8-59b8-45d6-9078-29423490982f-logs" (OuterVolumeSpecName: "logs") pod "d56c5ee8-59b8-45d6-9078-29423490982f" (UID: "d56c5ee8-59b8-45d6-9078-29423490982f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:18:28 crc kubenswrapper[4824]: I1006 10:18:28.640118 4824 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d56c5ee8-59b8-45d6-9078-29423490982f-logs\") on node \"crc\" DevicePath \"\"" Oct 06 10:18:28 crc kubenswrapper[4824]: I1006 10:18:28.645204 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d56c5ee8-59b8-45d6-9078-29423490982f-kube-api-access-hdzkz" (OuterVolumeSpecName: "kube-api-access-hdzkz") pod "d56c5ee8-59b8-45d6-9078-29423490982f" (UID: "d56c5ee8-59b8-45d6-9078-29423490982f"). InnerVolumeSpecName "kube-api-access-hdzkz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:18:28 crc kubenswrapper[4824]: I1006 10:18:28.653616 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 06 10:18:28 crc kubenswrapper[4824]: I1006 10:18:28.673462 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 06 10:18:28 crc kubenswrapper[4824]: I1006 10:18:28.674123 4824 scope.go:117] "RemoveContainer" containerID="11e6da3b70f0831659aade67abc845e1bb7b3a14433d49fd0924bdbb9518e1ad" Oct 06 10:18:28 crc kubenswrapper[4824]: I1006 10:18:28.687638 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d56c5ee8-59b8-45d6-9078-29423490982f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d56c5ee8-59b8-45d6-9078-29423490982f" (UID: "d56c5ee8-59b8-45d6-9078-29423490982f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:18:28 crc kubenswrapper[4824]: I1006 10:18:28.689192 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 06 10:18:28 crc kubenswrapper[4824]: I1006 10:18:28.707165 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d56c5ee8-59b8-45d6-9078-29423490982f-config-data" (OuterVolumeSpecName: "config-data") pod "d56c5ee8-59b8-45d6-9078-29423490982f" (UID: "d56c5ee8-59b8-45d6-9078-29423490982f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:18:28 crc kubenswrapper[4824]: I1006 10:18:28.723572 4824 scope.go:117] "RemoveContainer" containerID="488627b76866421d5a277cabe307c6c6ab3d23a1bca9671b420a9ca1953b960e" Oct 06 10:18:28 crc kubenswrapper[4824]: E1006 10:18:28.724272 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"488627b76866421d5a277cabe307c6c6ab3d23a1bca9671b420a9ca1953b960e\": container with ID starting with 488627b76866421d5a277cabe307c6c6ab3d23a1bca9671b420a9ca1953b960e not found: ID does not exist" containerID="488627b76866421d5a277cabe307c6c6ab3d23a1bca9671b420a9ca1953b960e" Oct 06 10:18:28 crc kubenswrapper[4824]: I1006 10:18:28.724341 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"488627b76866421d5a277cabe307c6c6ab3d23a1bca9671b420a9ca1953b960e"} err="failed to get container status \"488627b76866421d5a277cabe307c6c6ab3d23a1bca9671b420a9ca1953b960e\": rpc error: code = NotFound desc = could not find container \"488627b76866421d5a277cabe307c6c6ab3d23a1bca9671b420a9ca1953b960e\": container with ID starting with 488627b76866421d5a277cabe307c6c6ab3d23a1bca9671b420a9ca1953b960e not found: ID does not exist" Oct 06 10:18:28 crc kubenswrapper[4824]: I1006 10:18:28.724379 4824 scope.go:117] "RemoveContainer" containerID="11e6da3b70f0831659aade67abc845e1bb7b3a14433d49fd0924bdbb9518e1ad" Oct 06 10:18:28 crc kubenswrapper[4824]: E1006 10:18:28.724807 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"11e6da3b70f0831659aade67abc845e1bb7b3a14433d49fd0924bdbb9518e1ad\": container with ID starting with 11e6da3b70f0831659aade67abc845e1bb7b3a14433d49fd0924bdbb9518e1ad not found: ID does not exist" containerID="11e6da3b70f0831659aade67abc845e1bb7b3a14433d49fd0924bdbb9518e1ad" Oct 06 10:18:28 crc kubenswrapper[4824]: I1006 10:18:28.724840 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"11e6da3b70f0831659aade67abc845e1bb7b3a14433d49fd0924bdbb9518e1ad"} err="failed to get container status \"11e6da3b70f0831659aade67abc845e1bb7b3a14433d49fd0924bdbb9518e1ad\": rpc error: code = NotFound desc = could not find container \"11e6da3b70f0831659aade67abc845e1bb7b3a14433d49fd0924bdbb9518e1ad\": container with ID starting with 11e6da3b70f0831659aade67abc845e1bb7b3a14433d49fd0924bdbb9518e1ad not found: ID does not exist" Oct 06 10:18:28 crc kubenswrapper[4824]: I1006 10:18:28.724863 4824 scope.go:117] "RemoveContainer" containerID="5bbaaf09ee506a794b2f4f3000b75a5e8afaa034fab8e585936a61b6d460e76d" Oct 06 10:18:28 crc kubenswrapper[4824]: I1006 10:18:28.730503 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Oct 06 10:18:28 crc kubenswrapper[4824]: E1006 10:18:28.731161 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d56c5ee8-59b8-45d6-9078-29423490982f" containerName="nova-api-log" Oct 06 10:18:28 crc kubenswrapper[4824]: I1006 10:18:28.731186 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="d56c5ee8-59b8-45d6-9078-29423490982f" containerName="nova-api-log" Oct 06 10:18:28 crc kubenswrapper[4824]: E1006 10:18:28.731203 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d56c5ee8-59b8-45d6-9078-29423490982f" containerName="nova-api-api" Oct 06 10:18:28 crc kubenswrapper[4824]: I1006 10:18:28.731211 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="d56c5ee8-59b8-45d6-9078-29423490982f" containerName="nova-api-api" Oct 06 10:18:28 crc kubenswrapper[4824]: E1006 10:18:28.731225 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d365fcf-17fb-411e-ab2c-3c5dc8384b21" containerName="kube-state-metrics" Oct 06 10:18:28 crc kubenswrapper[4824]: I1006 10:18:28.731233 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d365fcf-17fb-411e-ab2c-3c5dc8384b21" containerName="kube-state-metrics" Oct 06 10:18:28 crc kubenswrapper[4824]: E1006 10:18:28.731277 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27cc4522-3bd4-41bd-9fb7-624c634a5288" containerName="nova-scheduler-scheduler" Oct 06 10:18:28 crc kubenswrapper[4824]: I1006 10:18:28.731286 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="27cc4522-3bd4-41bd-9fb7-624c634a5288" containerName="nova-scheduler-scheduler" Oct 06 10:18:28 crc kubenswrapper[4824]: I1006 10:18:28.731506 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="27cc4522-3bd4-41bd-9fb7-624c634a5288" containerName="nova-scheduler-scheduler" Oct 06 10:18:28 crc kubenswrapper[4824]: I1006 10:18:28.731526 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="d56c5ee8-59b8-45d6-9078-29423490982f" containerName="nova-api-log" Oct 06 10:18:28 crc kubenswrapper[4824]: I1006 10:18:28.731542 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d365fcf-17fb-411e-ab2c-3c5dc8384b21" containerName="kube-state-metrics" Oct 06 10:18:28 crc kubenswrapper[4824]: I1006 10:18:28.731563 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="d56c5ee8-59b8-45d6-9078-29423490982f" containerName="nova-api-api" Oct 06 10:18:28 crc kubenswrapper[4824]: I1006 10:18:28.732460 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 06 10:18:28 crc kubenswrapper[4824]: I1006 10:18:28.735788 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Oct 06 10:18:28 crc kubenswrapper[4824]: I1006 10:18:28.735815 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Oct 06 10:18:28 crc kubenswrapper[4824]: I1006 10:18:28.741811 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5cbxw\" (UniqueName: \"kubernetes.io/projected/27cc4522-3bd4-41bd-9fb7-624c634a5288-kube-api-access-5cbxw\") pod \"27cc4522-3bd4-41bd-9fb7-624c634a5288\" (UID: \"27cc4522-3bd4-41bd-9fb7-624c634a5288\") " Oct 06 10:18:28 crc kubenswrapper[4824]: I1006 10:18:28.742055 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27cc4522-3bd4-41bd-9fb7-624c634a5288-combined-ca-bundle\") pod \"27cc4522-3bd4-41bd-9fb7-624c634a5288\" (UID: \"27cc4522-3bd4-41bd-9fb7-624c634a5288\") " Oct 06 10:18:28 crc kubenswrapper[4824]: I1006 10:18:28.742287 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/27cc4522-3bd4-41bd-9fb7-624c634a5288-config-data\") pod \"27cc4522-3bd4-41bd-9fb7-624c634a5288\" (UID: \"27cc4522-3bd4-41bd-9fb7-624c634a5288\") " Oct 06 10:18:28 crc kubenswrapper[4824]: I1006 10:18:28.747555 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/58776586-f2ad-4b0e-bf7e-303324ffe7f2-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"58776586-f2ad-4b0e-bf7e-303324ffe7f2\") " pod="openstack/kube-state-metrics-0" Oct 06 10:18:28 crc kubenswrapper[4824]: I1006 10:18:28.747925 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/58776586-f2ad-4b0e-bf7e-303324ffe7f2-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"58776586-f2ad-4b0e-bf7e-303324ffe7f2\") " pod="openstack/kube-state-metrics-0" Oct 06 10:18:28 crc kubenswrapper[4824]: I1006 10:18:28.749535 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kfsjk\" (UniqueName: \"kubernetes.io/projected/58776586-f2ad-4b0e-bf7e-303324ffe7f2-kube-api-access-kfsjk\") pod \"kube-state-metrics-0\" (UID: \"58776586-f2ad-4b0e-bf7e-303324ffe7f2\") " pod="openstack/kube-state-metrics-0" Oct 06 10:18:28 crc kubenswrapper[4824]: I1006 10:18:28.749683 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58776586-f2ad-4b0e-bf7e-303324ffe7f2-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"58776586-f2ad-4b0e-bf7e-303324ffe7f2\") " pod="openstack/kube-state-metrics-0" Oct 06 10:18:28 crc kubenswrapper[4824]: I1006 10:18:28.749837 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d56c5ee8-59b8-45d6-9078-29423490982f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 10:18:28 crc kubenswrapper[4824]: I1006 10:18:28.749851 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d56c5ee8-59b8-45d6-9078-29423490982f-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 10:18:28 crc kubenswrapper[4824]: I1006 10:18:28.749863 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hdzkz\" (UniqueName: \"kubernetes.io/projected/d56c5ee8-59b8-45d6-9078-29423490982f-kube-api-access-hdzkz\") on node \"crc\" DevicePath \"\"" Oct 06 10:18:28 crc kubenswrapper[4824]: I1006 10:18:28.750658 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/27cc4522-3bd4-41bd-9fb7-624c634a5288-kube-api-access-5cbxw" (OuterVolumeSpecName: "kube-api-access-5cbxw") pod "27cc4522-3bd4-41bd-9fb7-624c634a5288" (UID: "27cc4522-3bd4-41bd-9fb7-624c634a5288"). InnerVolumeSpecName "kube-api-access-5cbxw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:18:28 crc kubenswrapper[4824]: I1006 10:18:28.768834 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 06 10:18:28 crc kubenswrapper[4824]: I1006 10:18:28.775879 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27cc4522-3bd4-41bd-9fb7-624c634a5288-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "27cc4522-3bd4-41bd-9fb7-624c634a5288" (UID: "27cc4522-3bd4-41bd-9fb7-624c634a5288"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:18:28 crc kubenswrapper[4824]: I1006 10:18:28.793426 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27cc4522-3bd4-41bd-9fb7-624c634a5288-config-data" (OuterVolumeSpecName: "config-data") pod "27cc4522-3bd4-41bd-9fb7-624c634a5288" (UID: "27cc4522-3bd4-41bd-9fb7-624c634a5288"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:18:28 crc kubenswrapper[4824]: I1006 10:18:28.851906 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/58776586-f2ad-4b0e-bf7e-303324ffe7f2-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"58776586-f2ad-4b0e-bf7e-303324ffe7f2\") " pod="openstack/kube-state-metrics-0" Oct 06 10:18:28 crc kubenswrapper[4824]: I1006 10:18:28.852004 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/58776586-f2ad-4b0e-bf7e-303324ffe7f2-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"58776586-f2ad-4b0e-bf7e-303324ffe7f2\") " pod="openstack/kube-state-metrics-0" Oct 06 10:18:28 crc kubenswrapper[4824]: I1006 10:18:28.852061 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kfsjk\" (UniqueName: \"kubernetes.io/projected/58776586-f2ad-4b0e-bf7e-303324ffe7f2-kube-api-access-kfsjk\") pod \"kube-state-metrics-0\" (UID: \"58776586-f2ad-4b0e-bf7e-303324ffe7f2\") " pod="openstack/kube-state-metrics-0" Oct 06 10:18:28 crc kubenswrapper[4824]: I1006 10:18:28.852094 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58776586-f2ad-4b0e-bf7e-303324ffe7f2-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"58776586-f2ad-4b0e-bf7e-303324ffe7f2\") " pod="openstack/kube-state-metrics-0" Oct 06 10:18:28 crc kubenswrapper[4824]: I1006 10:18:28.852611 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27cc4522-3bd4-41bd-9fb7-624c634a5288-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 10:18:28 crc kubenswrapper[4824]: I1006 10:18:28.852634 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/27cc4522-3bd4-41bd-9fb7-624c634a5288-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 10:18:28 crc kubenswrapper[4824]: I1006 10:18:28.852644 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5cbxw\" (UniqueName: \"kubernetes.io/projected/27cc4522-3bd4-41bd-9fb7-624c634a5288-kube-api-access-5cbxw\") on node \"crc\" DevicePath \"\"" Oct 06 10:18:28 crc kubenswrapper[4824]: I1006 10:18:28.855930 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/58776586-f2ad-4b0e-bf7e-303324ffe7f2-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"58776586-f2ad-4b0e-bf7e-303324ffe7f2\") " pod="openstack/kube-state-metrics-0" Oct 06 10:18:28 crc kubenswrapper[4824]: I1006 10:18:28.856159 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58776586-f2ad-4b0e-bf7e-303324ffe7f2-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"58776586-f2ad-4b0e-bf7e-303324ffe7f2\") " pod="openstack/kube-state-metrics-0" Oct 06 10:18:28 crc kubenswrapper[4824]: I1006 10:18:28.857802 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/58776586-f2ad-4b0e-bf7e-303324ffe7f2-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"58776586-f2ad-4b0e-bf7e-303324ffe7f2\") " pod="openstack/kube-state-metrics-0" Oct 06 10:18:28 crc kubenswrapper[4824]: I1006 10:18:28.872455 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kfsjk\" (UniqueName: \"kubernetes.io/projected/58776586-f2ad-4b0e-bf7e-303324ffe7f2-kube-api-access-kfsjk\") pod \"kube-state-metrics-0\" (UID: \"58776586-f2ad-4b0e-bf7e-303324ffe7f2\") " pod="openstack/kube-state-metrics-0" Oct 06 10:18:28 crc kubenswrapper[4824]: I1006 10:18:28.938214 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 06 10:18:28 crc kubenswrapper[4824]: I1006 10:18:28.973655 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 06 10:18:28 crc kubenswrapper[4824]: I1006 10:18:28.989318 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 06 10:18:28 crc kubenswrapper[4824]: I1006 10:18:28.992055 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 06 10:18:28 crc kubenswrapper[4824]: I1006 10:18:28.995261 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 06 10:18:29 crc kubenswrapper[4824]: I1006 10:18:29.000674 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 06 10:18:29 crc kubenswrapper[4824]: I1006 10:18:29.057894 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ebdd4e3c-8159-4542-b348-109f958f9a5f-config-data\") pod \"nova-api-0\" (UID: \"ebdd4e3c-8159-4542-b348-109f958f9a5f\") " pod="openstack/nova-api-0" Oct 06 10:18:29 crc kubenswrapper[4824]: I1006 10:18:29.058450 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q4gpc\" (UniqueName: \"kubernetes.io/projected/ebdd4e3c-8159-4542-b348-109f958f9a5f-kube-api-access-q4gpc\") pod \"nova-api-0\" (UID: \"ebdd4e3c-8159-4542-b348-109f958f9a5f\") " pod="openstack/nova-api-0" Oct 06 10:18:29 crc kubenswrapper[4824]: I1006 10:18:29.058608 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ebdd4e3c-8159-4542-b348-109f958f9a5f-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"ebdd4e3c-8159-4542-b348-109f958f9a5f\") " pod="openstack/nova-api-0" Oct 06 10:18:29 crc kubenswrapper[4824]: I1006 10:18:29.058667 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ebdd4e3c-8159-4542-b348-109f958f9a5f-logs\") pod \"nova-api-0\" (UID: \"ebdd4e3c-8159-4542-b348-109f958f9a5f\") " pod="openstack/nova-api-0" Oct 06 10:18:29 crc kubenswrapper[4824]: I1006 10:18:29.068245 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 06 10:18:29 crc kubenswrapper[4824]: I1006 10:18:29.160465 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q4gpc\" (UniqueName: \"kubernetes.io/projected/ebdd4e3c-8159-4542-b348-109f958f9a5f-kube-api-access-q4gpc\") pod \"nova-api-0\" (UID: \"ebdd4e3c-8159-4542-b348-109f958f9a5f\") " pod="openstack/nova-api-0" Oct 06 10:18:29 crc kubenswrapper[4824]: I1006 10:18:29.160533 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ebdd4e3c-8159-4542-b348-109f958f9a5f-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"ebdd4e3c-8159-4542-b348-109f958f9a5f\") " pod="openstack/nova-api-0" Oct 06 10:18:29 crc kubenswrapper[4824]: I1006 10:18:29.160557 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ebdd4e3c-8159-4542-b348-109f958f9a5f-logs\") pod \"nova-api-0\" (UID: \"ebdd4e3c-8159-4542-b348-109f958f9a5f\") " pod="openstack/nova-api-0" Oct 06 10:18:29 crc kubenswrapper[4824]: I1006 10:18:29.161348 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ebdd4e3c-8159-4542-b348-109f958f9a5f-logs\") pod \"nova-api-0\" (UID: \"ebdd4e3c-8159-4542-b348-109f958f9a5f\") " pod="openstack/nova-api-0" Oct 06 10:18:29 crc kubenswrapper[4824]: I1006 10:18:29.161619 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ebdd4e3c-8159-4542-b348-109f958f9a5f-config-data\") pod \"nova-api-0\" (UID: \"ebdd4e3c-8159-4542-b348-109f958f9a5f\") " pod="openstack/nova-api-0" Oct 06 10:18:29 crc kubenswrapper[4824]: I1006 10:18:29.166748 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ebdd4e3c-8159-4542-b348-109f958f9a5f-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"ebdd4e3c-8159-4542-b348-109f958f9a5f\") " pod="openstack/nova-api-0" Oct 06 10:18:29 crc kubenswrapper[4824]: I1006 10:18:29.167574 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ebdd4e3c-8159-4542-b348-109f958f9a5f-config-data\") pod \"nova-api-0\" (UID: \"ebdd4e3c-8159-4542-b348-109f958f9a5f\") " pod="openstack/nova-api-0" Oct 06 10:18:29 crc kubenswrapper[4824]: I1006 10:18:29.180417 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q4gpc\" (UniqueName: \"kubernetes.io/projected/ebdd4e3c-8159-4542-b348-109f958f9a5f-kube-api-access-q4gpc\") pod \"nova-api-0\" (UID: \"ebdd4e3c-8159-4542-b348-109f958f9a5f\") " pod="openstack/nova-api-0" Oct 06 10:18:29 crc kubenswrapper[4824]: I1006 10:18:29.295927 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d365fcf-17fb-411e-ab2c-3c5dc8384b21" path="/var/lib/kubelet/pods/1d365fcf-17fb-411e-ab2c-3c5dc8384b21/volumes" Oct 06 10:18:29 crc kubenswrapper[4824]: I1006 10:18:29.297067 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d56c5ee8-59b8-45d6-9078-29423490982f" path="/var/lib/kubelet/pods/d56c5ee8-59b8-45d6-9078-29423490982f/volumes" Oct 06 10:18:29 crc kubenswrapper[4824]: I1006 10:18:29.324757 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 06 10:18:29 crc kubenswrapper[4824]: I1006 10:18:29.432262 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 06 10:18:29 crc kubenswrapper[4824]: I1006 10:18:29.433078 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="539425e3-fd6d-4f71-88b4-80d1a4cf498e" containerName="ceilometer-central-agent" containerID="cri-o://0f485ce06461c003f61fb2c1fb9a1f77664462c0c522974b881b19f5457827f7" gracePeriod=30 Oct 06 10:18:29 crc kubenswrapper[4824]: I1006 10:18:29.433729 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="539425e3-fd6d-4f71-88b4-80d1a4cf498e" containerName="proxy-httpd" containerID="cri-o://6e1bfbb3d87cb6e39782a8b7402577417a374efcb86730702652bf68b748e638" gracePeriod=30 Oct 06 10:18:29 crc kubenswrapper[4824]: I1006 10:18:29.433795 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="539425e3-fd6d-4f71-88b4-80d1a4cf498e" containerName="sg-core" containerID="cri-o://a8e659ed6196e181667d71808ddae81bad11d0ceeba3a2923ec383131f1b2928" gracePeriod=30 Oct 06 10:18:29 crc kubenswrapper[4824]: I1006 10:18:29.433847 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="539425e3-fd6d-4f71-88b4-80d1a4cf498e" containerName="ceilometer-notification-agent" containerID="cri-o://3e8f7319f84c0db5b4c4b61cf8240040ec1a879aa1e37c178e94a242dff3714a" gracePeriod=30 Oct 06 10:18:29 crc kubenswrapper[4824]: I1006 10:18:29.604231 4824 generic.go:334] "Generic (PLEG): container finished" podID="539425e3-fd6d-4f71-88b4-80d1a4cf498e" containerID="a8e659ed6196e181667d71808ddae81bad11d0ceeba3a2923ec383131f1b2928" exitCode=2 Oct 06 10:18:29 crc kubenswrapper[4824]: I1006 10:18:29.604309 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"539425e3-fd6d-4f71-88b4-80d1a4cf498e","Type":"ContainerDied","Data":"a8e659ed6196e181667d71808ddae81bad11d0ceeba3a2923ec383131f1b2928"} Oct 06 10:18:29 crc kubenswrapper[4824]: I1006 10:18:29.604389 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 06 10:18:29 crc kubenswrapper[4824]: I1006 10:18:29.653372 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 06 10:18:29 crc kubenswrapper[4824]: I1006 10:18:29.666271 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 06 10:18:29 crc kubenswrapper[4824]: W1006 10:18:29.675683 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod58776586_f2ad_4b0e_bf7e_303324ffe7f2.slice/crio-85acec5f95ad67329427e840e85cc114ee4df3a88f34e91c49fc05f72957f6ec WatchSource:0}: Error finding container 85acec5f95ad67329427e840e85cc114ee4df3a88f34e91c49fc05f72957f6ec: Status 404 returned error can't find the container with id 85acec5f95ad67329427e840e85cc114ee4df3a88f34e91c49fc05f72957f6ec Oct 06 10:18:29 crc kubenswrapper[4824]: I1006 10:18:29.675799 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 06 10:18:29 crc kubenswrapper[4824]: I1006 10:18:29.686851 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 06 10:18:29 crc kubenswrapper[4824]: I1006 10:18:29.690548 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 06 10:18:29 crc kubenswrapper[4824]: I1006 10:18:29.695025 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 06 10:18:29 crc kubenswrapper[4824]: I1006 10:18:29.697929 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 06 10:18:29 crc kubenswrapper[4824]: I1006 10:18:29.701192 4824 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 06 10:18:29 crc kubenswrapper[4824]: I1006 10:18:29.879772 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mkf5f\" (UniqueName: \"kubernetes.io/projected/8958352a-0551-4af4-8e6b-6655ea3490f6-kube-api-access-mkf5f\") pod \"nova-scheduler-0\" (UID: \"8958352a-0551-4af4-8e6b-6655ea3490f6\") " pod="openstack/nova-scheduler-0" Oct 06 10:18:29 crc kubenswrapper[4824]: I1006 10:18:29.880139 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8958352a-0551-4af4-8e6b-6655ea3490f6-config-data\") pod \"nova-scheduler-0\" (UID: \"8958352a-0551-4af4-8e6b-6655ea3490f6\") " pod="openstack/nova-scheduler-0" Oct 06 10:18:29 crc kubenswrapper[4824]: I1006 10:18:29.880302 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8958352a-0551-4af4-8e6b-6655ea3490f6-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"8958352a-0551-4af4-8e6b-6655ea3490f6\") " pod="openstack/nova-scheduler-0" Oct 06 10:18:29 crc kubenswrapper[4824]: I1006 10:18:29.882821 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 06 10:18:29 crc kubenswrapper[4824]: W1006 10:18:29.890785 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podebdd4e3c_8159_4542_b348_109f958f9a5f.slice/crio-fbc6ee08dd6e8ac29a9ac95f2d4f31b70f3204164bc091e2d021368ba9a97f59 WatchSource:0}: Error finding container fbc6ee08dd6e8ac29a9ac95f2d4f31b70f3204164bc091e2d021368ba9a97f59: Status 404 returned error can't find the container with id fbc6ee08dd6e8ac29a9ac95f2d4f31b70f3204164bc091e2d021368ba9a97f59 Oct 06 10:18:29 crc kubenswrapper[4824]: I1006 10:18:29.981970 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mkf5f\" (UniqueName: \"kubernetes.io/projected/8958352a-0551-4af4-8e6b-6655ea3490f6-kube-api-access-mkf5f\") pod \"nova-scheduler-0\" (UID: \"8958352a-0551-4af4-8e6b-6655ea3490f6\") " pod="openstack/nova-scheduler-0" Oct 06 10:18:29 crc kubenswrapper[4824]: I1006 10:18:29.982132 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8958352a-0551-4af4-8e6b-6655ea3490f6-config-data\") pod \"nova-scheduler-0\" (UID: \"8958352a-0551-4af4-8e6b-6655ea3490f6\") " pod="openstack/nova-scheduler-0" Oct 06 10:18:29 crc kubenswrapper[4824]: I1006 10:18:29.982207 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8958352a-0551-4af4-8e6b-6655ea3490f6-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"8958352a-0551-4af4-8e6b-6655ea3490f6\") " pod="openstack/nova-scheduler-0" Oct 06 10:18:29 crc kubenswrapper[4824]: I1006 10:18:29.992854 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8958352a-0551-4af4-8e6b-6655ea3490f6-config-data\") pod \"nova-scheduler-0\" (UID: \"8958352a-0551-4af4-8e6b-6655ea3490f6\") " pod="openstack/nova-scheduler-0" Oct 06 10:18:29 crc kubenswrapper[4824]: I1006 10:18:29.993032 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8958352a-0551-4af4-8e6b-6655ea3490f6-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"8958352a-0551-4af4-8e6b-6655ea3490f6\") " pod="openstack/nova-scheduler-0" Oct 06 10:18:30 crc kubenswrapper[4824]: I1006 10:18:30.000936 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mkf5f\" (UniqueName: \"kubernetes.io/projected/8958352a-0551-4af4-8e6b-6655ea3490f6-kube-api-access-mkf5f\") pod \"nova-scheduler-0\" (UID: \"8958352a-0551-4af4-8e6b-6655ea3490f6\") " pod="openstack/nova-scheduler-0" Oct 06 10:18:30 crc kubenswrapper[4824]: I1006 10:18:30.020744 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 06 10:18:30 crc kubenswrapper[4824]: I1006 10:18:30.534611 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 06 10:18:30 crc kubenswrapper[4824]: W1006 10:18:30.549331 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8958352a_0551_4af4_8e6b_6655ea3490f6.slice/crio-263512d78658387a78e68953f51ba1fd17ebaecd6d3472c8a88a52845378388e WatchSource:0}: Error finding container 263512d78658387a78e68953f51ba1fd17ebaecd6d3472c8a88a52845378388e: Status 404 returned error can't find the container with id 263512d78658387a78e68953f51ba1fd17ebaecd6d3472c8a88a52845378388e Oct 06 10:18:30 crc kubenswrapper[4824]: I1006 10:18:30.627226 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"58776586-f2ad-4b0e-bf7e-303324ffe7f2","Type":"ContainerStarted","Data":"afffaf3b285539bfe4f281c10fd18bfd8570d5ebc52d25be6b641c7de8451d3f"} Oct 06 10:18:30 crc kubenswrapper[4824]: I1006 10:18:30.627756 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"58776586-f2ad-4b0e-bf7e-303324ffe7f2","Type":"ContainerStarted","Data":"85acec5f95ad67329427e840e85cc114ee4df3a88f34e91c49fc05f72957f6ec"} Oct 06 10:18:30 crc kubenswrapper[4824]: I1006 10:18:30.627780 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Oct 06 10:18:30 crc kubenswrapper[4824]: I1006 10:18:30.632613 4824 generic.go:334] "Generic (PLEG): container finished" podID="539425e3-fd6d-4f71-88b4-80d1a4cf498e" containerID="6e1bfbb3d87cb6e39782a8b7402577417a374efcb86730702652bf68b748e638" exitCode=0 Oct 06 10:18:30 crc kubenswrapper[4824]: I1006 10:18:30.632655 4824 generic.go:334] "Generic (PLEG): container finished" podID="539425e3-fd6d-4f71-88b4-80d1a4cf498e" containerID="0f485ce06461c003f61fb2c1fb9a1f77664462c0c522974b881b19f5457827f7" exitCode=0 Oct 06 10:18:30 crc kubenswrapper[4824]: I1006 10:18:30.632716 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"539425e3-fd6d-4f71-88b4-80d1a4cf498e","Type":"ContainerDied","Data":"6e1bfbb3d87cb6e39782a8b7402577417a374efcb86730702652bf68b748e638"} Oct 06 10:18:30 crc kubenswrapper[4824]: I1006 10:18:30.632752 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"539425e3-fd6d-4f71-88b4-80d1a4cf498e","Type":"ContainerDied","Data":"0f485ce06461c003f61fb2c1fb9a1f77664462c0c522974b881b19f5457827f7"} Oct 06 10:18:30 crc kubenswrapper[4824]: I1006 10:18:30.635864 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ebdd4e3c-8159-4542-b348-109f958f9a5f","Type":"ContainerStarted","Data":"00f934bf252754731c575469bc520b850aa42c980a66ac9f12ed730b0da97dbb"} Oct 06 10:18:30 crc kubenswrapper[4824]: I1006 10:18:30.635907 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ebdd4e3c-8159-4542-b348-109f958f9a5f","Type":"ContainerStarted","Data":"63be74dc92bd5218a7077af368d5c701873e4e262c7ead1808da0dd6d507e839"} Oct 06 10:18:30 crc kubenswrapper[4824]: I1006 10:18:30.635922 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ebdd4e3c-8159-4542-b348-109f958f9a5f","Type":"ContainerStarted","Data":"fbc6ee08dd6e8ac29a9ac95f2d4f31b70f3204164bc091e2d021368ba9a97f59"} Oct 06 10:18:30 crc kubenswrapper[4824]: I1006 10:18:30.638752 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"8958352a-0551-4af4-8e6b-6655ea3490f6","Type":"ContainerStarted","Data":"263512d78658387a78e68953f51ba1fd17ebaecd6d3472c8a88a52845378388e"} Oct 06 10:18:30 crc kubenswrapper[4824]: I1006 10:18:30.652075 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.237005257 podStartE2EDuration="2.652051513s" podCreationTimestamp="2025-10-06 10:18:28 +0000 UTC" firstStartedPulling="2025-10-06 10:18:29.700794183 +0000 UTC m=+1219.065217044" lastFinishedPulling="2025-10-06 10:18:30.115840429 +0000 UTC m=+1219.480263300" observedRunningTime="2025-10-06 10:18:30.64542115 +0000 UTC m=+1220.009844021" watchObservedRunningTime="2025-10-06 10:18:30.652051513 +0000 UTC m=+1220.016474374" Oct 06 10:18:30 crc kubenswrapper[4824]: I1006 10:18:30.679383 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.679351789 podStartE2EDuration="2.679351789s" podCreationTimestamp="2025-10-06 10:18:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:18:30.678121199 +0000 UTC m=+1220.042544060" watchObservedRunningTime="2025-10-06 10:18:30.679351789 +0000 UTC m=+1220.043774650" Oct 06 10:18:30 crc kubenswrapper[4824]: I1006 10:18:30.989488 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 06 10:18:30 crc kubenswrapper[4824]: I1006 10:18:30.989547 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 06 10:18:31 crc kubenswrapper[4824]: I1006 10:18:31.290183 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="27cc4522-3bd4-41bd-9fb7-624c634a5288" path="/var/lib/kubelet/pods/27cc4522-3bd4-41bd-9fb7-624c634a5288/volumes" Oct 06 10:18:31 crc kubenswrapper[4824]: I1006 10:18:31.653217 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"8958352a-0551-4af4-8e6b-6655ea3490f6","Type":"ContainerStarted","Data":"7933fab2e0964f7f5e7bfb0ad7a3d55be232f3319edbbfd02c0bebc60abc9188"} Oct 06 10:18:31 crc kubenswrapper[4824]: I1006 10:18:31.679828 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.679811807 podStartE2EDuration="2.679811807s" podCreationTimestamp="2025-10-06 10:18:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:18:31.677614703 +0000 UTC m=+1221.042037564" watchObservedRunningTime="2025-10-06 10:18:31.679811807 +0000 UTC m=+1221.044234668" Oct 06 10:18:33 crc kubenswrapper[4824]: I1006 10:18:33.907171 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Oct 06 10:18:35 crc kubenswrapper[4824]: I1006 10:18:35.021327 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 06 10:18:35 crc kubenswrapper[4824]: I1006 10:18:35.990203 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 06 10:18:35 crc kubenswrapper[4824]: I1006 10:18:35.990628 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 06 10:18:36 crc kubenswrapper[4824]: I1006 10:18:36.730099 4824 generic.go:334] "Generic (PLEG): container finished" podID="539425e3-fd6d-4f71-88b4-80d1a4cf498e" containerID="3e8f7319f84c0db5b4c4b61cf8240040ec1a879aa1e37c178e94a242dff3714a" exitCode=0 Oct 06 10:18:36 crc kubenswrapper[4824]: I1006 10:18:36.730146 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"539425e3-fd6d-4f71-88b4-80d1a4cf498e","Type":"ContainerDied","Data":"3e8f7319f84c0db5b4c4b61cf8240040ec1a879aa1e37c178e94a242dff3714a"} Oct 06 10:18:37 crc kubenswrapper[4824]: I1006 10:18:37.015329 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="38f30d16-bb2b-4a0b-aac6-64078e032eeb" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.196:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 06 10:18:37 crc kubenswrapper[4824]: I1006 10:18:37.015327 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="38f30d16-bb2b-4a0b-aac6-64078e032eeb" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.196:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 06 10:18:37 crc kubenswrapper[4824]: I1006 10:18:37.235875 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 06 10:18:37 crc kubenswrapper[4824]: I1006 10:18:37.350751 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/539425e3-fd6d-4f71-88b4-80d1a4cf498e-run-httpd\") pod \"539425e3-fd6d-4f71-88b4-80d1a4cf498e\" (UID: \"539425e3-fd6d-4f71-88b4-80d1a4cf498e\") " Oct 06 10:18:37 crc kubenswrapper[4824]: I1006 10:18:37.350833 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/539425e3-fd6d-4f71-88b4-80d1a4cf498e-combined-ca-bundle\") pod \"539425e3-fd6d-4f71-88b4-80d1a4cf498e\" (UID: \"539425e3-fd6d-4f71-88b4-80d1a4cf498e\") " Oct 06 10:18:37 crc kubenswrapper[4824]: I1006 10:18:37.351095 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/539425e3-fd6d-4f71-88b4-80d1a4cf498e-sg-core-conf-yaml\") pod \"539425e3-fd6d-4f71-88b4-80d1a4cf498e\" (UID: \"539425e3-fd6d-4f71-88b4-80d1a4cf498e\") " Oct 06 10:18:37 crc kubenswrapper[4824]: I1006 10:18:37.351301 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/539425e3-fd6d-4f71-88b4-80d1a4cf498e-config-data\") pod \"539425e3-fd6d-4f71-88b4-80d1a4cf498e\" (UID: \"539425e3-fd6d-4f71-88b4-80d1a4cf498e\") " Oct 06 10:18:37 crc kubenswrapper[4824]: I1006 10:18:37.351329 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/539425e3-fd6d-4f71-88b4-80d1a4cf498e-log-httpd\") pod \"539425e3-fd6d-4f71-88b4-80d1a4cf498e\" (UID: \"539425e3-fd6d-4f71-88b4-80d1a4cf498e\") " Oct 06 10:18:37 crc kubenswrapper[4824]: I1006 10:18:37.351401 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/539425e3-fd6d-4f71-88b4-80d1a4cf498e-scripts\") pod \"539425e3-fd6d-4f71-88b4-80d1a4cf498e\" (UID: \"539425e3-fd6d-4f71-88b4-80d1a4cf498e\") " Oct 06 10:18:37 crc kubenswrapper[4824]: I1006 10:18:37.351478 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6c62r\" (UniqueName: \"kubernetes.io/projected/539425e3-fd6d-4f71-88b4-80d1a4cf498e-kube-api-access-6c62r\") pod \"539425e3-fd6d-4f71-88b4-80d1a4cf498e\" (UID: \"539425e3-fd6d-4f71-88b4-80d1a4cf498e\") " Oct 06 10:18:37 crc kubenswrapper[4824]: I1006 10:18:37.351547 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/539425e3-fd6d-4f71-88b4-80d1a4cf498e-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "539425e3-fd6d-4f71-88b4-80d1a4cf498e" (UID: "539425e3-fd6d-4f71-88b4-80d1a4cf498e"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:18:37 crc kubenswrapper[4824]: I1006 10:18:37.351841 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/539425e3-fd6d-4f71-88b4-80d1a4cf498e-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "539425e3-fd6d-4f71-88b4-80d1a4cf498e" (UID: "539425e3-fd6d-4f71-88b4-80d1a4cf498e"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:18:37 crc kubenswrapper[4824]: I1006 10:18:37.352074 4824 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/539425e3-fd6d-4f71-88b4-80d1a4cf498e-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 06 10:18:37 crc kubenswrapper[4824]: I1006 10:18:37.352094 4824 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/539425e3-fd6d-4f71-88b4-80d1a4cf498e-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 06 10:18:37 crc kubenswrapper[4824]: I1006 10:18:37.357962 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/539425e3-fd6d-4f71-88b4-80d1a4cf498e-scripts" (OuterVolumeSpecName: "scripts") pod "539425e3-fd6d-4f71-88b4-80d1a4cf498e" (UID: "539425e3-fd6d-4f71-88b4-80d1a4cf498e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:18:37 crc kubenswrapper[4824]: I1006 10:18:37.375878 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/539425e3-fd6d-4f71-88b4-80d1a4cf498e-kube-api-access-6c62r" (OuterVolumeSpecName: "kube-api-access-6c62r") pod "539425e3-fd6d-4f71-88b4-80d1a4cf498e" (UID: "539425e3-fd6d-4f71-88b4-80d1a4cf498e"). InnerVolumeSpecName "kube-api-access-6c62r". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:18:37 crc kubenswrapper[4824]: I1006 10:18:37.394568 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/539425e3-fd6d-4f71-88b4-80d1a4cf498e-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "539425e3-fd6d-4f71-88b4-80d1a4cf498e" (UID: "539425e3-fd6d-4f71-88b4-80d1a4cf498e"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:18:37 crc kubenswrapper[4824]: I1006 10:18:37.443747 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/539425e3-fd6d-4f71-88b4-80d1a4cf498e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "539425e3-fd6d-4f71-88b4-80d1a4cf498e" (UID: "539425e3-fd6d-4f71-88b4-80d1a4cf498e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:18:37 crc kubenswrapper[4824]: I1006 10:18:37.454782 4824 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/539425e3-fd6d-4f71-88b4-80d1a4cf498e-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 06 10:18:37 crc kubenswrapper[4824]: I1006 10:18:37.454824 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/539425e3-fd6d-4f71-88b4-80d1a4cf498e-scripts\") on node \"crc\" DevicePath \"\"" Oct 06 10:18:37 crc kubenswrapper[4824]: I1006 10:18:37.454857 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6c62r\" (UniqueName: \"kubernetes.io/projected/539425e3-fd6d-4f71-88b4-80d1a4cf498e-kube-api-access-6c62r\") on node \"crc\" DevicePath \"\"" Oct 06 10:18:37 crc kubenswrapper[4824]: I1006 10:18:37.454872 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/539425e3-fd6d-4f71-88b4-80d1a4cf498e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 10:18:37 crc kubenswrapper[4824]: I1006 10:18:37.479786 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/539425e3-fd6d-4f71-88b4-80d1a4cf498e-config-data" (OuterVolumeSpecName: "config-data") pod "539425e3-fd6d-4f71-88b4-80d1a4cf498e" (UID: "539425e3-fd6d-4f71-88b4-80d1a4cf498e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:18:37 crc kubenswrapper[4824]: I1006 10:18:37.557024 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/539425e3-fd6d-4f71-88b4-80d1a4cf498e-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 10:18:37 crc kubenswrapper[4824]: I1006 10:18:37.746384 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"539425e3-fd6d-4f71-88b4-80d1a4cf498e","Type":"ContainerDied","Data":"e333ef4722e0b96c0abad57cd34ea3fa7c2674dcb5c6260623b43eb916bef206"} Oct 06 10:18:37 crc kubenswrapper[4824]: I1006 10:18:37.746453 4824 scope.go:117] "RemoveContainer" containerID="6e1bfbb3d87cb6e39782a8b7402577417a374efcb86730702652bf68b748e638" Oct 06 10:18:37 crc kubenswrapper[4824]: I1006 10:18:37.746607 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 06 10:18:37 crc kubenswrapper[4824]: I1006 10:18:37.799742 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 06 10:18:37 crc kubenswrapper[4824]: I1006 10:18:37.818213 4824 scope.go:117] "RemoveContainer" containerID="a8e659ed6196e181667d71808ddae81bad11d0ceeba3a2923ec383131f1b2928" Oct 06 10:18:37 crc kubenswrapper[4824]: I1006 10:18:37.829488 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 06 10:18:37 crc kubenswrapper[4824]: I1006 10:18:37.849447 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 06 10:18:37 crc kubenswrapper[4824]: E1006 10:18:37.850121 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="539425e3-fd6d-4f71-88b4-80d1a4cf498e" containerName="proxy-httpd" Oct 06 10:18:37 crc kubenswrapper[4824]: I1006 10:18:37.850154 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="539425e3-fd6d-4f71-88b4-80d1a4cf498e" containerName="proxy-httpd" Oct 06 10:18:37 crc kubenswrapper[4824]: E1006 10:18:37.850186 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="539425e3-fd6d-4f71-88b4-80d1a4cf498e" containerName="ceilometer-notification-agent" Oct 06 10:18:37 crc kubenswrapper[4824]: I1006 10:18:37.850198 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="539425e3-fd6d-4f71-88b4-80d1a4cf498e" containerName="ceilometer-notification-agent" Oct 06 10:18:37 crc kubenswrapper[4824]: E1006 10:18:37.850253 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="539425e3-fd6d-4f71-88b4-80d1a4cf498e" containerName="sg-core" Oct 06 10:18:37 crc kubenswrapper[4824]: I1006 10:18:37.850266 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="539425e3-fd6d-4f71-88b4-80d1a4cf498e" containerName="sg-core" Oct 06 10:18:37 crc kubenswrapper[4824]: E1006 10:18:37.850287 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="539425e3-fd6d-4f71-88b4-80d1a4cf498e" containerName="ceilometer-central-agent" Oct 06 10:18:37 crc kubenswrapper[4824]: I1006 10:18:37.850300 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="539425e3-fd6d-4f71-88b4-80d1a4cf498e" containerName="ceilometer-central-agent" Oct 06 10:18:37 crc kubenswrapper[4824]: I1006 10:18:37.850591 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="539425e3-fd6d-4f71-88b4-80d1a4cf498e" containerName="sg-core" Oct 06 10:18:37 crc kubenswrapper[4824]: I1006 10:18:37.850632 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="539425e3-fd6d-4f71-88b4-80d1a4cf498e" containerName="proxy-httpd" Oct 06 10:18:37 crc kubenswrapper[4824]: I1006 10:18:37.850653 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="539425e3-fd6d-4f71-88b4-80d1a4cf498e" containerName="ceilometer-notification-agent" Oct 06 10:18:37 crc kubenswrapper[4824]: I1006 10:18:37.850678 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="539425e3-fd6d-4f71-88b4-80d1a4cf498e" containerName="ceilometer-central-agent" Oct 06 10:18:37 crc kubenswrapper[4824]: I1006 10:18:37.853758 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 06 10:18:37 crc kubenswrapper[4824]: I1006 10:18:37.857090 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 06 10:18:37 crc kubenswrapper[4824]: I1006 10:18:37.857132 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 06 10:18:37 crc kubenswrapper[4824]: I1006 10:18:37.859539 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 06 10:18:37 crc kubenswrapper[4824]: I1006 10:18:37.879398 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 06 10:18:37 crc kubenswrapper[4824]: I1006 10:18:37.890229 4824 scope.go:117] "RemoveContainer" containerID="3e8f7319f84c0db5b4c4b61cf8240040ec1a879aa1e37c178e94a242dff3714a" Oct 06 10:18:37 crc kubenswrapper[4824]: I1006 10:18:37.918601 4824 scope.go:117] "RemoveContainer" containerID="0f485ce06461c003f61fb2c1fb9a1f77664462c0c522974b881b19f5457827f7" Oct 06 10:18:37 crc kubenswrapper[4824]: I1006 10:18:37.966460 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa150ff6-e61e-4deb-9a72-98fbcc8389b8-config-data\") pod \"ceilometer-0\" (UID: \"aa150ff6-e61e-4deb-9a72-98fbcc8389b8\") " pod="openstack/ceilometer-0" Oct 06 10:18:37 crc kubenswrapper[4824]: I1006 10:18:37.966536 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c9f8x\" (UniqueName: \"kubernetes.io/projected/aa150ff6-e61e-4deb-9a72-98fbcc8389b8-kube-api-access-c9f8x\") pod \"ceilometer-0\" (UID: \"aa150ff6-e61e-4deb-9a72-98fbcc8389b8\") " pod="openstack/ceilometer-0" Oct 06 10:18:37 crc kubenswrapper[4824]: I1006 10:18:37.966736 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aa150ff6-e61e-4deb-9a72-98fbcc8389b8-run-httpd\") pod \"ceilometer-0\" (UID: \"aa150ff6-e61e-4deb-9a72-98fbcc8389b8\") " pod="openstack/ceilometer-0" Oct 06 10:18:37 crc kubenswrapper[4824]: I1006 10:18:37.966876 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/aa150ff6-e61e-4deb-9a72-98fbcc8389b8-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"aa150ff6-e61e-4deb-9a72-98fbcc8389b8\") " pod="openstack/ceilometer-0" Oct 06 10:18:37 crc kubenswrapper[4824]: I1006 10:18:37.966930 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aa150ff6-e61e-4deb-9a72-98fbcc8389b8-scripts\") pod \"ceilometer-0\" (UID: \"aa150ff6-e61e-4deb-9a72-98fbcc8389b8\") " pod="openstack/ceilometer-0" Oct 06 10:18:37 crc kubenswrapper[4824]: I1006 10:18:37.966974 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa150ff6-e61e-4deb-9a72-98fbcc8389b8-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"aa150ff6-e61e-4deb-9a72-98fbcc8389b8\") " pod="openstack/ceilometer-0" Oct 06 10:18:37 crc kubenswrapper[4824]: I1006 10:18:37.967036 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/aa150ff6-e61e-4deb-9a72-98fbcc8389b8-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"aa150ff6-e61e-4deb-9a72-98fbcc8389b8\") " pod="openstack/ceilometer-0" Oct 06 10:18:37 crc kubenswrapper[4824]: I1006 10:18:37.967200 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aa150ff6-e61e-4deb-9a72-98fbcc8389b8-log-httpd\") pod \"ceilometer-0\" (UID: \"aa150ff6-e61e-4deb-9a72-98fbcc8389b8\") " pod="openstack/ceilometer-0" Oct 06 10:18:38 crc kubenswrapper[4824]: I1006 10:18:38.069340 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c9f8x\" (UniqueName: \"kubernetes.io/projected/aa150ff6-e61e-4deb-9a72-98fbcc8389b8-kube-api-access-c9f8x\") pod \"ceilometer-0\" (UID: \"aa150ff6-e61e-4deb-9a72-98fbcc8389b8\") " pod="openstack/ceilometer-0" Oct 06 10:18:38 crc kubenswrapper[4824]: I1006 10:18:38.069470 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aa150ff6-e61e-4deb-9a72-98fbcc8389b8-run-httpd\") pod \"ceilometer-0\" (UID: \"aa150ff6-e61e-4deb-9a72-98fbcc8389b8\") " pod="openstack/ceilometer-0" Oct 06 10:18:38 crc kubenswrapper[4824]: I1006 10:18:38.070121 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aa150ff6-e61e-4deb-9a72-98fbcc8389b8-run-httpd\") pod \"ceilometer-0\" (UID: \"aa150ff6-e61e-4deb-9a72-98fbcc8389b8\") " pod="openstack/ceilometer-0" Oct 06 10:18:38 crc kubenswrapper[4824]: I1006 10:18:38.070129 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/aa150ff6-e61e-4deb-9a72-98fbcc8389b8-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"aa150ff6-e61e-4deb-9a72-98fbcc8389b8\") " pod="openstack/ceilometer-0" Oct 06 10:18:38 crc kubenswrapper[4824]: I1006 10:18:38.070232 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aa150ff6-e61e-4deb-9a72-98fbcc8389b8-scripts\") pod \"ceilometer-0\" (UID: \"aa150ff6-e61e-4deb-9a72-98fbcc8389b8\") " pod="openstack/ceilometer-0" Oct 06 10:18:38 crc kubenswrapper[4824]: I1006 10:18:38.070270 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa150ff6-e61e-4deb-9a72-98fbcc8389b8-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"aa150ff6-e61e-4deb-9a72-98fbcc8389b8\") " pod="openstack/ceilometer-0" Oct 06 10:18:38 crc kubenswrapper[4824]: I1006 10:18:38.070295 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/aa150ff6-e61e-4deb-9a72-98fbcc8389b8-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"aa150ff6-e61e-4deb-9a72-98fbcc8389b8\") " pod="openstack/ceilometer-0" Oct 06 10:18:38 crc kubenswrapper[4824]: I1006 10:18:38.070442 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aa150ff6-e61e-4deb-9a72-98fbcc8389b8-log-httpd\") pod \"ceilometer-0\" (UID: \"aa150ff6-e61e-4deb-9a72-98fbcc8389b8\") " pod="openstack/ceilometer-0" Oct 06 10:18:38 crc kubenswrapper[4824]: I1006 10:18:38.070528 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa150ff6-e61e-4deb-9a72-98fbcc8389b8-config-data\") pod \"ceilometer-0\" (UID: \"aa150ff6-e61e-4deb-9a72-98fbcc8389b8\") " pod="openstack/ceilometer-0" Oct 06 10:18:38 crc kubenswrapper[4824]: I1006 10:18:38.071432 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aa150ff6-e61e-4deb-9a72-98fbcc8389b8-log-httpd\") pod \"ceilometer-0\" (UID: \"aa150ff6-e61e-4deb-9a72-98fbcc8389b8\") " pod="openstack/ceilometer-0" Oct 06 10:18:38 crc kubenswrapper[4824]: I1006 10:18:38.074389 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa150ff6-e61e-4deb-9a72-98fbcc8389b8-config-data\") pod \"ceilometer-0\" (UID: \"aa150ff6-e61e-4deb-9a72-98fbcc8389b8\") " pod="openstack/ceilometer-0" Oct 06 10:18:38 crc kubenswrapper[4824]: I1006 10:18:38.075156 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aa150ff6-e61e-4deb-9a72-98fbcc8389b8-scripts\") pod \"ceilometer-0\" (UID: \"aa150ff6-e61e-4deb-9a72-98fbcc8389b8\") " pod="openstack/ceilometer-0" Oct 06 10:18:38 crc kubenswrapper[4824]: I1006 10:18:38.076523 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa150ff6-e61e-4deb-9a72-98fbcc8389b8-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"aa150ff6-e61e-4deb-9a72-98fbcc8389b8\") " pod="openstack/ceilometer-0" Oct 06 10:18:38 crc kubenswrapper[4824]: I1006 10:18:38.085472 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/aa150ff6-e61e-4deb-9a72-98fbcc8389b8-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"aa150ff6-e61e-4deb-9a72-98fbcc8389b8\") " pod="openstack/ceilometer-0" Oct 06 10:18:38 crc kubenswrapper[4824]: I1006 10:18:38.085694 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c9f8x\" (UniqueName: \"kubernetes.io/projected/aa150ff6-e61e-4deb-9a72-98fbcc8389b8-kube-api-access-c9f8x\") pod \"ceilometer-0\" (UID: \"aa150ff6-e61e-4deb-9a72-98fbcc8389b8\") " pod="openstack/ceilometer-0" Oct 06 10:18:38 crc kubenswrapper[4824]: I1006 10:18:38.093156 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/aa150ff6-e61e-4deb-9a72-98fbcc8389b8-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"aa150ff6-e61e-4deb-9a72-98fbcc8389b8\") " pod="openstack/ceilometer-0" Oct 06 10:18:38 crc kubenswrapper[4824]: I1006 10:18:38.191667 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 06 10:18:38 crc kubenswrapper[4824]: I1006 10:18:38.710605 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 06 10:18:38 crc kubenswrapper[4824]: W1006 10:18:38.714706 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaa150ff6_e61e_4deb_9a72_98fbcc8389b8.slice/crio-082f4345738ef5d6215dfe58811ff8ec41b0a626ffeaaf0324e54c2f03b11902 WatchSource:0}: Error finding container 082f4345738ef5d6215dfe58811ff8ec41b0a626ffeaaf0324e54c2f03b11902: Status 404 returned error can't find the container with id 082f4345738ef5d6215dfe58811ff8ec41b0a626ffeaaf0324e54c2f03b11902 Oct 06 10:18:38 crc kubenswrapper[4824]: I1006 10:18:38.769903 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aa150ff6-e61e-4deb-9a72-98fbcc8389b8","Type":"ContainerStarted","Data":"082f4345738ef5d6215dfe58811ff8ec41b0a626ffeaaf0324e54c2f03b11902"} Oct 06 10:18:39 crc kubenswrapper[4824]: I1006 10:18:39.086304 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Oct 06 10:18:39 crc kubenswrapper[4824]: I1006 10:18:39.285578 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="539425e3-fd6d-4f71-88b4-80d1a4cf498e" path="/var/lib/kubelet/pods/539425e3-fd6d-4f71-88b4-80d1a4cf498e/volumes" Oct 06 10:18:39 crc kubenswrapper[4824]: I1006 10:18:39.325674 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 06 10:18:39 crc kubenswrapper[4824]: I1006 10:18:39.325761 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 06 10:18:40 crc kubenswrapper[4824]: I1006 10:18:40.021464 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 06 10:18:40 crc kubenswrapper[4824]: I1006 10:18:40.048365 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 06 10:18:40 crc kubenswrapper[4824]: I1006 10:18:40.410334 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="ebdd4e3c-8159-4542-b348-109f958f9a5f" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.198:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 06 10:18:40 crc kubenswrapper[4824]: I1006 10:18:40.410372 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="ebdd4e3c-8159-4542-b348-109f958f9a5f" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.198:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 06 10:18:40 crc kubenswrapper[4824]: I1006 10:18:40.806293 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aa150ff6-e61e-4deb-9a72-98fbcc8389b8","Type":"ContainerStarted","Data":"a520bed817b15e3c0e6355c65a738a99f59a361edb219a69a49fbac1f16671a5"} Oct 06 10:18:40 crc kubenswrapper[4824]: I1006 10:18:40.806496 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aa150ff6-e61e-4deb-9a72-98fbcc8389b8","Type":"ContainerStarted","Data":"e168848b51f45d398d90fb4b9ad42f43b4a7d30027bbfd84bb68c55a84747017"} Oct 06 10:18:40 crc kubenswrapper[4824]: I1006 10:18:40.856674 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 06 10:18:41 crc kubenswrapper[4824]: I1006 10:18:41.823929 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aa150ff6-e61e-4deb-9a72-98fbcc8389b8","Type":"ContainerStarted","Data":"a4ea9aec64f9ffadabc0fd0437a43c68ca18deb02a1e6848e8d3efc6bd6855c1"} Oct 06 10:18:42 crc kubenswrapper[4824]: I1006 10:18:42.838365 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aa150ff6-e61e-4deb-9a72-98fbcc8389b8","Type":"ContainerStarted","Data":"81c4f0456c769719e1bf6196c4a5067d84301428c756ca7b242e54aecb3f2bb1"} Oct 06 10:18:42 crc kubenswrapper[4824]: I1006 10:18:42.839196 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 06 10:18:42 crc kubenswrapper[4824]: I1006 10:18:42.864026 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.192503688 podStartE2EDuration="5.863970183s" podCreationTimestamp="2025-10-06 10:18:37 +0000 UTC" firstStartedPulling="2025-10-06 10:18:38.718269763 +0000 UTC m=+1228.082692624" lastFinishedPulling="2025-10-06 10:18:42.389736248 +0000 UTC m=+1231.754159119" observedRunningTime="2025-10-06 10:18:42.861501942 +0000 UTC m=+1232.225924803" watchObservedRunningTime="2025-10-06 10:18:42.863970183 +0000 UTC m=+1232.228393044" Oct 06 10:18:43 crc kubenswrapper[4824]: I1006 10:18:43.915891 4824 patch_prober.go:28] interesting pod/machine-config-daemon-khgzw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 10:18:43 crc kubenswrapper[4824]: I1006 10:18:43.916363 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 10:18:45 crc kubenswrapper[4824]: I1006 10:18:45.997547 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 06 10:18:46 crc kubenswrapper[4824]: I1006 10:18:46.000752 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 06 10:18:46 crc kubenswrapper[4824]: I1006 10:18:46.010030 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 06 10:18:46 crc kubenswrapper[4824]: I1006 10:18:46.747050 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 06 10:18:46 crc kubenswrapper[4824]: I1006 10:18:46.770162 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/627f8bd8-ea33-486d-ac9e-2cc7c0b2ae15-config-data\") pod \"627f8bd8-ea33-486d-ac9e-2cc7c0b2ae15\" (UID: \"627f8bd8-ea33-486d-ac9e-2cc7c0b2ae15\") " Oct 06 10:18:46 crc kubenswrapper[4824]: I1006 10:18:46.770329 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/627f8bd8-ea33-486d-ac9e-2cc7c0b2ae15-combined-ca-bundle\") pod \"627f8bd8-ea33-486d-ac9e-2cc7c0b2ae15\" (UID: \"627f8bd8-ea33-486d-ac9e-2cc7c0b2ae15\") " Oct 06 10:18:46 crc kubenswrapper[4824]: I1006 10:18:46.770545 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cxp5h\" (UniqueName: \"kubernetes.io/projected/627f8bd8-ea33-486d-ac9e-2cc7c0b2ae15-kube-api-access-cxp5h\") pod \"627f8bd8-ea33-486d-ac9e-2cc7c0b2ae15\" (UID: \"627f8bd8-ea33-486d-ac9e-2cc7c0b2ae15\") " Oct 06 10:18:46 crc kubenswrapper[4824]: I1006 10:18:46.793120 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/627f8bd8-ea33-486d-ac9e-2cc7c0b2ae15-kube-api-access-cxp5h" (OuterVolumeSpecName: "kube-api-access-cxp5h") pod "627f8bd8-ea33-486d-ac9e-2cc7c0b2ae15" (UID: "627f8bd8-ea33-486d-ac9e-2cc7c0b2ae15"). InnerVolumeSpecName "kube-api-access-cxp5h". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:18:46 crc kubenswrapper[4824]: I1006 10:18:46.818541 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/627f8bd8-ea33-486d-ac9e-2cc7c0b2ae15-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "627f8bd8-ea33-486d-ac9e-2cc7c0b2ae15" (UID: "627f8bd8-ea33-486d-ac9e-2cc7c0b2ae15"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:18:46 crc kubenswrapper[4824]: I1006 10:18:46.820276 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/627f8bd8-ea33-486d-ac9e-2cc7c0b2ae15-config-data" (OuterVolumeSpecName: "config-data") pod "627f8bd8-ea33-486d-ac9e-2cc7c0b2ae15" (UID: "627f8bd8-ea33-486d-ac9e-2cc7c0b2ae15"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:18:46 crc kubenswrapper[4824]: I1006 10:18:46.874366 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/627f8bd8-ea33-486d-ac9e-2cc7c0b2ae15-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 10:18:46 crc kubenswrapper[4824]: I1006 10:18:46.874411 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/627f8bd8-ea33-486d-ac9e-2cc7c0b2ae15-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 10:18:46 crc kubenswrapper[4824]: I1006 10:18:46.874427 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cxp5h\" (UniqueName: \"kubernetes.io/projected/627f8bd8-ea33-486d-ac9e-2cc7c0b2ae15-kube-api-access-cxp5h\") on node \"crc\" DevicePath \"\"" Oct 06 10:18:46 crc kubenswrapper[4824]: I1006 10:18:46.891106 4824 generic.go:334] "Generic (PLEG): container finished" podID="627f8bd8-ea33-486d-ac9e-2cc7c0b2ae15" containerID="64652f98eb9c1402f8d72a1b0d7d5a5e717088663152bb7f3cf9cce181f45f05" exitCode=137 Oct 06 10:18:46 crc kubenswrapper[4824]: I1006 10:18:46.894220 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 06 10:18:46 crc kubenswrapper[4824]: I1006 10:18:46.897411 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"627f8bd8-ea33-486d-ac9e-2cc7c0b2ae15","Type":"ContainerDied","Data":"64652f98eb9c1402f8d72a1b0d7d5a5e717088663152bb7f3cf9cce181f45f05"} Oct 06 10:18:46 crc kubenswrapper[4824]: I1006 10:18:46.897465 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"627f8bd8-ea33-486d-ac9e-2cc7c0b2ae15","Type":"ContainerDied","Data":"543492853f249ef04194d5b4be5598b2417659e4845101dc492848e8fc0ae67c"} Oct 06 10:18:46 crc kubenswrapper[4824]: I1006 10:18:46.897497 4824 scope.go:117] "RemoveContainer" containerID="64652f98eb9c1402f8d72a1b0d7d5a5e717088663152bb7f3cf9cce181f45f05" Oct 06 10:18:46 crc kubenswrapper[4824]: I1006 10:18:46.904573 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 06 10:18:46 crc kubenswrapper[4824]: I1006 10:18:46.924605 4824 scope.go:117] "RemoveContainer" containerID="64652f98eb9c1402f8d72a1b0d7d5a5e717088663152bb7f3cf9cce181f45f05" Oct 06 10:18:46 crc kubenswrapper[4824]: E1006 10:18:46.925352 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"64652f98eb9c1402f8d72a1b0d7d5a5e717088663152bb7f3cf9cce181f45f05\": container with ID starting with 64652f98eb9c1402f8d72a1b0d7d5a5e717088663152bb7f3cf9cce181f45f05 not found: ID does not exist" containerID="64652f98eb9c1402f8d72a1b0d7d5a5e717088663152bb7f3cf9cce181f45f05" Oct 06 10:18:46 crc kubenswrapper[4824]: I1006 10:18:46.925389 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"64652f98eb9c1402f8d72a1b0d7d5a5e717088663152bb7f3cf9cce181f45f05"} err="failed to get container status \"64652f98eb9c1402f8d72a1b0d7d5a5e717088663152bb7f3cf9cce181f45f05\": rpc error: code = NotFound desc = could not find container \"64652f98eb9c1402f8d72a1b0d7d5a5e717088663152bb7f3cf9cce181f45f05\": container with ID starting with 64652f98eb9c1402f8d72a1b0d7d5a5e717088663152bb7f3cf9cce181f45f05 not found: ID does not exist" Oct 06 10:18:46 crc kubenswrapper[4824]: I1006 10:18:46.965933 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 06 10:18:46 crc kubenswrapper[4824]: I1006 10:18:46.994164 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 06 10:18:47 crc kubenswrapper[4824]: I1006 10:18:47.009912 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 06 10:18:47 crc kubenswrapper[4824]: E1006 10:18:47.010494 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="627f8bd8-ea33-486d-ac9e-2cc7c0b2ae15" containerName="nova-cell1-novncproxy-novncproxy" Oct 06 10:18:47 crc kubenswrapper[4824]: I1006 10:18:47.010513 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="627f8bd8-ea33-486d-ac9e-2cc7c0b2ae15" containerName="nova-cell1-novncproxy-novncproxy" Oct 06 10:18:47 crc kubenswrapper[4824]: I1006 10:18:47.010755 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="627f8bd8-ea33-486d-ac9e-2cc7c0b2ae15" containerName="nova-cell1-novncproxy-novncproxy" Oct 06 10:18:47 crc kubenswrapper[4824]: I1006 10:18:47.011606 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 06 10:18:47 crc kubenswrapper[4824]: I1006 10:18:47.014419 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Oct 06 10:18:47 crc kubenswrapper[4824]: I1006 10:18:47.014732 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Oct 06 10:18:47 crc kubenswrapper[4824]: I1006 10:18:47.014833 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Oct 06 10:18:47 crc kubenswrapper[4824]: I1006 10:18:47.025795 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 06 10:18:47 crc kubenswrapper[4824]: I1006 10:18:47.078778 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db4434eb-50c9-4004-8ff3-752cffaa01b4-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"db4434eb-50c9-4004-8ff3-752cffaa01b4\") " pod="openstack/nova-cell1-novncproxy-0" Oct 06 10:18:47 crc kubenswrapper[4824]: I1006 10:18:47.078911 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/db4434eb-50c9-4004-8ff3-752cffaa01b4-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"db4434eb-50c9-4004-8ff3-752cffaa01b4\") " pod="openstack/nova-cell1-novncproxy-0" Oct 06 10:18:47 crc kubenswrapper[4824]: I1006 10:18:47.079035 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db4434eb-50c9-4004-8ff3-752cffaa01b4-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"db4434eb-50c9-4004-8ff3-752cffaa01b4\") " pod="openstack/nova-cell1-novncproxy-0" Oct 06 10:18:47 crc kubenswrapper[4824]: I1006 10:18:47.079264 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t58db\" (UniqueName: \"kubernetes.io/projected/db4434eb-50c9-4004-8ff3-752cffaa01b4-kube-api-access-t58db\") pod \"nova-cell1-novncproxy-0\" (UID: \"db4434eb-50c9-4004-8ff3-752cffaa01b4\") " pod="openstack/nova-cell1-novncproxy-0" Oct 06 10:18:47 crc kubenswrapper[4824]: I1006 10:18:47.079325 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/db4434eb-50c9-4004-8ff3-752cffaa01b4-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"db4434eb-50c9-4004-8ff3-752cffaa01b4\") " pod="openstack/nova-cell1-novncproxy-0" Oct 06 10:18:47 crc kubenswrapper[4824]: I1006 10:18:47.188374 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db4434eb-50c9-4004-8ff3-752cffaa01b4-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"db4434eb-50c9-4004-8ff3-752cffaa01b4\") " pod="openstack/nova-cell1-novncproxy-0" Oct 06 10:18:47 crc kubenswrapper[4824]: I1006 10:18:47.188571 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/db4434eb-50c9-4004-8ff3-752cffaa01b4-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"db4434eb-50c9-4004-8ff3-752cffaa01b4\") " pod="openstack/nova-cell1-novncproxy-0" Oct 06 10:18:47 crc kubenswrapper[4824]: I1006 10:18:47.188813 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db4434eb-50c9-4004-8ff3-752cffaa01b4-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"db4434eb-50c9-4004-8ff3-752cffaa01b4\") " pod="openstack/nova-cell1-novncproxy-0" Oct 06 10:18:47 crc kubenswrapper[4824]: I1006 10:18:47.189016 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t58db\" (UniqueName: \"kubernetes.io/projected/db4434eb-50c9-4004-8ff3-752cffaa01b4-kube-api-access-t58db\") pod \"nova-cell1-novncproxy-0\" (UID: \"db4434eb-50c9-4004-8ff3-752cffaa01b4\") " pod="openstack/nova-cell1-novncproxy-0" Oct 06 10:18:47 crc kubenswrapper[4824]: I1006 10:18:47.189062 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/db4434eb-50c9-4004-8ff3-752cffaa01b4-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"db4434eb-50c9-4004-8ff3-752cffaa01b4\") " pod="openstack/nova-cell1-novncproxy-0" Oct 06 10:18:47 crc kubenswrapper[4824]: I1006 10:18:47.193765 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db4434eb-50c9-4004-8ff3-752cffaa01b4-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"db4434eb-50c9-4004-8ff3-752cffaa01b4\") " pod="openstack/nova-cell1-novncproxy-0" Oct 06 10:18:47 crc kubenswrapper[4824]: I1006 10:18:47.193991 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db4434eb-50c9-4004-8ff3-752cffaa01b4-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"db4434eb-50c9-4004-8ff3-752cffaa01b4\") " pod="openstack/nova-cell1-novncproxy-0" Oct 06 10:18:47 crc kubenswrapper[4824]: I1006 10:18:47.194771 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/db4434eb-50c9-4004-8ff3-752cffaa01b4-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"db4434eb-50c9-4004-8ff3-752cffaa01b4\") " pod="openstack/nova-cell1-novncproxy-0" Oct 06 10:18:47 crc kubenswrapper[4824]: I1006 10:18:47.195084 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/db4434eb-50c9-4004-8ff3-752cffaa01b4-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"db4434eb-50c9-4004-8ff3-752cffaa01b4\") " pod="openstack/nova-cell1-novncproxy-0" Oct 06 10:18:47 crc kubenswrapper[4824]: I1006 10:18:47.208080 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t58db\" (UniqueName: \"kubernetes.io/projected/db4434eb-50c9-4004-8ff3-752cffaa01b4-kube-api-access-t58db\") pod \"nova-cell1-novncproxy-0\" (UID: \"db4434eb-50c9-4004-8ff3-752cffaa01b4\") " pod="openstack/nova-cell1-novncproxy-0" Oct 06 10:18:47 crc kubenswrapper[4824]: I1006 10:18:47.286168 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="627f8bd8-ea33-486d-ac9e-2cc7c0b2ae15" path="/var/lib/kubelet/pods/627f8bd8-ea33-486d-ac9e-2cc7c0b2ae15/volumes" Oct 06 10:18:47 crc kubenswrapper[4824]: I1006 10:18:47.340930 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 06 10:18:47 crc kubenswrapper[4824]: I1006 10:18:47.838601 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 06 10:18:47 crc kubenswrapper[4824]: I1006 10:18:47.904235 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"db4434eb-50c9-4004-8ff3-752cffaa01b4","Type":"ContainerStarted","Data":"da7719429de3b8c6598a700444e1641cb0c4c7a514aa8945f53ca60db36b90f5"} Oct 06 10:18:48 crc kubenswrapper[4824]: I1006 10:18:48.926224 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"db4434eb-50c9-4004-8ff3-752cffaa01b4","Type":"ContainerStarted","Data":"7438c5746c0a85c75fce9dcbfc9aa89f45b37c17ed6d45bf1518d4faadb6c88d"} Oct 06 10:18:48 crc kubenswrapper[4824]: I1006 10:18:48.944949 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.9449229150000003 podStartE2EDuration="2.944922915s" podCreationTimestamp="2025-10-06 10:18:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:18:48.94228155 +0000 UTC m=+1238.306704411" watchObservedRunningTime="2025-10-06 10:18:48.944922915 +0000 UTC m=+1238.309345776" Oct 06 10:18:49 crc kubenswrapper[4824]: I1006 10:18:49.333306 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 06 10:18:49 crc kubenswrapper[4824]: I1006 10:18:49.334332 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 06 10:18:49 crc kubenswrapper[4824]: I1006 10:18:49.334363 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 06 10:18:49 crc kubenswrapper[4824]: I1006 10:18:49.338351 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 06 10:18:49 crc kubenswrapper[4824]: I1006 10:18:49.937596 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 06 10:18:49 crc kubenswrapper[4824]: I1006 10:18:49.952012 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 06 10:18:50 crc kubenswrapper[4824]: I1006 10:18:50.165847 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-pk86t"] Oct 06 10:18:50 crc kubenswrapper[4824]: I1006 10:18:50.167887 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59cf4bdb65-pk86t" Oct 06 10:18:50 crc kubenswrapper[4824]: I1006 10:18:50.192416 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-pk86t"] Oct 06 10:18:50 crc kubenswrapper[4824]: I1006 10:18:50.263748 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hdrlh\" (UniqueName: \"kubernetes.io/projected/b940b381-928a-4f5d-b91e-0da628631f7c-kube-api-access-hdrlh\") pod \"dnsmasq-dns-59cf4bdb65-pk86t\" (UID: \"b940b381-928a-4f5d-b91e-0da628631f7c\") " pod="openstack/dnsmasq-dns-59cf4bdb65-pk86t" Oct 06 10:18:50 crc kubenswrapper[4824]: I1006 10:18:50.263814 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b940b381-928a-4f5d-b91e-0da628631f7c-dns-svc\") pod \"dnsmasq-dns-59cf4bdb65-pk86t\" (UID: \"b940b381-928a-4f5d-b91e-0da628631f7c\") " pod="openstack/dnsmasq-dns-59cf4bdb65-pk86t" Oct 06 10:18:50 crc kubenswrapper[4824]: I1006 10:18:50.263858 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b940b381-928a-4f5d-b91e-0da628631f7c-dns-swift-storage-0\") pod \"dnsmasq-dns-59cf4bdb65-pk86t\" (UID: \"b940b381-928a-4f5d-b91e-0da628631f7c\") " pod="openstack/dnsmasq-dns-59cf4bdb65-pk86t" Oct 06 10:18:50 crc kubenswrapper[4824]: I1006 10:18:50.263936 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b940b381-928a-4f5d-b91e-0da628631f7c-ovsdbserver-sb\") pod \"dnsmasq-dns-59cf4bdb65-pk86t\" (UID: \"b940b381-928a-4f5d-b91e-0da628631f7c\") " pod="openstack/dnsmasq-dns-59cf4bdb65-pk86t" Oct 06 10:18:50 crc kubenswrapper[4824]: I1006 10:18:50.264163 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b940b381-928a-4f5d-b91e-0da628631f7c-config\") pod \"dnsmasq-dns-59cf4bdb65-pk86t\" (UID: \"b940b381-928a-4f5d-b91e-0da628631f7c\") " pod="openstack/dnsmasq-dns-59cf4bdb65-pk86t" Oct 06 10:18:50 crc kubenswrapper[4824]: I1006 10:18:50.264222 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b940b381-928a-4f5d-b91e-0da628631f7c-ovsdbserver-nb\") pod \"dnsmasq-dns-59cf4bdb65-pk86t\" (UID: \"b940b381-928a-4f5d-b91e-0da628631f7c\") " pod="openstack/dnsmasq-dns-59cf4bdb65-pk86t" Oct 06 10:18:50 crc kubenswrapper[4824]: I1006 10:18:50.366095 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b940b381-928a-4f5d-b91e-0da628631f7c-config\") pod \"dnsmasq-dns-59cf4bdb65-pk86t\" (UID: \"b940b381-928a-4f5d-b91e-0da628631f7c\") " pod="openstack/dnsmasq-dns-59cf4bdb65-pk86t" Oct 06 10:18:50 crc kubenswrapper[4824]: I1006 10:18:50.366530 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b940b381-928a-4f5d-b91e-0da628631f7c-ovsdbserver-nb\") pod \"dnsmasq-dns-59cf4bdb65-pk86t\" (UID: \"b940b381-928a-4f5d-b91e-0da628631f7c\") " pod="openstack/dnsmasq-dns-59cf4bdb65-pk86t" Oct 06 10:18:50 crc kubenswrapper[4824]: I1006 10:18:50.366690 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hdrlh\" (UniqueName: \"kubernetes.io/projected/b940b381-928a-4f5d-b91e-0da628631f7c-kube-api-access-hdrlh\") pod \"dnsmasq-dns-59cf4bdb65-pk86t\" (UID: \"b940b381-928a-4f5d-b91e-0da628631f7c\") " pod="openstack/dnsmasq-dns-59cf4bdb65-pk86t" Oct 06 10:18:50 crc kubenswrapper[4824]: I1006 10:18:50.366795 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b940b381-928a-4f5d-b91e-0da628631f7c-dns-svc\") pod \"dnsmasq-dns-59cf4bdb65-pk86t\" (UID: \"b940b381-928a-4f5d-b91e-0da628631f7c\") " pod="openstack/dnsmasq-dns-59cf4bdb65-pk86t" Oct 06 10:18:50 crc kubenswrapper[4824]: I1006 10:18:50.366906 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b940b381-928a-4f5d-b91e-0da628631f7c-dns-swift-storage-0\") pod \"dnsmasq-dns-59cf4bdb65-pk86t\" (UID: \"b940b381-928a-4f5d-b91e-0da628631f7c\") " pod="openstack/dnsmasq-dns-59cf4bdb65-pk86t" Oct 06 10:18:50 crc kubenswrapper[4824]: I1006 10:18:50.367108 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b940b381-928a-4f5d-b91e-0da628631f7c-ovsdbserver-sb\") pod \"dnsmasq-dns-59cf4bdb65-pk86t\" (UID: \"b940b381-928a-4f5d-b91e-0da628631f7c\") " pod="openstack/dnsmasq-dns-59cf4bdb65-pk86t" Oct 06 10:18:50 crc kubenswrapper[4824]: I1006 10:18:50.367103 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b940b381-928a-4f5d-b91e-0da628631f7c-config\") pod \"dnsmasq-dns-59cf4bdb65-pk86t\" (UID: \"b940b381-928a-4f5d-b91e-0da628631f7c\") " pod="openstack/dnsmasq-dns-59cf4bdb65-pk86t" Oct 06 10:18:50 crc kubenswrapper[4824]: I1006 10:18:50.367516 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b940b381-928a-4f5d-b91e-0da628631f7c-ovsdbserver-nb\") pod \"dnsmasq-dns-59cf4bdb65-pk86t\" (UID: \"b940b381-928a-4f5d-b91e-0da628631f7c\") " pod="openstack/dnsmasq-dns-59cf4bdb65-pk86t" Oct 06 10:18:50 crc kubenswrapper[4824]: I1006 10:18:50.367665 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b940b381-928a-4f5d-b91e-0da628631f7c-dns-svc\") pod \"dnsmasq-dns-59cf4bdb65-pk86t\" (UID: \"b940b381-928a-4f5d-b91e-0da628631f7c\") " pod="openstack/dnsmasq-dns-59cf4bdb65-pk86t" Oct 06 10:18:50 crc kubenswrapper[4824]: I1006 10:18:50.367784 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b940b381-928a-4f5d-b91e-0da628631f7c-ovsdbserver-sb\") pod \"dnsmasq-dns-59cf4bdb65-pk86t\" (UID: \"b940b381-928a-4f5d-b91e-0da628631f7c\") " pod="openstack/dnsmasq-dns-59cf4bdb65-pk86t" Oct 06 10:18:50 crc kubenswrapper[4824]: I1006 10:18:50.368142 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b940b381-928a-4f5d-b91e-0da628631f7c-dns-swift-storage-0\") pod \"dnsmasq-dns-59cf4bdb65-pk86t\" (UID: \"b940b381-928a-4f5d-b91e-0da628631f7c\") " pod="openstack/dnsmasq-dns-59cf4bdb65-pk86t" Oct 06 10:18:50 crc kubenswrapper[4824]: I1006 10:18:50.401346 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hdrlh\" (UniqueName: \"kubernetes.io/projected/b940b381-928a-4f5d-b91e-0da628631f7c-kube-api-access-hdrlh\") pod \"dnsmasq-dns-59cf4bdb65-pk86t\" (UID: \"b940b381-928a-4f5d-b91e-0da628631f7c\") " pod="openstack/dnsmasq-dns-59cf4bdb65-pk86t" Oct 06 10:18:50 crc kubenswrapper[4824]: I1006 10:18:50.494263 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59cf4bdb65-pk86t" Oct 06 10:18:51 crc kubenswrapper[4824]: I1006 10:18:51.037301 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-pk86t"] Oct 06 10:18:51 crc kubenswrapper[4824]: W1006 10:18:51.045010 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb940b381_928a_4f5d_b91e_0da628631f7c.slice/crio-06242b7e1651eb30be31bba669ae59149898c09bd9ae14975275ecc623b0fddf WatchSource:0}: Error finding container 06242b7e1651eb30be31bba669ae59149898c09bd9ae14975275ecc623b0fddf: Status 404 returned error can't find the container with id 06242b7e1651eb30be31bba669ae59149898c09bd9ae14975275ecc623b0fddf Oct 06 10:18:51 crc kubenswrapper[4824]: I1006 10:18:51.956282 4824 generic.go:334] "Generic (PLEG): container finished" podID="b940b381-928a-4f5d-b91e-0da628631f7c" containerID="10ca06c8cefd1e8577231fd8670f5e205b5d1475d91bb02a7c3168387210b59a" exitCode=0 Oct 06 10:18:51 crc kubenswrapper[4824]: I1006 10:18:51.956336 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cf4bdb65-pk86t" event={"ID":"b940b381-928a-4f5d-b91e-0da628631f7c","Type":"ContainerDied","Data":"10ca06c8cefd1e8577231fd8670f5e205b5d1475d91bb02a7c3168387210b59a"} Oct 06 10:18:51 crc kubenswrapper[4824]: I1006 10:18:51.956885 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cf4bdb65-pk86t" event={"ID":"b940b381-928a-4f5d-b91e-0da628631f7c","Type":"ContainerStarted","Data":"06242b7e1651eb30be31bba669ae59149898c09bd9ae14975275ecc623b0fddf"} Oct 06 10:18:52 crc kubenswrapper[4824]: I1006 10:18:52.342097 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Oct 06 10:18:52 crc kubenswrapper[4824]: I1006 10:18:52.701032 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 06 10:18:52 crc kubenswrapper[4824]: I1006 10:18:52.701917 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="aa150ff6-e61e-4deb-9a72-98fbcc8389b8" containerName="ceilometer-central-agent" containerID="cri-o://e168848b51f45d398d90fb4b9ad42f43b4a7d30027bbfd84bb68c55a84747017" gracePeriod=30 Oct 06 10:18:52 crc kubenswrapper[4824]: I1006 10:18:52.702006 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="aa150ff6-e61e-4deb-9a72-98fbcc8389b8" containerName="sg-core" containerID="cri-o://a4ea9aec64f9ffadabc0fd0437a43c68ca18deb02a1e6848e8d3efc6bd6855c1" gracePeriod=30 Oct 06 10:18:52 crc kubenswrapper[4824]: I1006 10:18:52.702097 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="aa150ff6-e61e-4deb-9a72-98fbcc8389b8" containerName="proxy-httpd" containerID="cri-o://81c4f0456c769719e1bf6196c4a5067d84301428c756ca7b242e54aecb3f2bb1" gracePeriod=30 Oct 06 10:18:52 crc kubenswrapper[4824]: I1006 10:18:52.702106 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="aa150ff6-e61e-4deb-9a72-98fbcc8389b8" containerName="ceilometer-notification-agent" containerID="cri-o://a520bed817b15e3c0e6355c65a738a99f59a361edb219a69a49fbac1f16671a5" gracePeriod=30 Oct 06 10:18:52 crc kubenswrapper[4824]: I1006 10:18:52.722289 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="aa150ff6-e61e-4deb-9a72-98fbcc8389b8" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.200:3000/\": EOF" Oct 06 10:18:52 crc kubenswrapper[4824]: I1006 10:18:52.882833 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 06 10:18:53 crc kubenswrapper[4824]: I1006 10:18:53.061942 4824 generic.go:334] "Generic (PLEG): container finished" podID="aa150ff6-e61e-4deb-9a72-98fbcc8389b8" containerID="81c4f0456c769719e1bf6196c4a5067d84301428c756ca7b242e54aecb3f2bb1" exitCode=0 Oct 06 10:18:53 crc kubenswrapper[4824]: I1006 10:18:53.062156 4824 generic.go:334] "Generic (PLEG): container finished" podID="aa150ff6-e61e-4deb-9a72-98fbcc8389b8" containerID="a4ea9aec64f9ffadabc0fd0437a43c68ca18deb02a1e6848e8d3efc6bd6855c1" exitCode=2 Oct 06 10:18:53 crc kubenswrapper[4824]: I1006 10:18:53.062061 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aa150ff6-e61e-4deb-9a72-98fbcc8389b8","Type":"ContainerDied","Data":"81c4f0456c769719e1bf6196c4a5067d84301428c756ca7b242e54aecb3f2bb1"} Oct 06 10:18:53 crc kubenswrapper[4824]: I1006 10:18:53.062370 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aa150ff6-e61e-4deb-9a72-98fbcc8389b8","Type":"ContainerDied","Data":"a4ea9aec64f9ffadabc0fd0437a43c68ca18deb02a1e6848e8d3efc6bd6855c1"} Oct 06 10:18:53 crc kubenswrapper[4824]: I1006 10:18:53.063985 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cf4bdb65-pk86t" event={"ID":"b940b381-928a-4f5d-b91e-0da628631f7c","Type":"ContainerStarted","Data":"8071c18c77c91063dea49b3439242174955cb8078a65deffe8537ccc795d0394"} Oct 06 10:18:53 crc kubenswrapper[4824]: I1006 10:18:53.064324 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="ebdd4e3c-8159-4542-b348-109f958f9a5f" containerName="nova-api-log" containerID="cri-o://63be74dc92bd5218a7077af368d5c701873e4e262c7ead1808da0dd6d507e839" gracePeriod=30 Oct 06 10:18:53 crc kubenswrapper[4824]: I1006 10:18:53.064438 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="ebdd4e3c-8159-4542-b348-109f958f9a5f" containerName="nova-api-api" containerID="cri-o://00f934bf252754731c575469bc520b850aa42c980a66ac9f12ed730b0da97dbb" gracePeriod=30 Oct 06 10:18:53 crc kubenswrapper[4824]: I1006 10:18:53.102715 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-59cf4bdb65-pk86t" podStartSLOduration=3.102692717 podStartE2EDuration="3.102692717s" podCreationTimestamp="2025-10-06 10:18:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:18:53.096546803 +0000 UTC m=+1242.460969664" watchObservedRunningTime="2025-10-06 10:18:53.102692717 +0000 UTC m=+1242.467115578" Oct 06 10:18:53 crc kubenswrapper[4824]: I1006 10:18:53.644500 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 06 10:18:53 crc kubenswrapper[4824]: I1006 10:18:53.745769 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/aa150ff6-e61e-4deb-9a72-98fbcc8389b8-sg-core-conf-yaml\") pod \"aa150ff6-e61e-4deb-9a72-98fbcc8389b8\" (UID: \"aa150ff6-e61e-4deb-9a72-98fbcc8389b8\") " Oct 06 10:18:53 crc kubenswrapper[4824]: I1006 10:18:53.746169 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa150ff6-e61e-4deb-9a72-98fbcc8389b8-config-data\") pod \"aa150ff6-e61e-4deb-9a72-98fbcc8389b8\" (UID: \"aa150ff6-e61e-4deb-9a72-98fbcc8389b8\") " Oct 06 10:18:53 crc kubenswrapper[4824]: I1006 10:18:53.746258 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa150ff6-e61e-4deb-9a72-98fbcc8389b8-combined-ca-bundle\") pod \"aa150ff6-e61e-4deb-9a72-98fbcc8389b8\" (UID: \"aa150ff6-e61e-4deb-9a72-98fbcc8389b8\") " Oct 06 10:18:53 crc kubenswrapper[4824]: I1006 10:18:53.746458 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aa150ff6-e61e-4deb-9a72-98fbcc8389b8-scripts\") pod \"aa150ff6-e61e-4deb-9a72-98fbcc8389b8\" (UID: \"aa150ff6-e61e-4deb-9a72-98fbcc8389b8\") " Oct 06 10:18:53 crc kubenswrapper[4824]: I1006 10:18:53.746606 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c9f8x\" (UniqueName: \"kubernetes.io/projected/aa150ff6-e61e-4deb-9a72-98fbcc8389b8-kube-api-access-c9f8x\") pod \"aa150ff6-e61e-4deb-9a72-98fbcc8389b8\" (UID: \"aa150ff6-e61e-4deb-9a72-98fbcc8389b8\") " Oct 06 10:18:53 crc kubenswrapper[4824]: I1006 10:18:53.746709 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/aa150ff6-e61e-4deb-9a72-98fbcc8389b8-ceilometer-tls-certs\") pod \"aa150ff6-e61e-4deb-9a72-98fbcc8389b8\" (UID: \"aa150ff6-e61e-4deb-9a72-98fbcc8389b8\") " Oct 06 10:18:53 crc kubenswrapper[4824]: I1006 10:18:53.746797 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aa150ff6-e61e-4deb-9a72-98fbcc8389b8-run-httpd\") pod \"aa150ff6-e61e-4deb-9a72-98fbcc8389b8\" (UID: \"aa150ff6-e61e-4deb-9a72-98fbcc8389b8\") " Oct 06 10:18:53 crc kubenswrapper[4824]: I1006 10:18:53.746970 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aa150ff6-e61e-4deb-9a72-98fbcc8389b8-log-httpd\") pod \"aa150ff6-e61e-4deb-9a72-98fbcc8389b8\" (UID: \"aa150ff6-e61e-4deb-9a72-98fbcc8389b8\") " Oct 06 10:18:53 crc kubenswrapper[4824]: I1006 10:18:53.747462 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aa150ff6-e61e-4deb-9a72-98fbcc8389b8-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "aa150ff6-e61e-4deb-9a72-98fbcc8389b8" (UID: "aa150ff6-e61e-4deb-9a72-98fbcc8389b8"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:18:53 crc kubenswrapper[4824]: I1006 10:18:53.747879 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aa150ff6-e61e-4deb-9a72-98fbcc8389b8-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "aa150ff6-e61e-4deb-9a72-98fbcc8389b8" (UID: "aa150ff6-e61e-4deb-9a72-98fbcc8389b8"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:18:53 crc kubenswrapper[4824]: I1006 10:18:53.752637 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aa150ff6-e61e-4deb-9a72-98fbcc8389b8-kube-api-access-c9f8x" (OuterVolumeSpecName: "kube-api-access-c9f8x") pod "aa150ff6-e61e-4deb-9a72-98fbcc8389b8" (UID: "aa150ff6-e61e-4deb-9a72-98fbcc8389b8"). InnerVolumeSpecName "kube-api-access-c9f8x". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:18:53 crc kubenswrapper[4824]: I1006 10:18:53.754017 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa150ff6-e61e-4deb-9a72-98fbcc8389b8-scripts" (OuterVolumeSpecName: "scripts") pod "aa150ff6-e61e-4deb-9a72-98fbcc8389b8" (UID: "aa150ff6-e61e-4deb-9a72-98fbcc8389b8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:18:53 crc kubenswrapper[4824]: I1006 10:18:53.786637 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa150ff6-e61e-4deb-9a72-98fbcc8389b8-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "aa150ff6-e61e-4deb-9a72-98fbcc8389b8" (UID: "aa150ff6-e61e-4deb-9a72-98fbcc8389b8"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:18:53 crc kubenswrapper[4824]: I1006 10:18:53.843085 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa150ff6-e61e-4deb-9a72-98fbcc8389b8-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "aa150ff6-e61e-4deb-9a72-98fbcc8389b8" (UID: "aa150ff6-e61e-4deb-9a72-98fbcc8389b8"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:18:53 crc kubenswrapper[4824]: I1006 10:18:53.849431 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c9f8x\" (UniqueName: \"kubernetes.io/projected/aa150ff6-e61e-4deb-9a72-98fbcc8389b8-kube-api-access-c9f8x\") on node \"crc\" DevicePath \"\"" Oct 06 10:18:53 crc kubenswrapper[4824]: I1006 10:18:53.849460 4824 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/aa150ff6-e61e-4deb-9a72-98fbcc8389b8-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 06 10:18:53 crc kubenswrapper[4824]: I1006 10:18:53.849473 4824 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aa150ff6-e61e-4deb-9a72-98fbcc8389b8-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 06 10:18:53 crc kubenswrapper[4824]: I1006 10:18:53.849487 4824 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aa150ff6-e61e-4deb-9a72-98fbcc8389b8-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 06 10:18:53 crc kubenswrapper[4824]: I1006 10:18:53.849495 4824 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/aa150ff6-e61e-4deb-9a72-98fbcc8389b8-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 06 10:18:53 crc kubenswrapper[4824]: I1006 10:18:53.849503 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aa150ff6-e61e-4deb-9a72-98fbcc8389b8-scripts\") on node \"crc\" DevicePath \"\"" Oct 06 10:18:53 crc kubenswrapper[4824]: I1006 10:18:53.865427 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa150ff6-e61e-4deb-9a72-98fbcc8389b8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "aa150ff6-e61e-4deb-9a72-98fbcc8389b8" (UID: "aa150ff6-e61e-4deb-9a72-98fbcc8389b8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:18:53 crc kubenswrapper[4824]: I1006 10:18:53.918712 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa150ff6-e61e-4deb-9a72-98fbcc8389b8-config-data" (OuterVolumeSpecName: "config-data") pod "aa150ff6-e61e-4deb-9a72-98fbcc8389b8" (UID: "aa150ff6-e61e-4deb-9a72-98fbcc8389b8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:18:53 crc kubenswrapper[4824]: I1006 10:18:53.952243 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa150ff6-e61e-4deb-9a72-98fbcc8389b8-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 10:18:53 crc kubenswrapper[4824]: I1006 10:18:53.952303 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa150ff6-e61e-4deb-9a72-98fbcc8389b8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 10:18:54 crc kubenswrapper[4824]: I1006 10:18:54.078609 4824 generic.go:334] "Generic (PLEG): container finished" podID="aa150ff6-e61e-4deb-9a72-98fbcc8389b8" containerID="a520bed817b15e3c0e6355c65a738a99f59a361edb219a69a49fbac1f16671a5" exitCode=0 Oct 06 10:18:54 crc kubenswrapper[4824]: I1006 10:18:54.078943 4824 generic.go:334] "Generic (PLEG): container finished" podID="aa150ff6-e61e-4deb-9a72-98fbcc8389b8" containerID="e168848b51f45d398d90fb4b9ad42f43b4a7d30027bbfd84bb68c55a84747017" exitCode=0 Oct 06 10:18:54 crc kubenswrapper[4824]: I1006 10:18:54.078762 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aa150ff6-e61e-4deb-9a72-98fbcc8389b8","Type":"ContainerDied","Data":"a520bed817b15e3c0e6355c65a738a99f59a361edb219a69a49fbac1f16671a5"} Oct 06 10:18:54 crc kubenswrapper[4824]: I1006 10:18:54.078780 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 06 10:18:54 crc kubenswrapper[4824]: I1006 10:18:54.079067 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aa150ff6-e61e-4deb-9a72-98fbcc8389b8","Type":"ContainerDied","Data":"e168848b51f45d398d90fb4b9ad42f43b4a7d30027bbfd84bb68c55a84747017"} Oct 06 10:18:54 crc kubenswrapper[4824]: I1006 10:18:54.079115 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aa150ff6-e61e-4deb-9a72-98fbcc8389b8","Type":"ContainerDied","Data":"082f4345738ef5d6215dfe58811ff8ec41b0a626ffeaaf0324e54c2f03b11902"} Oct 06 10:18:54 crc kubenswrapper[4824]: I1006 10:18:54.079153 4824 scope.go:117] "RemoveContainer" containerID="81c4f0456c769719e1bf6196c4a5067d84301428c756ca7b242e54aecb3f2bb1" Oct 06 10:18:54 crc kubenswrapper[4824]: I1006 10:18:54.083542 4824 generic.go:334] "Generic (PLEG): container finished" podID="ebdd4e3c-8159-4542-b348-109f958f9a5f" containerID="63be74dc92bd5218a7077af368d5c701873e4e262c7ead1808da0dd6d507e839" exitCode=143 Oct 06 10:18:54 crc kubenswrapper[4824]: I1006 10:18:54.083601 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ebdd4e3c-8159-4542-b348-109f958f9a5f","Type":"ContainerDied","Data":"63be74dc92bd5218a7077af368d5c701873e4e262c7ead1808da0dd6d507e839"} Oct 06 10:18:54 crc kubenswrapper[4824]: I1006 10:18:54.084029 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-59cf4bdb65-pk86t" Oct 06 10:18:54 crc kubenswrapper[4824]: I1006 10:18:54.119201 4824 scope.go:117] "RemoveContainer" containerID="a4ea9aec64f9ffadabc0fd0437a43c68ca18deb02a1e6848e8d3efc6bd6855c1" Oct 06 10:18:54 crc kubenswrapper[4824]: I1006 10:18:54.130228 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 06 10:18:54 crc kubenswrapper[4824]: I1006 10:18:54.141860 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 06 10:18:54 crc kubenswrapper[4824]: I1006 10:18:54.160388 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 06 10:18:54 crc kubenswrapper[4824]: E1006 10:18:54.160920 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa150ff6-e61e-4deb-9a72-98fbcc8389b8" containerName="ceilometer-central-agent" Oct 06 10:18:54 crc kubenswrapper[4824]: I1006 10:18:54.160950 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa150ff6-e61e-4deb-9a72-98fbcc8389b8" containerName="ceilometer-central-agent" Oct 06 10:18:54 crc kubenswrapper[4824]: E1006 10:18:54.160997 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa150ff6-e61e-4deb-9a72-98fbcc8389b8" containerName="ceilometer-notification-agent" Oct 06 10:18:54 crc kubenswrapper[4824]: I1006 10:18:54.161005 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa150ff6-e61e-4deb-9a72-98fbcc8389b8" containerName="ceilometer-notification-agent" Oct 06 10:18:54 crc kubenswrapper[4824]: E1006 10:18:54.161022 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa150ff6-e61e-4deb-9a72-98fbcc8389b8" containerName="proxy-httpd" Oct 06 10:18:54 crc kubenswrapper[4824]: I1006 10:18:54.161028 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa150ff6-e61e-4deb-9a72-98fbcc8389b8" containerName="proxy-httpd" Oct 06 10:18:54 crc kubenswrapper[4824]: E1006 10:18:54.161051 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa150ff6-e61e-4deb-9a72-98fbcc8389b8" containerName="sg-core" Oct 06 10:18:54 crc kubenswrapper[4824]: I1006 10:18:54.161059 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa150ff6-e61e-4deb-9a72-98fbcc8389b8" containerName="sg-core" Oct 06 10:18:54 crc kubenswrapper[4824]: I1006 10:18:54.161297 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa150ff6-e61e-4deb-9a72-98fbcc8389b8" containerName="ceilometer-notification-agent" Oct 06 10:18:54 crc kubenswrapper[4824]: I1006 10:18:54.161317 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa150ff6-e61e-4deb-9a72-98fbcc8389b8" containerName="sg-core" Oct 06 10:18:54 crc kubenswrapper[4824]: I1006 10:18:54.161332 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa150ff6-e61e-4deb-9a72-98fbcc8389b8" containerName="proxy-httpd" Oct 06 10:18:54 crc kubenswrapper[4824]: I1006 10:18:54.161347 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa150ff6-e61e-4deb-9a72-98fbcc8389b8" containerName="ceilometer-central-agent" Oct 06 10:18:54 crc kubenswrapper[4824]: I1006 10:18:54.165872 4824 scope.go:117] "RemoveContainer" containerID="a520bed817b15e3c0e6355c65a738a99f59a361edb219a69a49fbac1f16671a5" Oct 06 10:18:54 crc kubenswrapper[4824]: I1006 10:18:54.166675 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 06 10:18:54 crc kubenswrapper[4824]: I1006 10:18:54.170861 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 06 10:18:54 crc kubenswrapper[4824]: I1006 10:18:54.170953 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 06 10:18:54 crc kubenswrapper[4824]: I1006 10:18:54.171169 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 06 10:18:54 crc kubenswrapper[4824]: I1006 10:18:54.196460 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 06 10:18:54 crc kubenswrapper[4824]: I1006 10:18:54.235176 4824 scope.go:117] "RemoveContainer" containerID="e168848b51f45d398d90fb4b9ad42f43b4a7d30027bbfd84bb68c55a84747017" Oct 06 10:18:54 crc kubenswrapper[4824]: I1006 10:18:54.257544 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3d253c1e-62ea-4559-8a58-63ddfbf02f88-run-httpd\") pod \"ceilometer-0\" (UID: \"3d253c1e-62ea-4559-8a58-63ddfbf02f88\") " pod="openstack/ceilometer-0" Oct 06 10:18:54 crc kubenswrapper[4824]: I1006 10:18:54.258171 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d253c1e-62ea-4559-8a58-63ddfbf02f88-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3d253c1e-62ea-4559-8a58-63ddfbf02f88\") " pod="openstack/ceilometer-0" Oct 06 10:18:54 crc kubenswrapper[4824]: I1006 10:18:54.258218 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3d253c1e-62ea-4559-8a58-63ddfbf02f88-scripts\") pod \"ceilometer-0\" (UID: \"3d253c1e-62ea-4559-8a58-63ddfbf02f88\") " pod="openstack/ceilometer-0" Oct 06 10:18:54 crc kubenswrapper[4824]: I1006 10:18:54.258239 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/3d253c1e-62ea-4559-8a58-63ddfbf02f88-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"3d253c1e-62ea-4559-8a58-63ddfbf02f88\") " pod="openstack/ceilometer-0" Oct 06 10:18:54 crc kubenswrapper[4824]: I1006 10:18:54.258277 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3d253c1e-62ea-4559-8a58-63ddfbf02f88-log-httpd\") pod \"ceilometer-0\" (UID: \"3d253c1e-62ea-4559-8a58-63ddfbf02f88\") " pod="openstack/ceilometer-0" Oct 06 10:18:54 crc kubenswrapper[4824]: I1006 10:18:54.258326 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3d253c1e-62ea-4559-8a58-63ddfbf02f88-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3d253c1e-62ea-4559-8a58-63ddfbf02f88\") " pod="openstack/ceilometer-0" Oct 06 10:18:54 crc kubenswrapper[4824]: I1006 10:18:54.258438 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-47dp2\" (UniqueName: \"kubernetes.io/projected/3d253c1e-62ea-4559-8a58-63ddfbf02f88-kube-api-access-47dp2\") pod \"ceilometer-0\" (UID: \"3d253c1e-62ea-4559-8a58-63ddfbf02f88\") " pod="openstack/ceilometer-0" Oct 06 10:18:54 crc kubenswrapper[4824]: I1006 10:18:54.258467 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3d253c1e-62ea-4559-8a58-63ddfbf02f88-config-data\") pod \"ceilometer-0\" (UID: \"3d253c1e-62ea-4559-8a58-63ddfbf02f88\") " pod="openstack/ceilometer-0" Oct 06 10:18:54 crc kubenswrapper[4824]: I1006 10:18:54.264459 4824 scope.go:117] "RemoveContainer" containerID="81c4f0456c769719e1bf6196c4a5067d84301428c756ca7b242e54aecb3f2bb1" Oct 06 10:18:54 crc kubenswrapper[4824]: E1006 10:18:54.265970 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"81c4f0456c769719e1bf6196c4a5067d84301428c756ca7b242e54aecb3f2bb1\": container with ID starting with 81c4f0456c769719e1bf6196c4a5067d84301428c756ca7b242e54aecb3f2bb1 not found: ID does not exist" containerID="81c4f0456c769719e1bf6196c4a5067d84301428c756ca7b242e54aecb3f2bb1" Oct 06 10:18:54 crc kubenswrapper[4824]: I1006 10:18:54.266118 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"81c4f0456c769719e1bf6196c4a5067d84301428c756ca7b242e54aecb3f2bb1"} err="failed to get container status \"81c4f0456c769719e1bf6196c4a5067d84301428c756ca7b242e54aecb3f2bb1\": rpc error: code = NotFound desc = could not find container \"81c4f0456c769719e1bf6196c4a5067d84301428c756ca7b242e54aecb3f2bb1\": container with ID starting with 81c4f0456c769719e1bf6196c4a5067d84301428c756ca7b242e54aecb3f2bb1 not found: ID does not exist" Oct 06 10:18:54 crc kubenswrapper[4824]: I1006 10:18:54.266146 4824 scope.go:117] "RemoveContainer" containerID="a4ea9aec64f9ffadabc0fd0437a43c68ca18deb02a1e6848e8d3efc6bd6855c1" Oct 06 10:18:54 crc kubenswrapper[4824]: E1006 10:18:54.267318 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a4ea9aec64f9ffadabc0fd0437a43c68ca18deb02a1e6848e8d3efc6bd6855c1\": container with ID starting with a4ea9aec64f9ffadabc0fd0437a43c68ca18deb02a1e6848e8d3efc6bd6855c1 not found: ID does not exist" containerID="a4ea9aec64f9ffadabc0fd0437a43c68ca18deb02a1e6848e8d3efc6bd6855c1" Oct 06 10:18:54 crc kubenswrapper[4824]: I1006 10:18:54.267340 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a4ea9aec64f9ffadabc0fd0437a43c68ca18deb02a1e6848e8d3efc6bd6855c1"} err="failed to get container status \"a4ea9aec64f9ffadabc0fd0437a43c68ca18deb02a1e6848e8d3efc6bd6855c1\": rpc error: code = NotFound desc = could not find container \"a4ea9aec64f9ffadabc0fd0437a43c68ca18deb02a1e6848e8d3efc6bd6855c1\": container with ID starting with a4ea9aec64f9ffadabc0fd0437a43c68ca18deb02a1e6848e8d3efc6bd6855c1 not found: ID does not exist" Oct 06 10:18:54 crc kubenswrapper[4824]: I1006 10:18:54.267353 4824 scope.go:117] "RemoveContainer" containerID="a520bed817b15e3c0e6355c65a738a99f59a361edb219a69a49fbac1f16671a5" Oct 06 10:18:54 crc kubenswrapper[4824]: E1006 10:18:54.267694 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a520bed817b15e3c0e6355c65a738a99f59a361edb219a69a49fbac1f16671a5\": container with ID starting with a520bed817b15e3c0e6355c65a738a99f59a361edb219a69a49fbac1f16671a5 not found: ID does not exist" containerID="a520bed817b15e3c0e6355c65a738a99f59a361edb219a69a49fbac1f16671a5" Oct 06 10:18:54 crc kubenswrapper[4824]: I1006 10:18:54.267737 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a520bed817b15e3c0e6355c65a738a99f59a361edb219a69a49fbac1f16671a5"} err="failed to get container status \"a520bed817b15e3c0e6355c65a738a99f59a361edb219a69a49fbac1f16671a5\": rpc error: code = NotFound desc = could not find container \"a520bed817b15e3c0e6355c65a738a99f59a361edb219a69a49fbac1f16671a5\": container with ID starting with a520bed817b15e3c0e6355c65a738a99f59a361edb219a69a49fbac1f16671a5 not found: ID does not exist" Oct 06 10:18:54 crc kubenswrapper[4824]: I1006 10:18:54.267764 4824 scope.go:117] "RemoveContainer" containerID="e168848b51f45d398d90fb4b9ad42f43b4a7d30027bbfd84bb68c55a84747017" Oct 06 10:18:54 crc kubenswrapper[4824]: E1006 10:18:54.268071 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e168848b51f45d398d90fb4b9ad42f43b4a7d30027bbfd84bb68c55a84747017\": container with ID starting with e168848b51f45d398d90fb4b9ad42f43b4a7d30027bbfd84bb68c55a84747017 not found: ID does not exist" containerID="e168848b51f45d398d90fb4b9ad42f43b4a7d30027bbfd84bb68c55a84747017" Oct 06 10:18:54 crc kubenswrapper[4824]: I1006 10:18:54.268096 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e168848b51f45d398d90fb4b9ad42f43b4a7d30027bbfd84bb68c55a84747017"} err="failed to get container status \"e168848b51f45d398d90fb4b9ad42f43b4a7d30027bbfd84bb68c55a84747017\": rpc error: code = NotFound desc = could not find container \"e168848b51f45d398d90fb4b9ad42f43b4a7d30027bbfd84bb68c55a84747017\": container with ID starting with e168848b51f45d398d90fb4b9ad42f43b4a7d30027bbfd84bb68c55a84747017 not found: ID does not exist" Oct 06 10:18:54 crc kubenswrapper[4824]: I1006 10:18:54.268114 4824 scope.go:117] "RemoveContainer" containerID="81c4f0456c769719e1bf6196c4a5067d84301428c756ca7b242e54aecb3f2bb1" Oct 06 10:18:54 crc kubenswrapper[4824]: I1006 10:18:54.268313 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"81c4f0456c769719e1bf6196c4a5067d84301428c756ca7b242e54aecb3f2bb1"} err="failed to get container status \"81c4f0456c769719e1bf6196c4a5067d84301428c756ca7b242e54aecb3f2bb1\": rpc error: code = NotFound desc = could not find container \"81c4f0456c769719e1bf6196c4a5067d84301428c756ca7b242e54aecb3f2bb1\": container with ID starting with 81c4f0456c769719e1bf6196c4a5067d84301428c756ca7b242e54aecb3f2bb1 not found: ID does not exist" Oct 06 10:18:54 crc kubenswrapper[4824]: I1006 10:18:54.268338 4824 scope.go:117] "RemoveContainer" containerID="a4ea9aec64f9ffadabc0fd0437a43c68ca18deb02a1e6848e8d3efc6bd6855c1" Oct 06 10:18:54 crc kubenswrapper[4824]: I1006 10:18:54.269271 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a4ea9aec64f9ffadabc0fd0437a43c68ca18deb02a1e6848e8d3efc6bd6855c1"} err="failed to get container status \"a4ea9aec64f9ffadabc0fd0437a43c68ca18deb02a1e6848e8d3efc6bd6855c1\": rpc error: code = NotFound desc = could not find container \"a4ea9aec64f9ffadabc0fd0437a43c68ca18deb02a1e6848e8d3efc6bd6855c1\": container with ID starting with a4ea9aec64f9ffadabc0fd0437a43c68ca18deb02a1e6848e8d3efc6bd6855c1 not found: ID does not exist" Oct 06 10:18:54 crc kubenswrapper[4824]: I1006 10:18:54.269295 4824 scope.go:117] "RemoveContainer" containerID="a520bed817b15e3c0e6355c65a738a99f59a361edb219a69a49fbac1f16671a5" Oct 06 10:18:54 crc kubenswrapper[4824]: I1006 10:18:54.269635 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a520bed817b15e3c0e6355c65a738a99f59a361edb219a69a49fbac1f16671a5"} err="failed to get container status \"a520bed817b15e3c0e6355c65a738a99f59a361edb219a69a49fbac1f16671a5\": rpc error: code = NotFound desc = could not find container \"a520bed817b15e3c0e6355c65a738a99f59a361edb219a69a49fbac1f16671a5\": container with ID starting with a520bed817b15e3c0e6355c65a738a99f59a361edb219a69a49fbac1f16671a5 not found: ID does not exist" Oct 06 10:18:54 crc kubenswrapper[4824]: I1006 10:18:54.269679 4824 scope.go:117] "RemoveContainer" containerID="e168848b51f45d398d90fb4b9ad42f43b4a7d30027bbfd84bb68c55a84747017" Oct 06 10:18:54 crc kubenswrapper[4824]: I1006 10:18:54.269992 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e168848b51f45d398d90fb4b9ad42f43b4a7d30027bbfd84bb68c55a84747017"} err="failed to get container status \"e168848b51f45d398d90fb4b9ad42f43b4a7d30027bbfd84bb68c55a84747017\": rpc error: code = NotFound desc = could not find container \"e168848b51f45d398d90fb4b9ad42f43b4a7d30027bbfd84bb68c55a84747017\": container with ID starting with e168848b51f45d398d90fb4b9ad42f43b4a7d30027bbfd84bb68c55a84747017 not found: ID does not exist" Oct 06 10:18:54 crc kubenswrapper[4824]: I1006 10:18:54.361210 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-47dp2\" (UniqueName: \"kubernetes.io/projected/3d253c1e-62ea-4559-8a58-63ddfbf02f88-kube-api-access-47dp2\") pod \"ceilometer-0\" (UID: \"3d253c1e-62ea-4559-8a58-63ddfbf02f88\") " pod="openstack/ceilometer-0" Oct 06 10:18:54 crc kubenswrapper[4824]: I1006 10:18:54.361335 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3d253c1e-62ea-4559-8a58-63ddfbf02f88-config-data\") pod \"ceilometer-0\" (UID: \"3d253c1e-62ea-4559-8a58-63ddfbf02f88\") " pod="openstack/ceilometer-0" Oct 06 10:18:54 crc kubenswrapper[4824]: I1006 10:18:54.361389 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3d253c1e-62ea-4559-8a58-63ddfbf02f88-run-httpd\") pod \"ceilometer-0\" (UID: \"3d253c1e-62ea-4559-8a58-63ddfbf02f88\") " pod="openstack/ceilometer-0" Oct 06 10:18:54 crc kubenswrapper[4824]: I1006 10:18:54.361490 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d253c1e-62ea-4559-8a58-63ddfbf02f88-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3d253c1e-62ea-4559-8a58-63ddfbf02f88\") " pod="openstack/ceilometer-0" Oct 06 10:18:54 crc kubenswrapper[4824]: I1006 10:18:54.361907 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3d253c1e-62ea-4559-8a58-63ddfbf02f88-run-httpd\") pod \"ceilometer-0\" (UID: \"3d253c1e-62ea-4559-8a58-63ddfbf02f88\") " pod="openstack/ceilometer-0" Oct 06 10:18:54 crc kubenswrapper[4824]: I1006 10:18:54.364521 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3d253c1e-62ea-4559-8a58-63ddfbf02f88-scripts\") pod \"ceilometer-0\" (UID: \"3d253c1e-62ea-4559-8a58-63ddfbf02f88\") " pod="openstack/ceilometer-0" Oct 06 10:18:54 crc kubenswrapper[4824]: I1006 10:18:54.364575 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/3d253c1e-62ea-4559-8a58-63ddfbf02f88-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"3d253c1e-62ea-4559-8a58-63ddfbf02f88\") " pod="openstack/ceilometer-0" Oct 06 10:18:54 crc kubenswrapper[4824]: I1006 10:18:54.364680 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3d253c1e-62ea-4559-8a58-63ddfbf02f88-log-httpd\") pod \"ceilometer-0\" (UID: \"3d253c1e-62ea-4559-8a58-63ddfbf02f88\") " pod="openstack/ceilometer-0" Oct 06 10:18:54 crc kubenswrapper[4824]: I1006 10:18:54.364787 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3d253c1e-62ea-4559-8a58-63ddfbf02f88-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3d253c1e-62ea-4559-8a58-63ddfbf02f88\") " pod="openstack/ceilometer-0" Oct 06 10:18:54 crc kubenswrapper[4824]: I1006 10:18:54.365201 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3d253c1e-62ea-4559-8a58-63ddfbf02f88-log-httpd\") pod \"ceilometer-0\" (UID: \"3d253c1e-62ea-4559-8a58-63ddfbf02f88\") " pod="openstack/ceilometer-0" Oct 06 10:18:54 crc kubenswrapper[4824]: I1006 10:18:54.367708 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3d253c1e-62ea-4559-8a58-63ddfbf02f88-scripts\") pod \"ceilometer-0\" (UID: \"3d253c1e-62ea-4559-8a58-63ddfbf02f88\") " pod="openstack/ceilometer-0" Oct 06 10:18:54 crc kubenswrapper[4824]: I1006 10:18:54.368163 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3d253c1e-62ea-4559-8a58-63ddfbf02f88-config-data\") pod \"ceilometer-0\" (UID: \"3d253c1e-62ea-4559-8a58-63ddfbf02f88\") " pod="openstack/ceilometer-0" Oct 06 10:18:54 crc kubenswrapper[4824]: I1006 10:18:54.368474 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/3d253c1e-62ea-4559-8a58-63ddfbf02f88-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"3d253c1e-62ea-4559-8a58-63ddfbf02f88\") " pod="openstack/ceilometer-0" Oct 06 10:18:54 crc kubenswrapper[4824]: I1006 10:18:54.369143 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d253c1e-62ea-4559-8a58-63ddfbf02f88-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3d253c1e-62ea-4559-8a58-63ddfbf02f88\") " pod="openstack/ceilometer-0" Oct 06 10:18:54 crc kubenswrapper[4824]: I1006 10:18:54.371883 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3d253c1e-62ea-4559-8a58-63ddfbf02f88-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3d253c1e-62ea-4559-8a58-63ddfbf02f88\") " pod="openstack/ceilometer-0" Oct 06 10:18:54 crc kubenswrapper[4824]: I1006 10:18:54.386626 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-47dp2\" (UniqueName: \"kubernetes.io/projected/3d253c1e-62ea-4559-8a58-63ddfbf02f88-kube-api-access-47dp2\") pod \"ceilometer-0\" (UID: \"3d253c1e-62ea-4559-8a58-63ddfbf02f88\") " pod="openstack/ceilometer-0" Oct 06 10:18:54 crc kubenswrapper[4824]: I1006 10:18:54.523006 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 06 10:18:55 crc kubenswrapper[4824]: I1006 10:18:55.142619 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 06 10:18:55 crc kubenswrapper[4824]: I1006 10:18:55.212472 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 06 10:18:55 crc kubenswrapper[4824]: W1006 10:18:55.216019 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3d253c1e_62ea_4559_8a58_63ddfbf02f88.slice/crio-33548dcb6b502aae91b4c41dde158ef8b72f252b9d2fb4e701d7f99b84e3f638 WatchSource:0}: Error finding container 33548dcb6b502aae91b4c41dde158ef8b72f252b9d2fb4e701d7f99b84e3f638: Status 404 returned error can't find the container with id 33548dcb6b502aae91b4c41dde158ef8b72f252b9d2fb4e701d7f99b84e3f638 Oct 06 10:18:55 crc kubenswrapper[4824]: I1006 10:18:55.298345 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aa150ff6-e61e-4deb-9a72-98fbcc8389b8" path="/var/lib/kubelet/pods/aa150ff6-e61e-4deb-9a72-98fbcc8389b8/volumes" Oct 06 10:18:56 crc kubenswrapper[4824]: I1006 10:18:56.104805 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3d253c1e-62ea-4559-8a58-63ddfbf02f88","Type":"ContainerStarted","Data":"33548dcb6b502aae91b4c41dde158ef8b72f252b9d2fb4e701d7f99b84e3f638"} Oct 06 10:18:56 crc kubenswrapper[4824]: I1006 10:18:56.744754 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 06 10:18:56 crc kubenswrapper[4824]: I1006 10:18:56.915129 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ebdd4e3c-8159-4542-b348-109f958f9a5f-logs\") pod \"ebdd4e3c-8159-4542-b348-109f958f9a5f\" (UID: \"ebdd4e3c-8159-4542-b348-109f958f9a5f\") " Oct 06 10:18:56 crc kubenswrapper[4824]: I1006 10:18:56.915317 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ebdd4e3c-8159-4542-b348-109f958f9a5f-combined-ca-bundle\") pod \"ebdd4e3c-8159-4542-b348-109f958f9a5f\" (UID: \"ebdd4e3c-8159-4542-b348-109f958f9a5f\") " Oct 06 10:18:56 crc kubenswrapper[4824]: I1006 10:18:56.915370 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q4gpc\" (UniqueName: \"kubernetes.io/projected/ebdd4e3c-8159-4542-b348-109f958f9a5f-kube-api-access-q4gpc\") pod \"ebdd4e3c-8159-4542-b348-109f958f9a5f\" (UID: \"ebdd4e3c-8159-4542-b348-109f958f9a5f\") " Oct 06 10:18:56 crc kubenswrapper[4824]: I1006 10:18:56.915421 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ebdd4e3c-8159-4542-b348-109f958f9a5f-config-data\") pod \"ebdd4e3c-8159-4542-b348-109f958f9a5f\" (UID: \"ebdd4e3c-8159-4542-b348-109f958f9a5f\") " Oct 06 10:18:56 crc kubenswrapper[4824]: I1006 10:18:56.918021 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ebdd4e3c-8159-4542-b348-109f958f9a5f-logs" (OuterVolumeSpecName: "logs") pod "ebdd4e3c-8159-4542-b348-109f958f9a5f" (UID: "ebdd4e3c-8159-4542-b348-109f958f9a5f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:18:56 crc kubenswrapper[4824]: I1006 10:18:56.954903 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ebdd4e3c-8159-4542-b348-109f958f9a5f-kube-api-access-q4gpc" (OuterVolumeSpecName: "kube-api-access-q4gpc") pod "ebdd4e3c-8159-4542-b348-109f958f9a5f" (UID: "ebdd4e3c-8159-4542-b348-109f958f9a5f"). InnerVolumeSpecName "kube-api-access-q4gpc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:18:56 crc kubenswrapper[4824]: I1006 10:18:56.993265 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ebdd4e3c-8159-4542-b348-109f958f9a5f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ebdd4e3c-8159-4542-b348-109f958f9a5f" (UID: "ebdd4e3c-8159-4542-b348-109f958f9a5f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:18:57 crc kubenswrapper[4824]: I1006 10:18:57.000160 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ebdd4e3c-8159-4542-b348-109f958f9a5f-config-data" (OuterVolumeSpecName: "config-data") pod "ebdd4e3c-8159-4542-b348-109f958f9a5f" (UID: "ebdd4e3c-8159-4542-b348-109f958f9a5f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:18:57 crc kubenswrapper[4824]: I1006 10:18:57.019413 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q4gpc\" (UniqueName: \"kubernetes.io/projected/ebdd4e3c-8159-4542-b348-109f958f9a5f-kube-api-access-q4gpc\") on node \"crc\" DevicePath \"\"" Oct 06 10:18:57 crc kubenswrapper[4824]: I1006 10:18:57.019435 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ebdd4e3c-8159-4542-b348-109f958f9a5f-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 10:18:57 crc kubenswrapper[4824]: I1006 10:18:57.019448 4824 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ebdd4e3c-8159-4542-b348-109f958f9a5f-logs\") on node \"crc\" DevicePath \"\"" Oct 06 10:18:57 crc kubenswrapper[4824]: I1006 10:18:57.019457 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ebdd4e3c-8159-4542-b348-109f958f9a5f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 10:18:57 crc kubenswrapper[4824]: I1006 10:18:57.119922 4824 generic.go:334] "Generic (PLEG): container finished" podID="ebdd4e3c-8159-4542-b348-109f958f9a5f" containerID="00f934bf252754731c575469bc520b850aa42c980a66ac9f12ed730b0da97dbb" exitCode=0 Oct 06 10:18:57 crc kubenswrapper[4824]: I1006 10:18:57.120010 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 06 10:18:57 crc kubenswrapper[4824]: I1006 10:18:57.120031 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ebdd4e3c-8159-4542-b348-109f958f9a5f","Type":"ContainerDied","Data":"00f934bf252754731c575469bc520b850aa42c980a66ac9f12ed730b0da97dbb"} Oct 06 10:18:57 crc kubenswrapper[4824]: I1006 10:18:57.120109 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ebdd4e3c-8159-4542-b348-109f958f9a5f","Type":"ContainerDied","Data":"fbc6ee08dd6e8ac29a9ac95f2d4f31b70f3204164bc091e2d021368ba9a97f59"} Oct 06 10:18:57 crc kubenswrapper[4824]: I1006 10:18:57.120132 4824 scope.go:117] "RemoveContainer" containerID="00f934bf252754731c575469bc520b850aa42c980a66ac9f12ed730b0da97dbb" Oct 06 10:18:57 crc kubenswrapper[4824]: I1006 10:18:57.123915 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3d253c1e-62ea-4559-8a58-63ddfbf02f88","Type":"ContainerStarted","Data":"3f30c1706a1fba29b1d927e96f36b1e41f6c7254fbe88b3af2d2c3a3c032287a"} Oct 06 10:18:57 crc kubenswrapper[4824]: I1006 10:18:57.124002 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3d253c1e-62ea-4559-8a58-63ddfbf02f88","Type":"ContainerStarted","Data":"9c6c907c9dd22e483a519752622d349c1247056f3b99903a47b37a098052e776"} Oct 06 10:18:57 crc kubenswrapper[4824]: I1006 10:18:57.153307 4824 scope.go:117] "RemoveContainer" containerID="63be74dc92bd5218a7077af368d5c701873e4e262c7ead1808da0dd6d507e839" Oct 06 10:18:57 crc kubenswrapper[4824]: I1006 10:18:57.182026 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 06 10:18:57 crc kubenswrapper[4824]: I1006 10:18:57.202867 4824 scope.go:117] "RemoveContainer" containerID="00f934bf252754731c575469bc520b850aa42c980a66ac9f12ed730b0da97dbb" Oct 06 10:18:57 crc kubenswrapper[4824]: I1006 10:18:57.207061 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 06 10:18:57 crc kubenswrapper[4824]: E1006 10:18:57.209448 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"00f934bf252754731c575469bc520b850aa42c980a66ac9f12ed730b0da97dbb\": container with ID starting with 00f934bf252754731c575469bc520b850aa42c980a66ac9f12ed730b0da97dbb not found: ID does not exist" containerID="00f934bf252754731c575469bc520b850aa42c980a66ac9f12ed730b0da97dbb" Oct 06 10:18:57 crc kubenswrapper[4824]: I1006 10:18:57.209482 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"00f934bf252754731c575469bc520b850aa42c980a66ac9f12ed730b0da97dbb"} err="failed to get container status \"00f934bf252754731c575469bc520b850aa42c980a66ac9f12ed730b0da97dbb\": rpc error: code = NotFound desc = could not find container \"00f934bf252754731c575469bc520b850aa42c980a66ac9f12ed730b0da97dbb\": container with ID starting with 00f934bf252754731c575469bc520b850aa42c980a66ac9f12ed730b0da97dbb not found: ID does not exist" Oct 06 10:18:57 crc kubenswrapper[4824]: I1006 10:18:57.209506 4824 scope.go:117] "RemoveContainer" containerID="63be74dc92bd5218a7077af368d5c701873e4e262c7ead1808da0dd6d507e839" Oct 06 10:18:57 crc kubenswrapper[4824]: E1006 10:18:57.210287 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"63be74dc92bd5218a7077af368d5c701873e4e262c7ead1808da0dd6d507e839\": container with ID starting with 63be74dc92bd5218a7077af368d5c701873e4e262c7ead1808da0dd6d507e839 not found: ID does not exist" containerID="63be74dc92bd5218a7077af368d5c701873e4e262c7ead1808da0dd6d507e839" Oct 06 10:18:57 crc kubenswrapper[4824]: I1006 10:18:57.210357 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"63be74dc92bd5218a7077af368d5c701873e4e262c7ead1808da0dd6d507e839"} err="failed to get container status \"63be74dc92bd5218a7077af368d5c701873e4e262c7ead1808da0dd6d507e839\": rpc error: code = NotFound desc = could not find container \"63be74dc92bd5218a7077af368d5c701873e4e262c7ead1808da0dd6d507e839\": container with ID starting with 63be74dc92bd5218a7077af368d5c701873e4e262c7ead1808da0dd6d507e839 not found: ID does not exist" Oct 06 10:18:57 crc kubenswrapper[4824]: I1006 10:18:57.219456 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 06 10:18:57 crc kubenswrapper[4824]: E1006 10:18:57.220126 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ebdd4e3c-8159-4542-b348-109f958f9a5f" containerName="nova-api-api" Oct 06 10:18:57 crc kubenswrapper[4824]: I1006 10:18:57.220152 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="ebdd4e3c-8159-4542-b348-109f958f9a5f" containerName="nova-api-api" Oct 06 10:18:57 crc kubenswrapper[4824]: E1006 10:18:57.220195 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ebdd4e3c-8159-4542-b348-109f958f9a5f" containerName="nova-api-log" Oct 06 10:18:57 crc kubenswrapper[4824]: I1006 10:18:57.220206 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="ebdd4e3c-8159-4542-b348-109f958f9a5f" containerName="nova-api-log" Oct 06 10:18:57 crc kubenswrapper[4824]: I1006 10:18:57.220439 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="ebdd4e3c-8159-4542-b348-109f958f9a5f" containerName="nova-api-api" Oct 06 10:18:57 crc kubenswrapper[4824]: I1006 10:18:57.220527 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="ebdd4e3c-8159-4542-b348-109f958f9a5f" containerName="nova-api-log" Oct 06 10:18:57 crc kubenswrapper[4824]: I1006 10:18:57.221901 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 06 10:18:57 crc kubenswrapper[4824]: I1006 10:18:57.224346 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Oct 06 10:18:57 crc kubenswrapper[4824]: I1006 10:18:57.224923 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Oct 06 10:18:57 crc kubenswrapper[4824]: I1006 10:18:57.225093 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 06 10:18:57 crc kubenswrapper[4824]: I1006 10:18:57.229354 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 06 10:18:57 crc kubenswrapper[4824]: I1006 10:18:57.288234 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ebdd4e3c-8159-4542-b348-109f958f9a5f" path="/var/lib/kubelet/pods/ebdd4e3c-8159-4542-b348-109f958f9a5f/volumes" Oct 06 10:18:57 crc kubenswrapper[4824]: I1006 10:18:57.325196 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2056be6d-e4b2-4f5e-91a7-ac5b2cb357ce-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"2056be6d-e4b2-4f5e-91a7-ac5b2cb357ce\") " pod="openstack/nova-api-0" Oct 06 10:18:57 crc kubenswrapper[4824]: I1006 10:18:57.325396 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2056be6d-e4b2-4f5e-91a7-ac5b2cb357ce-config-data\") pod \"nova-api-0\" (UID: \"2056be6d-e4b2-4f5e-91a7-ac5b2cb357ce\") " pod="openstack/nova-api-0" Oct 06 10:18:57 crc kubenswrapper[4824]: I1006 10:18:57.326141 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2056be6d-e4b2-4f5e-91a7-ac5b2cb357ce-public-tls-certs\") pod \"nova-api-0\" (UID: \"2056be6d-e4b2-4f5e-91a7-ac5b2cb357ce\") " pod="openstack/nova-api-0" Oct 06 10:18:57 crc kubenswrapper[4824]: I1006 10:18:57.326192 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2056be6d-e4b2-4f5e-91a7-ac5b2cb357ce-internal-tls-certs\") pod \"nova-api-0\" (UID: \"2056be6d-e4b2-4f5e-91a7-ac5b2cb357ce\") " pod="openstack/nova-api-0" Oct 06 10:18:57 crc kubenswrapper[4824]: I1006 10:18:57.326232 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6v8x6\" (UniqueName: \"kubernetes.io/projected/2056be6d-e4b2-4f5e-91a7-ac5b2cb357ce-kube-api-access-6v8x6\") pod \"nova-api-0\" (UID: \"2056be6d-e4b2-4f5e-91a7-ac5b2cb357ce\") " pod="openstack/nova-api-0" Oct 06 10:18:57 crc kubenswrapper[4824]: I1006 10:18:57.326297 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2056be6d-e4b2-4f5e-91a7-ac5b2cb357ce-logs\") pod \"nova-api-0\" (UID: \"2056be6d-e4b2-4f5e-91a7-ac5b2cb357ce\") " pod="openstack/nova-api-0" Oct 06 10:18:57 crc kubenswrapper[4824]: I1006 10:18:57.342391 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Oct 06 10:18:57 crc kubenswrapper[4824]: I1006 10:18:57.359878 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Oct 06 10:18:57 crc kubenswrapper[4824]: I1006 10:18:57.428495 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2056be6d-e4b2-4f5e-91a7-ac5b2cb357ce-public-tls-certs\") pod \"nova-api-0\" (UID: \"2056be6d-e4b2-4f5e-91a7-ac5b2cb357ce\") " pod="openstack/nova-api-0" Oct 06 10:18:57 crc kubenswrapper[4824]: I1006 10:18:57.428557 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2056be6d-e4b2-4f5e-91a7-ac5b2cb357ce-internal-tls-certs\") pod \"nova-api-0\" (UID: \"2056be6d-e4b2-4f5e-91a7-ac5b2cb357ce\") " pod="openstack/nova-api-0" Oct 06 10:18:57 crc kubenswrapper[4824]: I1006 10:18:57.428598 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6v8x6\" (UniqueName: \"kubernetes.io/projected/2056be6d-e4b2-4f5e-91a7-ac5b2cb357ce-kube-api-access-6v8x6\") pod \"nova-api-0\" (UID: \"2056be6d-e4b2-4f5e-91a7-ac5b2cb357ce\") " pod="openstack/nova-api-0" Oct 06 10:18:57 crc kubenswrapper[4824]: I1006 10:18:57.428636 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2056be6d-e4b2-4f5e-91a7-ac5b2cb357ce-logs\") pod \"nova-api-0\" (UID: \"2056be6d-e4b2-4f5e-91a7-ac5b2cb357ce\") " pod="openstack/nova-api-0" Oct 06 10:18:57 crc kubenswrapper[4824]: I1006 10:18:57.428735 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2056be6d-e4b2-4f5e-91a7-ac5b2cb357ce-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"2056be6d-e4b2-4f5e-91a7-ac5b2cb357ce\") " pod="openstack/nova-api-0" Oct 06 10:18:57 crc kubenswrapper[4824]: I1006 10:18:57.428761 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2056be6d-e4b2-4f5e-91a7-ac5b2cb357ce-config-data\") pod \"nova-api-0\" (UID: \"2056be6d-e4b2-4f5e-91a7-ac5b2cb357ce\") " pod="openstack/nova-api-0" Oct 06 10:18:57 crc kubenswrapper[4824]: I1006 10:18:57.431358 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2056be6d-e4b2-4f5e-91a7-ac5b2cb357ce-logs\") pod \"nova-api-0\" (UID: \"2056be6d-e4b2-4f5e-91a7-ac5b2cb357ce\") " pod="openstack/nova-api-0" Oct 06 10:18:57 crc kubenswrapper[4824]: I1006 10:18:57.435171 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2056be6d-e4b2-4f5e-91a7-ac5b2cb357ce-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"2056be6d-e4b2-4f5e-91a7-ac5b2cb357ce\") " pod="openstack/nova-api-0" Oct 06 10:18:57 crc kubenswrapper[4824]: I1006 10:18:57.435775 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2056be6d-e4b2-4f5e-91a7-ac5b2cb357ce-public-tls-certs\") pod \"nova-api-0\" (UID: \"2056be6d-e4b2-4f5e-91a7-ac5b2cb357ce\") " pod="openstack/nova-api-0" Oct 06 10:18:57 crc kubenswrapper[4824]: I1006 10:18:57.438532 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2056be6d-e4b2-4f5e-91a7-ac5b2cb357ce-internal-tls-certs\") pod \"nova-api-0\" (UID: \"2056be6d-e4b2-4f5e-91a7-ac5b2cb357ce\") " pod="openstack/nova-api-0" Oct 06 10:18:57 crc kubenswrapper[4824]: I1006 10:18:57.438628 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2056be6d-e4b2-4f5e-91a7-ac5b2cb357ce-config-data\") pod \"nova-api-0\" (UID: \"2056be6d-e4b2-4f5e-91a7-ac5b2cb357ce\") " pod="openstack/nova-api-0" Oct 06 10:18:57 crc kubenswrapper[4824]: I1006 10:18:57.451375 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6v8x6\" (UniqueName: \"kubernetes.io/projected/2056be6d-e4b2-4f5e-91a7-ac5b2cb357ce-kube-api-access-6v8x6\") pod \"nova-api-0\" (UID: \"2056be6d-e4b2-4f5e-91a7-ac5b2cb357ce\") " pod="openstack/nova-api-0" Oct 06 10:18:57 crc kubenswrapper[4824]: I1006 10:18:57.550377 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 06 10:18:58 crc kubenswrapper[4824]: I1006 10:18:58.122351 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 06 10:18:58 crc kubenswrapper[4824]: W1006 10:18:58.136907 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2056be6d_e4b2_4f5e_91a7_ac5b2cb357ce.slice/crio-c494e47b029f92169e5a3026fbf8da5e2515ce94d4201282d80a1f48afffecec WatchSource:0}: Error finding container c494e47b029f92169e5a3026fbf8da5e2515ce94d4201282d80a1f48afffecec: Status 404 returned error can't find the container with id c494e47b029f92169e5a3026fbf8da5e2515ce94d4201282d80a1f48afffecec Oct 06 10:18:58 crc kubenswrapper[4824]: I1006 10:18:58.137653 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3d253c1e-62ea-4559-8a58-63ddfbf02f88","Type":"ContainerStarted","Data":"cf85709e75fd7d0c3ff1554711bfb2a9a2714722216a83d5f5d0d6d377feb21a"} Oct 06 10:18:58 crc kubenswrapper[4824]: I1006 10:18:58.169339 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Oct 06 10:18:58 crc kubenswrapper[4824]: I1006 10:18:58.689872 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-mrzrt"] Oct 06 10:18:58 crc kubenswrapper[4824]: I1006 10:18:58.691957 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-mrzrt" Oct 06 10:18:58 crc kubenswrapper[4824]: I1006 10:18:58.695083 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Oct 06 10:18:58 crc kubenswrapper[4824]: I1006 10:18:58.695320 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Oct 06 10:18:58 crc kubenswrapper[4824]: I1006 10:18:58.702450 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-mrzrt"] Oct 06 10:18:58 crc kubenswrapper[4824]: I1006 10:18:58.765302 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tksh9\" (UniqueName: \"kubernetes.io/projected/06af5e40-91c0-45c9-9890-ffe7673be037-kube-api-access-tksh9\") pod \"nova-cell1-cell-mapping-mrzrt\" (UID: \"06af5e40-91c0-45c9-9890-ffe7673be037\") " pod="openstack/nova-cell1-cell-mapping-mrzrt" Oct 06 10:18:58 crc kubenswrapper[4824]: I1006 10:18:58.765353 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06af5e40-91c0-45c9-9890-ffe7673be037-config-data\") pod \"nova-cell1-cell-mapping-mrzrt\" (UID: \"06af5e40-91c0-45c9-9890-ffe7673be037\") " pod="openstack/nova-cell1-cell-mapping-mrzrt" Oct 06 10:18:58 crc kubenswrapper[4824]: I1006 10:18:58.765627 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06af5e40-91c0-45c9-9890-ffe7673be037-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-mrzrt\" (UID: \"06af5e40-91c0-45c9-9890-ffe7673be037\") " pod="openstack/nova-cell1-cell-mapping-mrzrt" Oct 06 10:18:58 crc kubenswrapper[4824]: I1006 10:18:58.765708 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/06af5e40-91c0-45c9-9890-ffe7673be037-scripts\") pod \"nova-cell1-cell-mapping-mrzrt\" (UID: \"06af5e40-91c0-45c9-9890-ffe7673be037\") " pod="openstack/nova-cell1-cell-mapping-mrzrt" Oct 06 10:18:58 crc kubenswrapper[4824]: I1006 10:18:58.868771 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06af5e40-91c0-45c9-9890-ffe7673be037-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-mrzrt\" (UID: \"06af5e40-91c0-45c9-9890-ffe7673be037\") " pod="openstack/nova-cell1-cell-mapping-mrzrt" Oct 06 10:18:58 crc kubenswrapper[4824]: I1006 10:18:58.868904 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/06af5e40-91c0-45c9-9890-ffe7673be037-scripts\") pod \"nova-cell1-cell-mapping-mrzrt\" (UID: \"06af5e40-91c0-45c9-9890-ffe7673be037\") " pod="openstack/nova-cell1-cell-mapping-mrzrt" Oct 06 10:18:58 crc kubenswrapper[4824]: I1006 10:18:58.869024 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tksh9\" (UniqueName: \"kubernetes.io/projected/06af5e40-91c0-45c9-9890-ffe7673be037-kube-api-access-tksh9\") pod \"nova-cell1-cell-mapping-mrzrt\" (UID: \"06af5e40-91c0-45c9-9890-ffe7673be037\") " pod="openstack/nova-cell1-cell-mapping-mrzrt" Oct 06 10:18:58 crc kubenswrapper[4824]: I1006 10:18:58.869060 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06af5e40-91c0-45c9-9890-ffe7673be037-config-data\") pod \"nova-cell1-cell-mapping-mrzrt\" (UID: \"06af5e40-91c0-45c9-9890-ffe7673be037\") " pod="openstack/nova-cell1-cell-mapping-mrzrt" Oct 06 10:18:58 crc kubenswrapper[4824]: I1006 10:18:58.874728 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06af5e40-91c0-45c9-9890-ffe7673be037-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-mrzrt\" (UID: \"06af5e40-91c0-45c9-9890-ffe7673be037\") " pod="openstack/nova-cell1-cell-mapping-mrzrt" Oct 06 10:18:58 crc kubenswrapper[4824]: I1006 10:18:58.877558 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/06af5e40-91c0-45c9-9890-ffe7673be037-scripts\") pod \"nova-cell1-cell-mapping-mrzrt\" (UID: \"06af5e40-91c0-45c9-9890-ffe7673be037\") " pod="openstack/nova-cell1-cell-mapping-mrzrt" Oct 06 10:18:58 crc kubenswrapper[4824]: I1006 10:18:58.878833 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06af5e40-91c0-45c9-9890-ffe7673be037-config-data\") pod \"nova-cell1-cell-mapping-mrzrt\" (UID: \"06af5e40-91c0-45c9-9890-ffe7673be037\") " pod="openstack/nova-cell1-cell-mapping-mrzrt" Oct 06 10:18:58 crc kubenswrapper[4824]: I1006 10:18:58.898835 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tksh9\" (UniqueName: \"kubernetes.io/projected/06af5e40-91c0-45c9-9890-ffe7673be037-kube-api-access-tksh9\") pod \"nova-cell1-cell-mapping-mrzrt\" (UID: \"06af5e40-91c0-45c9-9890-ffe7673be037\") " pod="openstack/nova-cell1-cell-mapping-mrzrt" Oct 06 10:18:59 crc kubenswrapper[4824]: I1006 10:18:59.014058 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-mrzrt" Oct 06 10:18:59 crc kubenswrapper[4824]: I1006 10:18:59.186150 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2056be6d-e4b2-4f5e-91a7-ac5b2cb357ce","Type":"ContainerStarted","Data":"ba945d970f2119d285f28c46084b1b59b9567dd3cd65490e0d58c9f33bf072a7"} Oct 06 10:18:59 crc kubenswrapper[4824]: I1006 10:18:59.186929 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2056be6d-e4b2-4f5e-91a7-ac5b2cb357ce","Type":"ContainerStarted","Data":"80156743cc95e672e11830faa0c625af7a756c3dc73c8b5f5c1110102ff75d85"} Oct 06 10:18:59 crc kubenswrapper[4824]: I1006 10:18:59.186957 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2056be6d-e4b2-4f5e-91a7-ac5b2cb357ce","Type":"ContainerStarted","Data":"c494e47b029f92169e5a3026fbf8da5e2515ce94d4201282d80a1f48afffecec"} Oct 06 10:18:59 crc kubenswrapper[4824]: I1006 10:18:59.235739 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.235700654 podStartE2EDuration="2.235700654s" podCreationTimestamp="2025-10-06 10:18:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:18:59.228077064 +0000 UTC m=+1248.592499925" watchObservedRunningTime="2025-10-06 10:18:59.235700654 +0000 UTC m=+1248.600123515" Oct 06 10:18:59 crc kubenswrapper[4824]: I1006 10:18:59.392082 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-mrzrt"] Oct 06 10:18:59 crc kubenswrapper[4824]: W1006 10:18:59.427028 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod06af5e40_91c0_45c9_9890_ffe7673be037.slice/crio-b22ab225ecdd418baf887bf650cabf7e2c5b61cbdee9c57216aea63df67f2d61 WatchSource:0}: Error finding container b22ab225ecdd418baf887bf650cabf7e2c5b61cbdee9c57216aea63df67f2d61: Status 404 returned error can't find the container with id b22ab225ecdd418baf887bf650cabf7e2c5b61cbdee9c57216aea63df67f2d61 Oct 06 10:19:00 crc kubenswrapper[4824]: I1006 10:19:00.186485 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3d253c1e-62ea-4559-8a58-63ddfbf02f88","Type":"ContainerStarted","Data":"2789458d38ec144e6b8ef6b7975ad8b6c887df4c48ca454883585e3088496351"} Oct 06 10:19:00 crc kubenswrapper[4824]: I1006 10:19:00.186602 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3d253c1e-62ea-4559-8a58-63ddfbf02f88" containerName="ceilometer-central-agent" containerID="cri-o://9c6c907c9dd22e483a519752622d349c1247056f3b99903a47b37a098052e776" gracePeriod=30 Oct 06 10:19:00 crc kubenswrapper[4824]: I1006 10:19:00.186641 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3d253c1e-62ea-4559-8a58-63ddfbf02f88" containerName="ceilometer-notification-agent" containerID="cri-o://3f30c1706a1fba29b1d927e96f36b1e41f6c7254fbe88b3af2d2c3a3c032287a" gracePeriod=30 Oct 06 10:19:00 crc kubenswrapper[4824]: I1006 10:19:00.186644 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3d253c1e-62ea-4559-8a58-63ddfbf02f88" containerName="sg-core" containerID="cri-o://cf85709e75fd7d0c3ff1554711bfb2a9a2714722216a83d5f5d0d6d377feb21a" gracePeriod=30 Oct 06 10:19:00 crc kubenswrapper[4824]: I1006 10:19:00.187050 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 06 10:19:00 crc kubenswrapper[4824]: I1006 10:19:00.186668 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3d253c1e-62ea-4559-8a58-63ddfbf02f88" containerName="proxy-httpd" containerID="cri-o://2789458d38ec144e6b8ef6b7975ad8b6c887df4c48ca454883585e3088496351" gracePeriod=30 Oct 06 10:19:00 crc kubenswrapper[4824]: I1006 10:19:00.192270 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-mrzrt" event={"ID":"06af5e40-91c0-45c9-9890-ffe7673be037","Type":"ContainerStarted","Data":"100570af227d742275ce56ad92c0165761a6c359e8215e08efe07f27d976decc"} Oct 06 10:19:00 crc kubenswrapper[4824]: I1006 10:19:00.192322 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-mrzrt" event={"ID":"06af5e40-91c0-45c9-9890-ffe7673be037","Type":"ContainerStarted","Data":"b22ab225ecdd418baf887bf650cabf7e2c5b61cbdee9c57216aea63df67f2d61"} Oct 06 10:19:00 crc kubenswrapper[4824]: I1006 10:19:00.224482 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.170437656 podStartE2EDuration="6.224451026s" podCreationTimestamp="2025-10-06 10:18:54 +0000 UTC" firstStartedPulling="2025-10-06 10:18:55.218716844 +0000 UTC m=+1244.583139705" lastFinishedPulling="2025-10-06 10:18:59.272730214 +0000 UTC m=+1248.637153075" observedRunningTime="2025-10-06 10:19:00.222075727 +0000 UTC m=+1249.586498588" watchObservedRunningTime="2025-10-06 10:19:00.224451026 +0000 UTC m=+1249.588873887" Oct 06 10:19:00 crc kubenswrapper[4824]: I1006 10:19:00.496181 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-59cf4bdb65-pk86t" Oct 06 10:19:00 crc kubenswrapper[4824]: I1006 10:19:00.521849 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-mrzrt" podStartSLOduration=2.521824645 podStartE2EDuration="2.521824645s" podCreationTimestamp="2025-10-06 10:18:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:19:00.251458564 +0000 UTC m=+1249.615881435" watchObservedRunningTime="2025-10-06 10:19:00.521824645 +0000 UTC m=+1249.886247506" Oct 06 10:19:00 crc kubenswrapper[4824]: I1006 10:19:00.585199 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-vg6jp"] Oct 06 10:19:00 crc kubenswrapper[4824]: I1006 10:19:00.585508 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-845d6d6f59-vg6jp" podUID="48fc2f94-426b-4056-88db-4445309b7e08" containerName="dnsmasq-dns" containerID="cri-o://311bea70f6af6e61a096848b0d42449d4d2c5caf0e2671a1a2dcef94afab3072" gracePeriod=10 Oct 06 10:19:01 crc kubenswrapper[4824]: I1006 10:19:01.160342 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-845d6d6f59-vg6jp" Oct 06 10:19:01 crc kubenswrapper[4824]: I1006 10:19:01.208659 4824 generic.go:334] "Generic (PLEG): container finished" podID="3d253c1e-62ea-4559-8a58-63ddfbf02f88" containerID="2789458d38ec144e6b8ef6b7975ad8b6c887df4c48ca454883585e3088496351" exitCode=0 Oct 06 10:19:01 crc kubenswrapper[4824]: I1006 10:19:01.208698 4824 generic.go:334] "Generic (PLEG): container finished" podID="3d253c1e-62ea-4559-8a58-63ddfbf02f88" containerID="cf85709e75fd7d0c3ff1554711bfb2a9a2714722216a83d5f5d0d6d377feb21a" exitCode=2 Oct 06 10:19:01 crc kubenswrapper[4824]: I1006 10:19:01.208707 4824 generic.go:334] "Generic (PLEG): container finished" podID="3d253c1e-62ea-4559-8a58-63ddfbf02f88" containerID="3f30c1706a1fba29b1d927e96f36b1e41f6c7254fbe88b3af2d2c3a3c032287a" exitCode=0 Oct 06 10:19:01 crc kubenswrapper[4824]: I1006 10:19:01.208758 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3d253c1e-62ea-4559-8a58-63ddfbf02f88","Type":"ContainerDied","Data":"2789458d38ec144e6b8ef6b7975ad8b6c887df4c48ca454883585e3088496351"} Oct 06 10:19:01 crc kubenswrapper[4824]: I1006 10:19:01.208825 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3d253c1e-62ea-4559-8a58-63ddfbf02f88","Type":"ContainerDied","Data":"cf85709e75fd7d0c3ff1554711bfb2a9a2714722216a83d5f5d0d6d377feb21a"} Oct 06 10:19:01 crc kubenswrapper[4824]: I1006 10:19:01.208845 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3d253c1e-62ea-4559-8a58-63ddfbf02f88","Type":"ContainerDied","Data":"3f30c1706a1fba29b1d927e96f36b1e41f6c7254fbe88b3af2d2c3a3c032287a"} Oct 06 10:19:01 crc kubenswrapper[4824]: I1006 10:19:01.210929 4824 generic.go:334] "Generic (PLEG): container finished" podID="48fc2f94-426b-4056-88db-4445309b7e08" containerID="311bea70f6af6e61a096848b0d42449d4d2c5caf0e2671a1a2dcef94afab3072" exitCode=0 Oct 06 10:19:01 crc kubenswrapper[4824]: I1006 10:19:01.212255 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-845d6d6f59-vg6jp" Oct 06 10:19:01 crc kubenswrapper[4824]: I1006 10:19:01.212826 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845d6d6f59-vg6jp" event={"ID":"48fc2f94-426b-4056-88db-4445309b7e08","Type":"ContainerDied","Data":"311bea70f6af6e61a096848b0d42449d4d2c5caf0e2671a1a2dcef94afab3072"} Oct 06 10:19:01 crc kubenswrapper[4824]: I1006 10:19:01.212857 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845d6d6f59-vg6jp" event={"ID":"48fc2f94-426b-4056-88db-4445309b7e08","Type":"ContainerDied","Data":"01a56efed5bc48d24af8943161b071a81b6bab26718103cafb5fa94368de1122"} Oct 06 10:19:01 crc kubenswrapper[4824]: I1006 10:19:01.212875 4824 scope.go:117] "RemoveContainer" containerID="311bea70f6af6e61a096848b0d42449d4d2c5caf0e2671a1a2dcef94afab3072" Oct 06 10:19:01 crc kubenswrapper[4824]: I1006 10:19:01.232166 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/48fc2f94-426b-4056-88db-4445309b7e08-dns-swift-storage-0\") pod \"48fc2f94-426b-4056-88db-4445309b7e08\" (UID: \"48fc2f94-426b-4056-88db-4445309b7e08\") " Oct 06 10:19:01 crc kubenswrapper[4824]: I1006 10:19:01.232258 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/48fc2f94-426b-4056-88db-4445309b7e08-ovsdbserver-nb\") pod \"48fc2f94-426b-4056-88db-4445309b7e08\" (UID: \"48fc2f94-426b-4056-88db-4445309b7e08\") " Oct 06 10:19:01 crc kubenswrapper[4824]: I1006 10:19:01.232348 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/48fc2f94-426b-4056-88db-4445309b7e08-dns-svc\") pod \"48fc2f94-426b-4056-88db-4445309b7e08\" (UID: \"48fc2f94-426b-4056-88db-4445309b7e08\") " Oct 06 10:19:01 crc kubenswrapper[4824]: I1006 10:19:01.232599 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/48fc2f94-426b-4056-88db-4445309b7e08-ovsdbserver-sb\") pod \"48fc2f94-426b-4056-88db-4445309b7e08\" (UID: \"48fc2f94-426b-4056-88db-4445309b7e08\") " Oct 06 10:19:01 crc kubenswrapper[4824]: I1006 10:19:01.232704 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4wvk8\" (UniqueName: \"kubernetes.io/projected/48fc2f94-426b-4056-88db-4445309b7e08-kube-api-access-4wvk8\") pod \"48fc2f94-426b-4056-88db-4445309b7e08\" (UID: \"48fc2f94-426b-4056-88db-4445309b7e08\") " Oct 06 10:19:01 crc kubenswrapper[4824]: I1006 10:19:01.232765 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/48fc2f94-426b-4056-88db-4445309b7e08-config\") pod \"48fc2f94-426b-4056-88db-4445309b7e08\" (UID: \"48fc2f94-426b-4056-88db-4445309b7e08\") " Oct 06 10:19:01 crc kubenswrapper[4824]: I1006 10:19:01.260231 4824 scope.go:117] "RemoveContainer" containerID="fd2ae08b0d41823968cf87f77265c104dcc97a748c00a0eeaf761cf405f8abe5" Oct 06 10:19:01 crc kubenswrapper[4824]: I1006 10:19:01.268172 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/48fc2f94-426b-4056-88db-4445309b7e08-kube-api-access-4wvk8" (OuterVolumeSpecName: "kube-api-access-4wvk8") pod "48fc2f94-426b-4056-88db-4445309b7e08" (UID: "48fc2f94-426b-4056-88db-4445309b7e08"). InnerVolumeSpecName "kube-api-access-4wvk8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:19:01 crc kubenswrapper[4824]: I1006 10:19:01.318443 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/48fc2f94-426b-4056-88db-4445309b7e08-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "48fc2f94-426b-4056-88db-4445309b7e08" (UID: "48fc2f94-426b-4056-88db-4445309b7e08"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:19:01 crc kubenswrapper[4824]: I1006 10:19:01.326289 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/48fc2f94-426b-4056-88db-4445309b7e08-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "48fc2f94-426b-4056-88db-4445309b7e08" (UID: "48fc2f94-426b-4056-88db-4445309b7e08"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:19:01 crc kubenswrapper[4824]: I1006 10:19:01.327746 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/48fc2f94-426b-4056-88db-4445309b7e08-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "48fc2f94-426b-4056-88db-4445309b7e08" (UID: "48fc2f94-426b-4056-88db-4445309b7e08"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:19:01 crc kubenswrapper[4824]: I1006 10:19:01.336717 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4wvk8\" (UniqueName: \"kubernetes.io/projected/48fc2f94-426b-4056-88db-4445309b7e08-kube-api-access-4wvk8\") on node \"crc\" DevicePath \"\"" Oct 06 10:19:01 crc kubenswrapper[4824]: I1006 10:19:01.336790 4824 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/48fc2f94-426b-4056-88db-4445309b7e08-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 06 10:19:01 crc kubenswrapper[4824]: I1006 10:19:01.336804 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/48fc2f94-426b-4056-88db-4445309b7e08-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 06 10:19:01 crc kubenswrapper[4824]: I1006 10:19:01.336817 4824 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/48fc2f94-426b-4056-88db-4445309b7e08-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 06 10:19:01 crc kubenswrapper[4824]: I1006 10:19:01.340032 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/48fc2f94-426b-4056-88db-4445309b7e08-config" (OuterVolumeSpecName: "config") pod "48fc2f94-426b-4056-88db-4445309b7e08" (UID: "48fc2f94-426b-4056-88db-4445309b7e08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:19:01 crc kubenswrapper[4824]: I1006 10:19:01.354882 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/48fc2f94-426b-4056-88db-4445309b7e08-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "48fc2f94-426b-4056-88db-4445309b7e08" (UID: "48fc2f94-426b-4056-88db-4445309b7e08"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:19:01 crc kubenswrapper[4824]: I1006 10:19:01.431560 4824 scope.go:117] "RemoveContainer" containerID="311bea70f6af6e61a096848b0d42449d4d2c5caf0e2671a1a2dcef94afab3072" Oct 06 10:19:01 crc kubenswrapper[4824]: E1006 10:19:01.432643 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"311bea70f6af6e61a096848b0d42449d4d2c5caf0e2671a1a2dcef94afab3072\": container with ID starting with 311bea70f6af6e61a096848b0d42449d4d2c5caf0e2671a1a2dcef94afab3072 not found: ID does not exist" containerID="311bea70f6af6e61a096848b0d42449d4d2c5caf0e2671a1a2dcef94afab3072" Oct 06 10:19:01 crc kubenswrapper[4824]: I1006 10:19:01.432719 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"311bea70f6af6e61a096848b0d42449d4d2c5caf0e2671a1a2dcef94afab3072"} err="failed to get container status \"311bea70f6af6e61a096848b0d42449d4d2c5caf0e2671a1a2dcef94afab3072\": rpc error: code = NotFound desc = could not find container \"311bea70f6af6e61a096848b0d42449d4d2c5caf0e2671a1a2dcef94afab3072\": container with ID starting with 311bea70f6af6e61a096848b0d42449d4d2c5caf0e2671a1a2dcef94afab3072 not found: ID does not exist" Oct 06 10:19:01 crc kubenswrapper[4824]: I1006 10:19:01.432765 4824 scope.go:117] "RemoveContainer" containerID="fd2ae08b0d41823968cf87f77265c104dcc97a748c00a0eeaf761cf405f8abe5" Oct 06 10:19:01 crc kubenswrapper[4824]: E1006 10:19:01.433456 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fd2ae08b0d41823968cf87f77265c104dcc97a748c00a0eeaf761cf405f8abe5\": container with ID starting with fd2ae08b0d41823968cf87f77265c104dcc97a748c00a0eeaf761cf405f8abe5 not found: ID does not exist" containerID="fd2ae08b0d41823968cf87f77265c104dcc97a748c00a0eeaf761cf405f8abe5" Oct 06 10:19:01 crc kubenswrapper[4824]: I1006 10:19:01.433483 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fd2ae08b0d41823968cf87f77265c104dcc97a748c00a0eeaf761cf405f8abe5"} err="failed to get container status \"fd2ae08b0d41823968cf87f77265c104dcc97a748c00a0eeaf761cf405f8abe5\": rpc error: code = NotFound desc = could not find container \"fd2ae08b0d41823968cf87f77265c104dcc97a748c00a0eeaf761cf405f8abe5\": container with ID starting with fd2ae08b0d41823968cf87f77265c104dcc97a748c00a0eeaf761cf405f8abe5 not found: ID does not exist" Oct 06 10:19:01 crc kubenswrapper[4824]: I1006 10:19:01.439156 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/48fc2f94-426b-4056-88db-4445309b7e08-config\") on node \"crc\" DevicePath \"\"" Oct 06 10:19:01 crc kubenswrapper[4824]: I1006 10:19:01.439203 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/48fc2f94-426b-4056-88db-4445309b7e08-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 06 10:19:01 crc kubenswrapper[4824]: I1006 10:19:01.550704 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-vg6jp"] Oct 06 10:19:01 crc kubenswrapper[4824]: I1006 10:19:01.558760 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-vg6jp"] Oct 06 10:19:02 crc kubenswrapper[4824]: I1006 10:19:02.759660 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 06 10:19:02 crc kubenswrapper[4824]: I1006 10:19:02.870971 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3d253c1e-62ea-4559-8a58-63ddfbf02f88-config-data\") pod \"3d253c1e-62ea-4559-8a58-63ddfbf02f88\" (UID: \"3d253c1e-62ea-4559-8a58-63ddfbf02f88\") " Oct 06 10:19:02 crc kubenswrapper[4824]: I1006 10:19:02.871245 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3d253c1e-62ea-4559-8a58-63ddfbf02f88-log-httpd\") pod \"3d253c1e-62ea-4559-8a58-63ddfbf02f88\" (UID: \"3d253c1e-62ea-4559-8a58-63ddfbf02f88\") " Oct 06 10:19:02 crc kubenswrapper[4824]: I1006 10:19:02.871287 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3d253c1e-62ea-4559-8a58-63ddfbf02f88-scripts\") pod \"3d253c1e-62ea-4559-8a58-63ddfbf02f88\" (UID: \"3d253c1e-62ea-4559-8a58-63ddfbf02f88\") " Oct 06 10:19:02 crc kubenswrapper[4824]: I1006 10:19:02.871368 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d253c1e-62ea-4559-8a58-63ddfbf02f88-combined-ca-bundle\") pod \"3d253c1e-62ea-4559-8a58-63ddfbf02f88\" (UID: \"3d253c1e-62ea-4559-8a58-63ddfbf02f88\") " Oct 06 10:19:02 crc kubenswrapper[4824]: I1006 10:19:02.871448 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-47dp2\" (UniqueName: \"kubernetes.io/projected/3d253c1e-62ea-4559-8a58-63ddfbf02f88-kube-api-access-47dp2\") pod \"3d253c1e-62ea-4559-8a58-63ddfbf02f88\" (UID: \"3d253c1e-62ea-4559-8a58-63ddfbf02f88\") " Oct 06 10:19:02 crc kubenswrapper[4824]: I1006 10:19:02.871491 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/3d253c1e-62ea-4559-8a58-63ddfbf02f88-ceilometer-tls-certs\") pod \"3d253c1e-62ea-4559-8a58-63ddfbf02f88\" (UID: \"3d253c1e-62ea-4559-8a58-63ddfbf02f88\") " Oct 06 10:19:02 crc kubenswrapper[4824]: I1006 10:19:02.871533 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3d253c1e-62ea-4559-8a58-63ddfbf02f88-run-httpd\") pod \"3d253c1e-62ea-4559-8a58-63ddfbf02f88\" (UID: \"3d253c1e-62ea-4559-8a58-63ddfbf02f88\") " Oct 06 10:19:02 crc kubenswrapper[4824]: I1006 10:19:02.871633 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3d253c1e-62ea-4559-8a58-63ddfbf02f88-sg-core-conf-yaml\") pod \"3d253c1e-62ea-4559-8a58-63ddfbf02f88\" (UID: \"3d253c1e-62ea-4559-8a58-63ddfbf02f88\") " Oct 06 10:19:02 crc kubenswrapper[4824]: I1006 10:19:02.872008 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3d253c1e-62ea-4559-8a58-63ddfbf02f88-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "3d253c1e-62ea-4559-8a58-63ddfbf02f88" (UID: "3d253c1e-62ea-4559-8a58-63ddfbf02f88"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:19:02 crc kubenswrapper[4824]: I1006 10:19:02.872828 4824 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3d253c1e-62ea-4559-8a58-63ddfbf02f88-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 06 10:19:02 crc kubenswrapper[4824]: I1006 10:19:02.873934 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3d253c1e-62ea-4559-8a58-63ddfbf02f88-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "3d253c1e-62ea-4559-8a58-63ddfbf02f88" (UID: "3d253c1e-62ea-4559-8a58-63ddfbf02f88"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:19:02 crc kubenswrapper[4824]: I1006 10:19:02.896582 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3d253c1e-62ea-4559-8a58-63ddfbf02f88-scripts" (OuterVolumeSpecName: "scripts") pod "3d253c1e-62ea-4559-8a58-63ddfbf02f88" (UID: "3d253c1e-62ea-4559-8a58-63ddfbf02f88"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:19:02 crc kubenswrapper[4824]: I1006 10:19:02.924352 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3d253c1e-62ea-4559-8a58-63ddfbf02f88-kube-api-access-47dp2" (OuterVolumeSpecName: "kube-api-access-47dp2") pod "3d253c1e-62ea-4559-8a58-63ddfbf02f88" (UID: "3d253c1e-62ea-4559-8a58-63ddfbf02f88"). InnerVolumeSpecName "kube-api-access-47dp2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:19:02 crc kubenswrapper[4824]: I1006 10:19:02.938076 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3d253c1e-62ea-4559-8a58-63ddfbf02f88-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "3d253c1e-62ea-4559-8a58-63ddfbf02f88" (UID: "3d253c1e-62ea-4559-8a58-63ddfbf02f88"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:19:02 crc kubenswrapper[4824]: I1006 10:19:02.975942 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-47dp2\" (UniqueName: \"kubernetes.io/projected/3d253c1e-62ea-4559-8a58-63ddfbf02f88-kube-api-access-47dp2\") on node \"crc\" DevicePath \"\"" Oct 06 10:19:02 crc kubenswrapper[4824]: I1006 10:19:02.976013 4824 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3d253c1e-62ea-4559-8a58-63ddfbf02f88-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 06 10:19:02 crc kubenswrapper[4824]: I1006 10:19:02.976029 4824 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3d253c1e-62ea-4559-8a58-63ddfbf02f88-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 06 10:19:02 crc kubenswrapper[4824]: I1006 10:19:02.976042 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3d253c1e-62ea-4559-8a58-63ddfbf02f88-scripts\") on node \"crc\" DevicePath \"\"" Oct 06 10:19:02 crc kubenswrapper[4824]: I1006 10:19:02.992784 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3d253c1e-62ea-4559-8a58-63ddfbf02f88-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "3d253c1e-62ea-4559-8a58-63ddfbf02f88" (UID: "3d253c1e-62ea-4559-8a58-63ddfbf02f88"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:19:03 crc kubenswrapper[4824]: I1006 10:19:03.013646 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3d253c1e-62ea-4559-8a58-63ddfbf02f88-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3d253c1e-62ea-4559-8a58-63ddfbf02f88" (UID: "3d253c1e-62ea-4559-8a58-63ddfbf02f88"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:19:03 crc kubenswrapper[4824]: I1006 10:19:03.053308 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3d253c1e-62ea-4559-8a58-63ddfbf02f88-config-data" (OuterVolumeSpecName: "config-data") pod "3d253c1e-62ea-4559-8a58-63ddfbf02f88" (UID: "3d253c1e-62ea-4559-8a58-63ddfbf02f88"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:19:03 crc kubenswrapper[4824]: I1006 10:19:03.078341 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d253c1e-62ea-4559-8a58-63ddfbf02f88-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 10:19:03 crc kubenswrapper[4824]: I1006 10:19:03.078397 4824 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/3d253c1e-62ea-4559-8a58-63ddfbf02f88-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 06 10:19:03 crc kubenswrapper[4824]: I1006 10:19:03.078409 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3d253c1e-62ea-4559-8a58-63ddfbf02f88-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 10:19:03 crc kubenswrapper[4824]: I1006 10:19:03.248352 4824 generic.go:334] "Generic (PLEG): container finished" podID="3d253c1e-62ea-4559-8a58-63ddfbf02f88" containerID="9c6c907c9dd22e483a519752622d349c1247056f3b99903a47b37a098052e776" exitCode=0 Oct 06 10:19:03 crc kubenswrapper[4824]: I1006 10:19:03.248420 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3d253c1e-62ea-4559-8a58-63ddfbf02f88","Type":"ContainerDied","Data":"9c6c907c9dd22e483a519752622d349c1247056f3b99903a47b37a098052e776"} Oct 06 10:19:03 crc kubenswrapper[4824]: I1006 10:19:03.248478 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3d253c1e-62ea-4559-8a58-63ddfbf02f88","Type":"ContainerDied","Data":"33548dcb6b502aae91b4c41dde158ef8b72f252b9d2fb4e701d7f99b84e3f638"} Oct 06 10:19:03 crc kubenswrapper[4824]: I1006 10:19:03.248493 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 06 10:19:03 crc kubenswrapper[4824]: I1006 10:19:03.248511 4824 scope.go:117] "RemoveContainer" containerID="2789458d38ec144e6b8ef6b7975ad8b6c887df4c48ca454883585e3088496351" Oct 06 10:19:03 crc kubenswrapper[4824]: I1006 10:19:03.272508 4824 scope.go:117] "RemoveContainer" containerID="cf85709e75fd7d0c3ff1554711bfb2a9a2714722216a83d5f5d0d6d377feb21a" Oct 06 10:19:03 crc kubenswrapper[4824]: I1006 10:19:03.317718 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="48fc2f94-426b-4056-88db-4445309b7e08" path="/var/lib/kubelet/pods/48fc2f94-426b-4056-88db-4445309b7e08/volumes" Oct 06 10:19:03 crc kubenswrapper[4824]: I1006 10:19:03.318753 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 06 10:19:03 crc kubenswrapper[4824]: I1006 10:19:03.318793 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 06 10:19:03 crc kubenswrapper[4824]: I1006 10:19:03.373542 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 06 10:19:03 crc kubenswrapper[4824]: E1006 10:19:03.374520 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48fc2f94-426b-4056-88db-4445309b7e08" containerName="dnsmasq-dns" Oct 06 10:19:03 crc kubenswrapper[4824]: I1006 10:19:03.375052 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="48fc2f94-426b-4056-88db-4445309b7e08" containerName="dnsmasq-dns" Oct 06 10:19:03 crc kubenswrapper[4824]: E1006 10:19:03.375150 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d253c1e-62ea-4559-8a58-63ddfbf02f88" containerName="sg-core" Oct 06 10:19:03 crc kubenswrapper[4824]: I1006 10:19:03.375204 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d253c1e-62ea-4559-8a58-63ddfbf02f88" containerName="sg-core" Oct 06 10:19:03 crc kubenswrapper[4824]: E1006 10:19:03.375299 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d253c1e-62ea-4559-8a58-63ddfbf02f88" containerName="ceilometer-central-agent" Oct 06 10:19:03 crc kubenswrapper[4824]: I1006 10:19:03.375355 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d253c1e-62ea-4559-8a58-63ddfbf02f88" containerName="ceilometer-central-agent" Oct 06 10:19:03 crc kubenswrapper[4824]: E1006 10:19:03.375408 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d253c1e-62ea-4559-8a58-63ddfbf02f88" containerName="ceilometer-notification-agent" Oct 06 10:19:03 crc kubenswrapper[4824]: I1006 10:19:03.375457 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d253c1e-62ea-4559-8a58-63ddfbf02f88" containerName="ceilometer-notification-agent" Oct 06 10:19:03 crc kubenswrapper[4824]: E1006 10:19:03.375533 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d253c1e-62ea-4559-8a58-63ddfbf02f88" containerName="proxy-httpd" Oct 06 10:19:03 crc kubenswrapper[4824]: I1006 10:19:03.375582 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d253c1e-62ea-4559-8a58-63ddfbf02f88" containerName="proxy-httpd" Oct 06 10:19:03 crc kubenswrapper[4824]: E1006 10:19:03.375656 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48fc2f94-426b-4056-88db-4445309b7e08" containerName="init" Oct 06 10:19:03 crc kubenswrapper[4824]: I1006 10:19:03.375776 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="48fc2f94-426b-4056-88db-4445309b7e08" containerName="init" Oct 06 10:19:03 crc kubenswrapper[4824]: I1006 10:19:03.376073 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="3d253c1e-62ea-4559-8a58-63ddfbf02f88" containerName="ceilometer-notification-agent" Oct 06 10:19:03 crc kubenswrapper[4824]: I1006 10:19:03.376153 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="3d253c1e-62ea-4559-8a58-63ddfbf02f88" containerName="sg-core" Oct 06 10:19:03 crc kubenswrapper[4824]: I1006 10:19:03.376228 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="3d253c1e-62ea-4559-8a58-63ddfbf02f88" containerName="proxy-httpd" Oct 06 10:19:03 crc kubenswrapper[4824]: I1006 10:19:03.376298 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="48fc2f94-426b-4056-88db-4445309b7e08" containerName="dnsmasq-dns" Oct 06 10:19:03 crc kubenswrapper[4824]: I1006 10:19:03.376364 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="3d253c1e-62ea-4559-8a58-63ddfbf02f88" containerName="ceilometer-central-agent" Oct 06 10:19:03 crc kubenswrapper[4824]: I1006 10:19:03.374694 4824 scope.go:117] "RemoveContainer" containerID="3f30c1706a1fba29b1d927e96f36b1e41f6c7254fbe88b3af2d2c3a3c032287a" Oct 06 10:19:03 crc kubenswrapper[4824]: I1006 10:19:03.379772 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 06 10:19:03 crc kubenswrapper[4824]: I1006 10:19:03.384379 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 06 10:19:03 crc kubenswrapper[4824]: I1006 10:19:03.384617 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 06 10:19:03 crc kubenswrapper[4824]: I1006 10:19:03.384850 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 06 10:19:03 crc kubenswrapper[4824]: I1006 10:19:03.393622 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 06 10:19:03 crc kubenswrapper[4824]: I1006 10:19:03.420964 4824 scope.go:117] "RemoveContainer" containerID="9c6c907c9dd22e483a519752622d349c1247056f3b99903a47b37a098052e776" Oct 06 10:19:03 crc kubenswrapper[4824]: I1006 10:19:03.453299 4824 scope.go:117] "RemoveContainer" containerID="2789458d38ec144e6b8ef6b7975ad8b6c887df4c48ca454883585e3088496351" Oct 06 10:19:03 crc kubenswrapper[4824]: E1006 10:19:03.457363 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2789458d38ec144e6b8ef6b7975ad8b6c887df4c48ca454883585e3088496351\": container with ID starting with 2789458d38ec144e6b8ef6b7975ad8b6c887df4c48ca454883585e3088496351 not found: ID does not exist" containerID="2789458d38ec144e6b8ef6b7975ad8b6c887df4c48ca454883585e3088496351" Oct 06 10:19:03 crc kubenswrapper[4824]: I1006 10:19:03.457500 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2789458d38ec144e6b8ef6b7975ad8b6c887df4c48ca454883585e3088496351"} err="failed to get container status \"2789458d38ec144e6b8ef6b7975ad8b6c887df4c48ca454883585e3088496351\": rpc error: code = NotFound desc = could not find container \"2789458d38ec144e6b8ef6b7975ad8b6c887df4c48ca454883585e3088496351\": container with ID starting with 2789458d38ec144e6b8ef6b7975ad8b6c887df4c48ca454883585e3088496351 not found: ID does not exist" Oct 06 10:19:03 crc kubenswrapper[4824]: I1006 10:19:03.457611 4824 scope.go:117] "RemoveContainer" containerID="cf85709e75fd7d0c3ff1554711bfb2a9a2714722216a83d5f5d0d6d377feb21a" Oct 06 10:19:03 crc kubenswrapper[4824]: E1006 10:19:03.458138 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cf85709e75fd7d0c3ff1554711bfb2a9a2714722216a83d5f5d0d6d377feb21a\": container with ID starting with cf85709e75fd7d0c3ff1554711bfb2a9a2714722216a83d5f5d0d6d377feb21a not found: ID does not exist" containerID="cf85709e75fd7d0c3ff1554711bfb2a9a2714722216a83d5f5d0d6d377feb21a" Oct 06 10:19:03 crc kubenswrapper[4824]: I1006 10:19:03.458362 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cf85709e75fd7d0c3ff1554711bfb2a9a2714722216a83d5f5d0d6d377feb21a"} err="failed to get container status \"cf85709e75fd7d0c3ff1554711bfb2a9a2714722216a83d5f5d0d6d377feb21a\": rpc error: code = NotFound desc = could not find container \"cf85709e75fd7d0c3ff1554711bfb2a9a2714722216a83d5f5d0d6d377feb21a\": container with ID starting with cf85709e75fd7d0c3ff1554711bfb2a9a2714722216a83d5f5d0d6d377feb21a not found: ID does not exist" Oct 06 10:19:03 crc kubenswrapper[4824]: I1006 10:19:03.458511 4824 scope.go:117] "RemoveContainer" containerID="3f30c1706a1fba29b1d927e96f36b1e41f6c7254fbe88b3af2d2c3a3c032287a" Oct 06 10:19:03 crc kubenswrapper[4824]: E1006 10:19:03.459085 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3f30c1706a1fba29b1d927e96f36b1e41f6c7254fbe88b3af2d2c3a3c032287a\": container with ID starting with 3f30c1706a1fba29b1d927e96f36b1e41f6c7254fbe88b3af2d2c3a3c032287a not found: ID does not exist" containerID="3f30c1706a1fba29b1d927e96f36b1e41f6c7254fbe88b3af2d2c3a3c032287a" Oct 06 10:19:03 crc kubenswrapper[4824]: I1006 10:19:03.459158 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3f30c1706a1fba29b1d927e96f36b1e41f6c7254fbe88b3af2d2c3a3c032287a"} err="failed to get container status \"3f30c1706a1fba29b1d927e96f36b1e41f6c7254fbe88b3af2d2c3a3c032287a\": rpc error: code = NotFound desc = could not find container \"3f30c1706a1fba29b1d927e96f36b1e41f6c7254fbe88b3af2d2c3a3c032287a\": container with ID starting with 3f30c1706a1fba29b1d927e96f36b1e41f6c7254fbe88b3af2d2c3a3c032287a not found: ID does not exist" Oct 06 10:19:03 crc kubenswrapper[4824]: I1006 10:19:03.459201 4824 scope.go:117] "RemoveContainer" containerID="9c6c907c9dd22e483a519752622d349c1247056f3b99903a47b37a098052e776" Oct 06 10:19:03 crc kubenswrapper[4824]: E1006 10:19:03.459720 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9c6c907c9dd22e483a519752622d349c1247056f3b99903a47b37a098052e776\": container with ID starting with 9c6c907c9dd22e483a519752622d349c1247056f3b99903a47b37a098052e776 not found: ID does not exist" containerID="9c6c907c9dd22e483a519752622d349c1247056f3b99903a47b37a098052e776" Oct 06 10:19:03 crc kubenswrapper[4824]: I1006 10:19:03.459813 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9c6c907c9dd22e483a519752622d349c1247056f3b99903a47b37a098052e776"} err="failed to get container status \"9c6c907c9dd22e483a519752622d349c1247056f3b99903a47b37a098052e776\": rpc error: code = NotFound desc = could not find container \"9c6c907c9dd22e483a519752622d349c1247056f3b99903a47b37a098052e776\": container with ID starting with 9c6c907c9dd22e483a519752622d349c1247056f3b99903a47b37a098052e776 not found: ID does not exist" Oct 06 10:19:03 crc kubenswrapper[4824]: I1006 10:19:03.495733 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/369bbc94-27b7-4016-9a43-1e20b28f4323-config-data\") pod \"ceilometer-0\" (UID: \"369bbc94-27b7-4016-9a43-1e20b28f4323\") " pod="openstack/ceilometer-0" Oct 06 10:19:03 crc kubenswrapper[4824]: I1006 10:19:03.497294 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/369bbc94-27b7-4016-9a43-1e20b28f4323-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"369bbc94-27b7-4016-9a43-1e20b28f4323\") " pod="openstack/ceilometer-0" Oct 06 10:19:03 crc kubenswrapper[4824]: I1006 10:19:03.497436 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/369bbc94-27b7-4016-9a43-1e20b28f4323-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"369bbc94-27b7-4016-9a43-1e20b28f4323\") " pod="openstack/ceilometer-0" Oct 06 10:19:03 crc kubenswrapper[4824]: I1006 10:19:03.497579 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/369bbc94-27b7-4016-9a43-1e20b28f4323-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"369bbc94-27b7-4016-9a43-1e20b28f4323\") " pod="openstack/ceilometer-0" Oct 06 10:19:03 crc kubenswrapper[4824]: I1006 10:19:03.497669 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/369bbc94-27b7-4016-9a43-1e20b28f4323-log-httpd\") pod \"ceilometer-0\" (UID: \"369bbc94-27b7-4016-9a43-1e20b28f4323\") " pod="openstack/ceilometer-0" Oct 06 10:19:03 crc kubenswrapper[4824]: I1006 10:19:03.497758 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/369bbc94-27b7-4016-9a43-1e20b28f4323-scripts\") pod \"ceilometer-0\" (UID: \"369bbc94-27b7-4016-9a43-1e20b28f4323\") " pod="openstack/ceilometer-0" Oct 06 10:19:03 crc kubenswrapper[4824]: I1006 10:19:03.497896 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bhbzq\" (UniqueName: \"kubernetes.io/projected/369bbc94-27b7-4016-9a43-1e20b28f4323-kube-api-access-bhbzq\") pod \"ceilometer-0\" (UID: \"369bbc94-27b7-4016-9a43-1e20b28f4323\") " pod="openstack/ceilometer-0" Oct 06 10:19:03 crc kubenswrapper[4824]: I1006 10:19:03.498092 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/369bbc94-27b7-4016-9a43-1e20b28f4323-run-httpd\") pod \"ceilometer-0\" (UID: \"369bbc94-27b7-4016-9a43-1e20b28f4323\") " pod="openstack/ceilometer-0" Oct 06 10:19:03 crc kubenswrapper[4824]: I1006 10:19:03.600562 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/369bbc94-27b7-4016-9a43-1e20b28f4323-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"369bbc94-27b7-4016-9a43-1e20b28f4323\") " pod="openstack/ceilometer-0" Oct 06 10:19:03 crc kubenswrapper[4824]: I1006 10:19:03.600643 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/369bbc94-27b7-4016-9a43-1e20b28f4323-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"369bbc94-27b7-4016-9a43-1e20b28f4323\") " pod="openstack/ceilometer-0" Oct 06 10:19:03 crc kubenswrapper[4824]: I1006 10:19:03.600684 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/369bbc94-27b7-4016-9a43-1e20b28f4323-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"369bbc94-27b7-4016-9a43-1e20b28f4323\") " pod="openstack/ceilometer-0" Oct 06 10:19:03 crc kubenswrapper[4824]: I1006 10:19:03.600747 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/369bbc94-27b7-4016-9a43-1e20b28f4323-log-httpd\") pod \"ceilometer-0\" (UID: \"369bbc94-27b7-4016-9a43-1e20b28f4323\") " pod="openstack/ceilometer-0" Oct 06 10:19:03 crc kubenswrapper[4824]: I1006 10:19:03.600772 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/369bbc94-27b7-4016-9a43-1e20b28f4323-scripts\") pod \"ceilometer-0\" (UID: \"369bbc94-27b7-4016-9a43-1e20b28f4323\") " pod="openstack/ceilometer-0" Oct 06 10:19:03 crc kubenswrapper[4824]: I1006 10:19:03.600828 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bhbzq\" (UniqueName: \"kubernetes.io/projected/369bbc94-27b7-4016-9a43-1e20b28f4323-kube-api-access-bhbzq\") pod \"ceilometer-0\" (UID: \"369bbc94-27b7-4016-9a43-1e20b28f4323\") " pod="openstack/ceilometer-0" Oct 06 10:19:03 crc kubenswrapper[4824]: I1006 10:19:03.600865 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/369bbc94-27b7-4016-9a43-1e20b28f4323-run-httpd\") pod \"ceilometer-0\" (UID: \"369bbc94-27b7-4016-9a43-1e20b28f4323\") " pod="openstack/ceilometer-0" Oct 06 10:19:03 crc kubenswrapper[4824]: I1006 10:19:03.600997 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/369bbc94-27b7-4016-9a43-1e20b28f4323-config-data\") pod \"ceilometer-0\" (UID: \"369bbc94-27b7-4016-9a43-1e20b28f4323\") " pod="openstack/ceilometer-0" Oct 06 10:19:03 crc kubenswrapper[4824]: I1006 10:19:03.601934 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/369bbc94-27b7-4016-9a43-1e20b28f4323-log-httpd\") pod \"ceilometer-0\" (UID: \"369bbc94-27b7-4016-9a43-1e20b28f4323\") " pod="openstack/ceilometer-0" Oct 06 10:19:03 crc kubenswrapper[4824]: I1006 10:19:03.602464 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/369bbc94-27b7-4016-9a43-1e20b28f4323-run-httpd\") pod \"ceilometer-0\" (UID: \"369bbc94-27b7-4016-9a43-1e20b28f4323\") " pod="openstack/ceilometer-0" Oct 06 10:19:03 crc kubenswrapper[4824]: I1006 10:19:03.608457 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/369bbc94-27b7-4016-9a43-1e20b28f4323-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"369bbc94-27b7-4016-9a43-1e20b28f4323\") " pod="openstack/ceilometer-0" Oct 06 10:19:03 crc kubenswrapper[4824]: I1006 10:19:03.609536 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/369bbc94-27b7-4016-9a43-1e20b28f4323-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"369bbc94-27b7-4016-9a43-1e20b28f4323\") " pod="openstack/ceilometer-0" Oct 06 10:19:03 crc kubenswrapper[4824]: I1006 10:19:03.611537 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/369bbc94-27b7-4016-9a43-1e20b28f4323-scripts\") pod \"ceilometer-0\" (UID: \"369bbc94-27b7-4016-9a43-1e20b28f4323\") " pod="openstack/ceilometer-0" Oct 06 10:19:03 crc kubenswrapper[4824]: I1006 10:19:03.619437 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/369bbc94-27b7-4016-9a43-1e20b28f4323-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"369bbc94-27b7-4016-9a43-1e20b28f4323\") " pod="openstack/ceilometer-0" Oct 06 10:19:03 crc kubenswrapper[4824]: I1006 10:19:03.619595 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/369bbc94-27b7-4016-9a43-1e20b28f4323-config-data\") pod \"ceilometer-0\" (UID: \"369bbc94-27b7-4016-9a43-1e20b28f4323\") " pod="openstack/ceilometer-0" Oct 06 10:19:03 crc kubenswrapper[4824]: I1006 10:19:03.629577 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bhbzq\" (UniqueName: \"kubernetes.io/projected/369bbc94-27b7-4016-9a43-1e20b28f4323-kube-api-access-bhbzq\") pod \"ceilometer-0\" (UID: \"369bbc94-27b7-4016-9a43-1e20b28f4323\") " pod="openstack/ceilometer-0" Oct 06 10:19:03 crc kubenswrapper[4824]: I1006 10:19:03.715113 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 06 10:19:04 crc kubenswrapper[4824]: I1006 10:19:04.207575 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 06 10:19:04 crc kubenswrapper[4824]: I1006 10:19:04.261172 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"369bbc94-27b7-4016-9a43-1e20b28f4323","Type":"ContainerStarted","Data":"673c60b9a61a67f679841672f27f7094c3172ab7c8f367662863b071d9ce364e"} Oct 06 10:19:05 crc kubenswrapper[4824]: I1006 10:19:05.340917 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3d253c1e-62ea-4559-8a58-63ddfbf02f88" path="/var/lib/kubelet/pods/3d253c1e-62ea-4559-8a58-63ddfbf02f88/volumes" Oct 06 10:19:05 crc kubenswrapper[4824]: I1006 10:19:05.342185 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"369bbc94-27b7-4016-9a43-1e20b28f4323","Type":"ContainerStarted","Data":"8b39fa35275ed6b16d9a24efc3d824b9b28c95bc6c4e5297b3eae897931e6579"} Oct 06 10:19:05 crc kubenswrapper[4824]: I1006 10:19:05.970684 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-845d6d6f59-vg6jp" podUID="48fc2f94-426b-4056-88db-4445309b7e08" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.190:5353: i/o timeout" Oct 06 10:19:06 crc kubenswrapper[4824]: I1006 10:19:06.368614 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"369bbc94-27b7-4016-9a43-1e20b28f4323","Type":"ContainerStarted","Data":"55d0020ccd4f644734a041f118ccac8b66bd67075445321910792b5b364c0772"} Oct 06 10:19:06 crc kubenswrapper[4824]: I1006 10:19:06.377604 4824 generic.go:334] "Generic (PLEG): container finished" podID="06af5e40-91c0-45c9-9890-ffe7673be037" containerID="100570af227d742275ce56ad92c0165761a6c359e8215e08efe07f27d976decc" exitCode=0 Oct 06 10:19:06 crc kubenswrapper[4824]: I1006 10:19:06.377682 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-mrzrt" event={"ID":"06af5e40-91c0-45c9-9890-ffe7673be037","Type":"ContainerDied","Data":"100570af227d742275ce56ad92c0165761a6c359e8215e08efe07f27d976decc"} Oct 06 10:19:07 crc kubenswrapper[4824]: I1006 10:19:07.394449 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"369bbc94-27b7-4016-9a43-1e20b28f4323","Type":"ContainerStarted","Data":"d3ea50a25bcfcf39da6392031781864f6791f6e2e03e3b109b35029b89051ca3"} Oct 06 10:19:07 crc kubenswrapper[4824]: I1006 10:19:07.551706 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 06 10:19:07 crc kubenswrapper[4824]: I1006 10:19:07.551778 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 06 10:19:08 crc kubenswrapper[4824]: I1006 10:19:08.043613 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-mrzrt" Oct 06 10:19:08 crc kubenswrapper[4824]: I1006 10:19:08.110853 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06af5e40-91c0-45c9-9890-ffe7673be037-config-data\") pod \"06af5e40-91c0-45c9-9890-ffe7673be037\" (UID: \"06af5e40-91c0-45c9-9890-ffe7673be037\") " Oct 06 10:19:08 crc kubenswrapper[4824]: I1006 10:19:08.111756 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tksh9\" (UniqueName: \"kubernetes.io/projected/06af5e40-91c0-45c9-9890-ffe7673be037-kube-api-access-tksh9\") pod \"06af5e40-91c0-45c9-9890-ffe7673be037\" (UID: \"06af5e40-91c0-45c9-9890-ffe7673be037\") " Oct 06 10:19:08 crc kubenswrapper[4824]: I1006 10:19:08.111801 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/06af5e40-91c0-45c9-9890-ffe7673be037-scripts\") pod \"06af5e40-91c0-45c9-9890-ffe7673be037\" (UID: \"06af5e40-91c0-45c9-9890-ffe7673be037\") " Oct 06 10:19:08 crc kubenswrapper[4824]: I1006 10:19:08.111879 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06af5e40-91c0-45c9-9890-ffe7673be037-combined-ca-bundle\") pod \"06af5e40-91c0-45c9-9890-ffe7673be037\" (UID: \"06af5e40-91c0-45c9-9890-ffe7673be037\") " Oct 06 10:19:08 crc kubenswrapper[4824]: I1006 10:19:08.120413 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06af5e40-91c0-45c9-9890-ffe7673be037-scripts" (OuterVolumeSpecName: "scripts") pod "06af5e40-91c0-45c9-9890-ffe7673be037" (UID: "06af5e40-91c0-45c9-9890-ffe7673be037"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:19:08 crc kubenswrapper[4824]: I1006 10:19:08.129385 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/06af5e40-91c0-45c9-9890-ffe7673be037-kube-api-access-tksh9" (OuterVolumeSpecName: "kube-api-access-tksh9") pod "06af5e40-91c0-45c9-9890-ffe7673be037" (UID: "06af5e40-91c0-45c9-9890-ffe7673be037"). InnerVolumeSpecName "kube-api-access-tksh9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:19:08 crc kubenswrapper[4824]: I1006 10:19:08.160603 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06af5e40-91c0-45c9-9890-ffe7673be037-config-data" (OuterVolumeSpecName: "config-data") pod "06af5e40-91c0-45c9-9890-ffe7673be037" (UID: "06af5e40-91c0-45c9-9890-ffe7673be037"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:19:08 crc kubenswrapper[4824]: I1006 10:19:08.161248 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06af5e40-91c0-45c9-9890-ffe7673be037-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "06af5e40-91c0-45c9-9890-ffe7673be037" (UID: "06af5e40-91c0-45c9-9890-ffe7673be037"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:19:08 crc kubenswrapper[4824]: I1006 10:19:08.214903 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tksh9\" (UniqueName: \"kubernetes.io/projected/06af5e40-91c0-45c9-9890-ffe7673be037-kube-api-access-tksh9\") on node \"crc\" DevicePath \"\"" Oct 06 10:19:08 crc kubenswrapper[4824]: I1006 10:19:08.215439 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/06af5e40-91c0-45c9-9890-ffe7673be037-scripts\") on node \"crc\" DevicePath \"\"" Oct 06 10:19:08 crc kubenswrapper[4824]: I1006 10:19:08.215536 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06af5e40-91c0-45c9-9890-ffe7673be037-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 10:19:08 crc kubenswrapper[4824]: I1006 10:19:08.215612 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06af5e40-91c0-45c9-9890-ffe7673be037-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 10:19:08 crc kubenswrapper[4824]: I1006 10:19:08.417614 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"369bbc94-27b7-4016-9a43-1e20b28f4323","Type":"ContainerStarted","Data":"333f17108a5c2ea5d476b0b350949963f54582ee214f08d0cfae1253519a62cd"} Oct 06 10:19:08 crc kubenswrapper[4824]: I1006 10:19:08.420422 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 06 10:19:08 crc kubenswrapper[4824]: I1006 10:19:08.425319 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-mrzrt" event={"ID":"06af5e40-91c0-45c9-9890-ffe7673be037","Type":"ContainerDied","Data":"b22ab225ecdd418baf887bf650cabf7e2c5b61cbdee9c57216aea63df67f2d61"} Oct 06 10:19:08 crc kubenswrapper[4824]: I1006 10:19:08.425386 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b22ab225ecdd418baf887bf650cabf7e2c5b61cbdee9c57216aea63df67f2d61" Oct 06 10:19:08 crc kubenswrapper[4824]: I1006 10:19:08.425493 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-mrzrt" Oct 06 10:19:08 crc kubenswrapper[4824]: I1006 10:19:08.457631 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.808172691 podStartE2EDuration="5.457602033s" podCreationTimestamp="2025-10-06 10:19:03 +0000 UTC" firstStartedPulling="2025-10-06 10:19:04.223032324 +0000 UTC m=+1253.587455205" lastFinishedPulling="2025-10-06 10:19:07.872461686 +0000 UTC m=+1257.236884547" observedRunningTime="2025-10-06 10:19:08.445478499 +0000 UTC m=+1257.809901370" watchObservedRunningTime="2025-10-06 10:19:08.457602033 +0000 UTC m=+1257.822024904" Oct 06 10:19:08 crc kubenswrapper[4824]: I1006 10:19:08.566322 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="2056be6d-e4b2-4f5e-91a7-ac5b2cb357ce" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.204:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 06 10:19:08 crc kubenswrapper[4824]: I1006 10:19:08.566383 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="2056be6d-e4b2-4f5e-91a7-ac5b2cb357ce" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.204:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 06 10:19:08 crc kubenswrapper[4824]: I1006 10:19:08.627335 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 06 10:19:08 crc kubenswrapper[4824]: I1006 10:19:08.628167 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="8958352a-0551-4af4-8e6b-6655ea3490f6" containerName="nova-scheduler-scheduler" containerID="cri-o://7933fab2e0964f7f5e7bfb0ad7a3d55be232f3319edbbfd02c0bebc60abc9188" gracePeriod=30 Oct 06 10:19:08 crc kubenswrapper[4824]: I1006 10:19:08.644386 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 06 10:19:08 crc kubenswrapper[4824]: I1006 10:19:08.645221 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="2056be6d-e4b2-4f5e-91a7-ac5b2cb357ce" containerName="nova-api-api" containerID="cri-o://ba945d970f2119d285f28c46084b1b59b9567dd3cd65490e0d58c9f33bf072a7" gracePeriod=30 Oct 06 10:19:08 crc kubenswrapper[4824]: I1006 10:19:08.645499 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="2056be6d-e4b2-4f5e-91a7-ac5b2cb357ce" containerName="nova-api-log" containerID="cri-o://80156743cc95e672e11830faa0c625af7a756c3dc73c8b5f5c1110102ff75d85" gracePeriod=30 Oct 06 10:19:08 crc kubenswrapper[4824]: I1006 10:19:08.660796 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 06 10:19:08 crc kubenswrapper[4824]: I1006 10:19:08.661392 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="38f30d16-bb2b-4a0b-aac6-64078e032eeb" containerName="nova-metadata-log" containerID="cri-o://e582782b73e98e3ec11c149376d44e09b4ebf2e98421da7607dfbc12b3b526a9" gracePeriod=30 Oct 06 10:19:08 crc kubenswrapper[4824]: I1006 10:19:08.661499 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="38f30d16-bb2b-4a0b-aac6-64078e032eeb" containerName="nova-metadata-metadata" containerID="cri-o://aa67c97d929e74e959591f108ea9503ddd0ba9c3e4d21ac0322b719670baaf66" gracePeriod=30 Oct 06 10:19:09 crc kubenswrapper[4824]: I1006 10:19:09.439443 4824 generic.go:334] "Generic (PLEG): container finished" podID="38f30d16-bb2b-4a0b-aac6-64078e032eeb" containerID="e582782b73e98e3ec11c149376d44e09b4ebf2e98421da7607dfbc12b3b526a9" exitCode=143 Oct 06 10:19:09 crc kubenswrapper[4824]: I1006 10:19:09.439509 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"38f30d16-bb2b-4a0b-aac6-64078e032eeb","Type":"ContainerDied","Data":"e582782b73e98e3ec11c149376d44e09b4ebf2e98421da7607dfbc12b3b526a9"} Oct 06 10:19:09 crc kubenswrapper[4824]: I1006 10:19:09.444184 4824 generic.go:334] "Generic (PLEG): container finished" podID="2056be6d-e4b2-4f5e-91a7-ac5b2cb357ce" containerID="80156743cc95e672e11830faa0c625af7a756c3dc73c8b5f5c1110102ff75d85" exitCode=143 Oct 06 10:19:09 crc kubenswrapper[4824]: I1006 10:19:09.444299 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2056be6d-e4b2-4f5e-91a7-ac5b2cb357ce","Type":"ContainerDied","Data":"80156743cc95e672e11830faa0c625af7a756c3dc73c8b5f5c1110102ff75d85"} Oct 06 10:19:10 crc kubenswrapper[4824]: E1006 10:19:10.023224 4824 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="7933fab2e0964f7f5e7bfb0ad7a3d55be232f3319edbbfd02c0bebc60abc9188" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 06 10:19:10 crc kubenswrapper[4824]: E1006 10:19:10.024648 4824 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="7933fab2e0964f7f5e7bfb0ad7a3d55be232f3319edbbfd02c0bebc60abc9188" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 06 10:19:10 crc kubenswrapper[4824]: E1006 10:19:10.025805 4824 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="7933fab2e0964f7f5e7bfb0ad7a3d55be232f3319edbbfd02c0bebc60abc9188" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 06 10:19:10 crc kubenswrapper[4824]: E1006 10:19:10.025858 4824 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="8958352a-0551-4af4-8e6b-6655ea3490f6" containerName="nova-scheduler-scheduler" Oct 06 10:19:12 crc kubenswrapper[4824]: I1006 10:19:12.086467 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="38f30d16-bb2b-4a0b-aac6-64078e032eeb" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.196:8775/\": read tcp 10.217.0.2:33914->10.217.0.196:8775: read: connection reset by peer" Oct 06 10:19:12 crc kubenswrapper[4824]: I1006 10:19:12.086566 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="38f30d16-bb2b-4a0b-aac6-64078e032eeb" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.196:8775/\": read tcp 10.217.0.2:33926->10.217.0.196:8775: read: connection reset by peer" Oct 06 10:19:12 crc kubenswrapper[4824]: I1006 10:19:12.504751 4824 generic.go:334] "Generic (PLEG): container finished" podID="38f30d16-bb2b-4a0b-aac6-64078e032eeb" containerID="aa67c97d929e74e959591f108ea9503ddd0ba9c3e4d21ac0322b719670baaf66" exitCode=0 Oct 06 10:19:12 crc kubenswrapper[4824]: I1006 10:19:12.504871 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"38f30d16-bb2b-4a0b-aac6-64078e032eeb","Type":"ContainerDied","Data":"aa67c97d929e74e959591f108ea9503ddd0ba9c3e4d21ac0322b719670baaf66"} Oct 06 10:19:12 crc kubenswrapper[4824]: I1006 10:19:12.721443 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 06 10:19:12 crc kubenswrapper[4824]: I1006 10:19:12.851264 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/38f30d16-bb2b-4a0b-aac6-64078e032eeb-combined-ca-bundle\") pod \"38f30d16-bb2b-4a0b-aac6-64078e032eeb\" (UID: \"38f30d16-bb2b-4a0b-aac6-64078e032eeb\") " Oct 06 10:19:12 crc kubenswrapper[4824]: I1006 10:19:12.851384 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/38f30d16-bb2b-4a0b-aac6-64078e032eeb-config-data\") pod \"38f30d16-bb2b-4a0b-aac6-64078e032eeb\" (UID: \"38f30d16-bb2b-4a0b-aac6-64078e032eeb\") " Oct 06 10:19:12 crc kubenswrapper[4824]: I1006 10:19:12.851513 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vh8r5\" (UniqueName: \"kubernetes.io/projected/38f30d16-bb2b-4a0b-aac6-64078e032eeb-kube-api-access-vh8r5\") pod \"38f30d16-bb2b-4a0b-aac6-64078e032eeb\" (UID: \"38f30d16-bb2b-4a0b-aac6-64078e032eeb\") " Oct 06 10:19:12 crc kubenswrapper[4824]: I1006 10:19:12.851753 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/38f30d16-bb2b-4a0b-aac6-64078e032eeb-logs\") pod \"38f30d16-bb2b-4a0b-aac6-64078e032eeb\" (UID: \"38f30d16-bb2b-4a0b-aac6-64078e032eeb\") " Oct 06 10:19:12 crc kubenswrapper[4824]: I1006 10:19:12.851785 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/38f30d16-bb2b-4a0b-aac6-64078e032eeb-nova-metadata-tls-certs\") pod \"38f30d16-bb2b-4a0b-aac6-64078e032eeb\" (UID: \"38f30d16-bb2b-4a0b-aac6-64078e032eeb\") " Oct 06 10:19:12 crc kubenswrapper[4824]: I1006 10:19:12.852610 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/38f30d16-bb2b-4a0b-aac6-64078e032eeb-logs" (OuterVolumeSpecName: "logs") pod "38f30d16-bb2b-4a0b-aac6-64078e032eeb" (UID: "38f30d16-bb2b-4a0b-aac6-64078e032eeb"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:19:12 crc kubenswrapper[4824]: I1006 10:19:12.853088 4824 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/38f30d16-bb2b-4a0b-aac6-64078e032eeb-logs\") on node \"crc\" DevicePath \"\"" Oct 06 10:19:12 crc kubenswrapper[4824]: I1006 10:19:12.861529 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/38f30d16-bb2b-4a0b-aac6-64078e032eeb-kube-api-access-vh8r5" (OuterVolumeSpecName: "kube-api-access-vh8r5") pod "38f30d16-bb2b-4a0b-aac6-64078e032eeb" (UID: "38f30d16-bb2b-4a0b-aac6-64078e032eeb"). InnerVolumeSpecName "kube-api-access-vh8r5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:19:12 crc kubenswrapper[4824]: I1006 10:19:12.894907 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/38f30d16-bb2b-4a0b-aac6-64078e032eeb-config-data" (OuterVolumeSpecName: "config-data") pod "38f30d16-bb2b-4a0b-aac6-64078e032eeb" (UID: "38f30d16-bb2b-4a0b-aac6-64078e032eeb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:19:12 crc kubenswrapper[4824]: I1006 10:19:12.927141 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/38f30d16-bb2b-4a0b-aac6-64078e032eeb-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "38f30d16-bb2b-4a0b-aac6-64078e032eeb" (UID: "38f30d16-bb2b-4a0b-aac6-64078e032eeb"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:19:12 crc kubenswrapper[4824]: I1006 10:19:12.929162 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/38f30d16-bb2b-4a0b-aac6-64078e032eeb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "38f30d16-bb2b-4a0b-aac6-64078e032eeb" (UID: "38f30d16-bb2b-4a0b-aac6-64078e032eeb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:19:12 crc kubenswrapper[4824]: I1006 10:19:12.956141 4824 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/38f30d16-bb2b-4a0b-aac6-64078e032eeb-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 06 10:19:12 crc kubenswrapper[4824]: I1006 10:19:12.956191 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/38f30d16-bb2b-4a0b-aac6-64078e032eeb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 10:19:12 crc kubenswrapper[4824]: I1006 10:19:12.956201 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/38f30d16-bb2b-4a0b-aac6-64078e032eeb-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 10:19:12 crc kubenswrapper[4824]: I1006 10:19:12.956210 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vh8r5\" (UniqueName: \"kubernetes.io/projected/38f30d16-bb2b-4a0b-aac6-64078e032eeb-kube-api-access-vh8r5\") on node \"crc\" DevicePath \"\"" Oct 06 10:19:13 crc kubenswrapper[4824]: I1006 10:19:13.547873 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"38f30d16-bb2b-4a0b-aac6-64078e032eeb","Type":"ContainerDied","Data":"bd41727b70698dc94ba3e3fb0c13a2822d04f71771ae3b6ad6d7626d71132a91"} Oct 06 10:19:13 crc kubenswrapper[4824]: I1006 10:19:13.548025 4824 scope.go:117] "RemoveContainer" containerID="aa67c97d929e74e959591f108ea9503ddd0ba9c3e4d21ac0322b719670baaf66" Oct 06 10:19:13 crc kubenswrapper[4824]: I1006 10:19:13.548279 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 06 10:19:13 crc kubenswrapper[4824]: I1006 10:19:13.587714 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 06 10:19:13 crc kubenswrapper[4824]: I1006 10:19:13.593783 4824 scope.go:117] "RemoveContainer" containerID="e582782b73e98e3ec11c149376d44e09b4ebf2e98421da7607dfbc12b3b526a9" Oct 06 10:19:13 crc kubenswrapper[4824]: I1006 10:19:13.600784 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 06 10:19:13 crc kubenswrapper[4824]: I1006 10:19:13.618217 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 06 10:19:13 crc kubenswrapper[4824]: E1006 10:19:13.618734 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06af5e40-91c0-45c9-9890-ffe7673be037" containerName="nova-manage" Oct 06 10:19:13 crc kubenswrapper[4824]: I1006 10:19:13.618757 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="06af5e40-91c0-45c9-9890-ffe7673be037" containerName="nova-manage" Oct 06 10:19:13 crc kubenswrapper[4824]: E1006 10:19:13.618794 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38f30d16-bb2b-4a0b-aac6-64078e032eeb" containerName="nova-metadata-log" Oct 06 10:19:13 crc kubenswrapper[4824]: I1006 10:19:13.618805 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="38f30d16-bb2b-4a0b-aac6-64078e032eeb" containerName="nova-metadata-log" Oct 06 10:19:13 crc kubenswrapper[4824]: E1006 10:19:13.618815 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38f30d16-bb2b-4a0b-aac6-64078e032eeb" containerName="nova-metadata-metadata" Oct 06 10:19:13 crc kubenswrapper[4824]: I1006 10:19:13.618823 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="38f30d16-bb2b-4a0b-aac6-64078e032eeb" containerName="nova-metadata-metadata" Oct 06 10:19:13 crc kubenswrapper[4824]: I1006 10:19:13.619136 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="06af5e40-91c0-45c9-9890-ffe7673be037" containerName="nova-manage" Oct 06 10:19:13 crc kubenswrapper[4824]: I1006 10:19:13.619273 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="38f30d16-bb2b-4a0b-aac6-64078e032eeb" containerName="nova-metadata-metadata" Oct 06 10:19:13 crc kubenswrapper[4824]: I1006 10:19:13.619306 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="38f30d16-bb2b-4a0b-aac6-64078e032eeb" containerName="nova-metadata-log" Oct 06 10:19:13 crc kubenswrapper[4824]: I1006 10:19:13.620643 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 06 10:19:13 crc kubenswrapper[4824]: I1006 10:19:13.626623 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 06 10:19:13 crc kubenswrapper[4824]: I1006 10:19:13.627454 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 06 10:19:13 crc kubenswrapper[4824]: I1006 10:19:13.652284 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 06 10:19:13 crc kubenswrapper[4824]: I1006 10:19:13.675006 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/754abe5c-3e1b-4a4a-b5a7-619326c06846-logs\") pod \"nova-metadata-0\" (UID: \"754abe5c-3e1b-4a4a-b5a7-619326c06846\") " pod="openstack/nova-metadata-0" Oct 06 10:19:13 crc kubenswrapper[4824]: I1006 10:19:13.675481 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/754abe5c-3e1b-4a4a-b5a7-619326c06846-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"754abe5c-3e1b-4a4a-b5a7-619326c06846\") " pod="openstack/nova-metadata-0" Oct 06 10:19:13 crc kubenswrapper[4824]: I1006 10:19:13.675673 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/754abe5c-3e1b-4a4a-b5a7-619326c06846-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"754abe5c-3e1b-4a4a-b5a7-619326c06846\") " pod="openstack/nova-metadata-0" Oct 06 10:19:13 crc kubenswrapper[4824]: I1006 10:19:13.675803 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/754abe5c-3e1b-4a4a-b5a7-619326c06846-config-data\") pod \"nova-metadata-0\" (UID: \"754abe5c-3e1b-4a4a-b5a7-619326c06846\") " pod="openstack/nova-metadata-0" Oct 06 10:19:13 crc kubenswrapper[4824]: I1006 10:19:13.675927 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sdqpv\" (UniqueName: \"kubernetes.io/projected/754abe5c-3e1b-4a4a-b5a7-619326c06846-kube-api-access-sdqpv\") pod \"nova-metadata-0\" (UID: \"754abe5c-3e1b-4a4a-b5a7-619326c06846\") " pod="openstack/nova-metadata-0" Oct 06 10:19:13 crc kubenswrapper[4824]: I1006 10:19:13.778073 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/754abe5c-3e1b-4a4a-b5a7-619326c06846-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"754abe5c-3e1b-4a4a-b5a7-619326c06846\") " pod="openstack/nova-metadata-0" Oct 06 10:19:13 crc kubenswrapper[4824]: I1006 10:19:13.778133 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/754abe5c-3e1b-4a4a-b5a7-619326c06846-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"754abe5c-3e1b-4a4a-b5a7-619326c06846\") " pod="openstack/nova-metadata-0" Oct 06 10:19:13 crc kubenswrapper[4824]: I1006 10:19:13.778174 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/754abe5c-3e1b-4a4a-b5a7-619326c06846-config-data\") pod \"nova-metadata-0\" (UID: \"754abe5c-3e1b-4a4a-b5a7-619326c06846\") " pod="openstack/nova-metadata-0" Oct 06 10:19:13 crc kubenswrapper[4824]: I1006 10:19:13.778207 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sdqpv\" (UniqueName: \"kubernetes.io/projected/754abe5c-3e1b-4a4a-b5a7-619326c06846-kube-api-access-sdqpv\") pod \"nova-metadata-0\" (UID: \"754abe5c-3e1b-4a4a-b5a7-619326c06846\") " pod="openstack/nova-metadata-0" Oct 06 10:19:13 crc kubenswrapper[4824]: I1006 10:19:13.778255 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/754abe5c-3e1b-4a4a-b5a7-619326c06846-logs\") pod \"nova-metadata-0\" (UID: \"754abe5c-3e1b-4a4a-b5a7-619326c06846\") " pod="openstack/nova-metadata-0" Oct 06 10:19:13 crc kubenswrapper[4824]: I1006 10:19:13.778711 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/754abe5c-3e1b-4a4a-b5a7-619326c06846-logs\") pod \"nova-metadata-0\" (UID: \"754abe5c-3e1b-4a4a-b5a7-619326c06846\") " pod="openstack/nova-metadata-0" Oct 06 10:19:13 crc kubenswrapper[4824]: I1006 10:19:13.786751 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/754abe5c-3e1b-4a4a-b5a7-619326c06846-config-data\") pod \"nova-metadata-0\" (UID: \"754abe5c-3e1b-4a4a-b5a7-619326c06846\") " pod="openstack/nova-metadata-0" Oct 06 10:19:13 crc kubenswrapper[4824]: I1006 10:19:13.786738 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/754abe5c-3e1b-4a4a-b5a7-619326c06846-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"754abe5c-3e1b-4a4a-b5a7-619326c06846\") " pod="openstack/nova-metadata-0" Oct 06 10:19:13 crc kubenswrapper[4824]: I1006 10:19:13.788038 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/754abe5c-3e1b-4a4a-b5a7-619326c06846-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"754abe5c-3e1b-4a4a-b5a7-619326c06846\") " pod="openstack/nova-metadata-0" Oct 06 10:19:13 crc kubenswrapper[4824]: I1006 10:19:13.798081 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sdqpv\" (UniqueName: \"kubernetes.io/projected/754abe5c-3e1b-4a4a-b5a7-619326c06846-kube-api-access-sdqpv\") pod \"nova-metadata-0\" (UID: \"754abe5c-3e1b-4a4a-b5a7-619326c06846\") " pod="openstack/nova-metadata-0" Oct 06 10:19:13 crc kubenswrapper[4824]: I1006 10:19:13.915218 4824 patch_prober.go:28] interesting pod/machine-config-daemon-khgzw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 10:19:13 crc kubenswrapper[4824]: I1006 10:19:13.915733 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 10:19:13 crc kubenswrapper[4824]: I1006 10:19:13.915857 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" Oct 06 10:19:13 crc kubenswrapper[4824]: I1006 10:19:13.916714 4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ad63fa5818247f7e35384bdeb2bf047f095a2d6eed77601da04ba8f26adf2f37"} pod="openshift-machine-config-operator/machine-config-daemon-khgzw" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 06 10:19:13 crc kubenswrapper[4824]: I1006 10:19:13.916854 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" containerName="machine-config-daemon" containerID="cri-o://ad63fa5818247f7e35384bdeb2bf047f095a2d6eed77601da04ba8f26adf2f37" gracePeriod=600 Oct 06 10:19:13 crc kubenswrapper[4824]: I1006 10:19:13.954793 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 06 10:19:14 crc kubenswrapper[4824]: I1006 10:19:14.425646 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 06 10:19:14 crc kubenswrapper[4824]: I1006 10:19:14.493311 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8958352a-0551-4af4-8e6b-6655ea3490f6-combined-ca-bundle\") pod \"8958352a-0551-4af4-8e6b-6655ea3490f6\" (UID: \"8958352a-0551-4af4-8e6b-6655ea3490f6\") " Oct 06 10:19:14 crc kubenswrapper[4824]: I1006 10:19:14.493542 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mkf5f\" (UniqueName: \"kubernetes.io/projected/8958352a-0551-4af4-8e6b-6655ea3490f6-kube-api-access-mkf5f\") pod \"8958352a-0551-4af4-8e6b-6655ea3490f6\" (UID: \"8958352a-0551-4af4-8e6b-6655ea3490f6\") " Oct 06 10:19:14 crc kubenswrapper[4824]: I1006 10:19:14.493588 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8958352a-0551-4af4-8e6b-6655ea3490f6-config-data\") pod \"8958352a-0551-4af4-8e6b-6655ea3490f6\" (UID: \"8958352a-0551-4af4-8e6b-6655ea3490f6\") " Oct 06 10:19:14 crc kubenswrapper[4824]: I1006 10:19:14.504236 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8958352a-0551-4af4-8e6b-6655ea3490f6-kube-api-access-mkf5f" (OuterVolumeSpecName: "kube-api-access-mkf5f") pod "8958352a-0551-4af4-8e6b-6655ea3490f6" (UID: "8958352a-0551-4af4-8e6b-6655ea3490f6"). InnerVolumeSpecName "kube-api-access-mkf5f". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:19:14 crc kubenswrapper[4824]: I1006 10:19:14.543215 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 06 10:19:14 crc kubenswrapper[4824]: I1006 10:19:14.571406 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8958352a-0551-4af4-8e6b-6655ea3490f6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8958352a-0551-4af4-8e6b-6655ea3490f6" (UID: "8958352a-0551-4af4-8e6b-6655ea3490f6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:19:14 crc kubenswrapper[4824]: I1006 10:19:14.581605 4824 generic.go:334] "Generic (PLEG): container finished" podID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" containerID="ad63fa5818247f7e35384bdeb2bf047f095a2d6eed77601da04ba8f26adf2f37" exitCode=0 Oct 06 10:19:14 crc kubenswrapper[4824]: I1006 10:19:14.581676 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" event={"ID":"e1140c8c-93f1-4fce-8c06-a69561ff8a6d","Type":"ContainerDied","Data":"ad63fa5818247f7e35384bdeb2bf047f095a2d6eed77601da04ba8f26adf2f37"} Oct 06 10:19:14 crc kubenswrapper[4824]: I1006 10:19:14.581713 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" event={"ID":"e1140c8c-93f1-4fce-8c06-a69561ff8a6d","Type":"ContainerStarted","Data":"7f680663f7fe4c1e2c1c711b5dc3a94dfad5110e11cc00fd374e4d6a79007f09"} Oct 06 10:19:14 crc kubenswrapper[4824]: I1006 10:19:14.581736 4824 scope.go:117] "RemoveContainer" containerID="797b5c3c4be7e935c4c45dea62e8a371c48208db7d2ac607bd9b1a5fbc29630b" Oct 06 10:19:14 crc kubenswrapper[4824]: I1006 10:19:14.596595 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8958352a-0551-4af4-8e6b-6655ea3490f6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 10:19:14 crc kubenswrapper[4824]: I1006 10:19:14.596643 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mkf5f\" (UniqueName: \"kubernetes.io/projected/8958352a-0551-4af4-8e6b-6655ea3490f6-kube-api-access-mkf5f\") on node \"crc\" DevicePath \"\"" Oct 06 10:19:14 crc kubenswrapper[4824]: I1006 10:19:14.602182 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8958352a-0551-4af4-8e6b-6655ea3490f6-config-data" (OuterVolumeSpecName: "config-data") pod "8958352a-0551-4af4-8e6b-6655ea3490f6" (UID: "8958352a-0551-4af4-8e6b-6655ea3490f6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:19:14 crc kubenswrapper[4824]: I1006 10:19:14.611491 4824 generic.go:334] "Generic (PLEG): container finished" podID="2056be6d-e4b2-4f5e-91a7-ac5b2cb357ce" containerID="ba945d970f2119d285f28c46084b1b59b9567dd3cd65490e0d58c9f33bf072a7" exitCode=0 Oct 06 10:19:14 crc kubenswrapper[4824]: I1006 10:19:14.611587 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2056be6d-e4b2-4f5e-91a7-ac5b2cb357ce","Type":"ContainerDied","Data":"ba945d970f2119d285f28c46084b1b59b9567dd3cd65490e0d58c9f33bf072a7"} Oct 06 10:19:14 crc kubenswrapper[4824]: I1006 10:19:14.613760 4824 generic.go:334] "Generic (PLEG): container finished" podID="8958352a-0551-4af4-8e6b-6655ea3490f6" containerID="7933fab2e0964f7f5e7bfb0ad7a3d55be232f3319edbbfd02c0bebc60abc9188" exitCode=0 Oct 06 10:19:14 crc kubenswrapper[4824]: I1006 10:19:14.613787 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"8958352a-0551-4af4-8e6b-6655ea3490f6","Type":"ContainerDied","Data":"7933fab2e0964f7f5e7bfb0ad7a3d55be232f3319edbbfd02c0bebc60abc9188"} Oct 06 10:19:14 crc kubenswrapper[4824]: I1006 10:19:14.613804 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"8958352a-0551-4af4-8e6b-6655ea3490f6","Type":"ContainerDied","Data":"263512d78658387a78e68953f51ba1fd17ebaecd6d3472c8a88a52845378388e"} Oct 06 10:19:14 crc kubenswrapper[4824]: I1006 10:19:14.613873 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 06 10:19:14 crc kubenswrapper[4824]: I1006 10:19:14.655108 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 06 10:19:14 crc kubenswrapper[4824]: I1006 10:19:14.669887 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 06 10:19:14 crc kubenswrapper[4824]: I1006 10:19:14.690608 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 06 10:19:14 crc kubenswrapper[4824]: E1006 10:19:14.691289 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8958352a-0551-4af4-8e6b-6655ea3490f6" containerName="nova-scheduler-scheduler" Oct 06 10:19:14 crc kubenswrapper[4824]: I1006 10:19:14.691326 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="8958352a-0551-4af4-8e6b-6655ea3490f6" containerName="nova-scheduler-scheduler" Oct 06 10:19:14 crc kubenswrapper[4824]: I1006 10:19:14.691562 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="8958352a-0551-4af4-8e6b-6655ea3490f6" containerName="nova-scheduler-scheduler" Oct 06 10:19:14 crc kubenswrapper[4824]: I1006 10:19:14.692484 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 06 10:19:14 crc kubenswrapper[4824]: I1006 10:19:14.697204 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 06 10:19:14 crc kubenswrapper[4824]: I1006 10:19:14.707917 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 06 10:19:14 crc kubenswrapper[4824]: I1006 10:19:14.708323 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 06 10:19:14 crc kubenswrapper[4824]: I1006 10:19:14.723315 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8958352a-0551-4af4-8e6b-6655ea3490f6-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 10:19:14 crc kubenswrapper[4824]: I1006 10:19:14.727145 4824 scope.go:117] "RemoveContainer" containerID="7933fab2e0964f7f5e7bfb0ad7a3d55be232f3319edbbfd02c0bebc60abc9188" Oct 06 10:19:14 crc kubenswrapper[4824]: I1006 10:19:14.761554 4824 scope.go:117] "RemoveContainer" containerID="7933fab2e0964f7f5e7bfb0ad7a3d55be232f3319edbbfd02c0bebc60abc9188" Oct 06 10:19:14 crc kubenswrapper[4824]: E1006 10:19:14.762471 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7933fab2e0964f7f5e7bfb0ad7a3d55be232f3319edbbfd02c0bebc60abc9188\": container with ID starting with 7933fab2e0964f7f5e7bfb0ad7a3d55be232f3319edbbfd02c0bebc60abc9188 not found: ID does not exist" containerID="7933fab2e0964f7f5e7bfb0ad7a3d55be232f3319edbbfd02c0bebc60abc9188" Oct 06 10:19:14 crc kubenswrapper[4824]: I1006 10:19:14.762514 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7933fab2e0964f7f5e7bfb0ad7a3d55be232f3319edbbfd02c0bebc60abc9188"} err="failed to get container status \"7933fab2e0964f7f5e7bfb0ad7a3d55be232f3319edbbfd02c0bebc60abc9188\": rpc error: code = NotFound desc = could not find container \"7933fab2e0964f7f5e7bfb0ad7a3d55be232f3319edbbfd02c0bebc60abc9188\": container with ID starting with 7933fab2e0964f7f5e7bfb0ad7a3d55be232f3319edbbfd02c0bebc60abc9188 not found: ID does not exist" Oct 06 10:19:14 crc kubenswrapper[4824]: I1006 10:19:14.824701 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2056be6d-e4b2-4f5e-91a7-ac5b2cb357ce-config-data\") pod \"2056be6d-e4b2-4f5e-91a7-ac5b2cb357ce\" (UID: \"2056be6d-e4b2-4f5e-91a7-ac5b2cb357ce\") " Oct 06 10:19:14 crc kubenswrapper[4824]: I1006 10:19:14.824790 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6v8x6\" (UniqueName: \"kubernetes.io/projected/2056be6d-e4b2-4f5e-91a7-ac5b2cb357ce-kube-api-access-6v8x6\") pod \"2056be6d-e4b2-4f5e-91a7-ac5b2cb357ce\" (UID: \"2056be6d-e4b2-4f5e-91a7-ac5b2cb357ce\") " Oct 06 10:19:14 crc kubenswrapper[4824]: I1006 10:19:14.825064 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2056be6d-e4b2-4f5e-91a7-ac5b2cb357ce-combined-ca-bundle\") pod \"2056be6d-e4b2-4f5e-91a7-ac5b2cb357ce\" (UID: \"2056be6d-e4b2-4f5e-91a7-ac5b2cb357ce\") " Oct 06 10:19:14 crc kubenswrapper[4824]: I1006 10:19:14.825115 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2056be6d-e4b2-4f5e-91a7-ac5b2cb357ce-logs\") pod \"2056be6d-e4b2-4f5e-91a7-ac5b2cb357ce\" (UID: \"2056be6d-e4b2-4f5e-91a7-ac5b2cb357ce\") " Oct 06 10:19:14 crc kubenswrapper[4824]: I1006 10:19:14.825185 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2056be6d-e4b2-4f5e-91a7-ac5b2cb357ce-internal-tls-certs\") pod \"2056be6d-e4b2-4f5e-91a7-ac5b2cb357ce\" (UID: \"2056be6d-e4b2-4f5e-91a7-ac5b2cb357ce\") " Oct 06 10:19:14 crc kubenswrapper[4824]: I1006 10:19:14.825329 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2056be6d-e4b2-4f5e-91a7-ac5b2cb357ce-public-tls-certs\") pod \"2056be6d-e4b2-4f5e-91a7-ac5b2cb357ce\" (UID: \"2056be6d-e4b2-4f5e-91a7-ac5b2cb357ce\") " Oct 06 10:19:14 crc kubenswrapper[4824]: I1006 10:19:14.825773 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1166f1ef-43be-42e7-9d5f-84252fccfb69-config-data\") pod \"nova-scheduler-0\" (UID: \"1166f1ef-43be-42e7-9d5f-84252fccfb69\") " pod="openstack/nova-scheduler-0" Oct 06 10:19:14 crc kubenswrapper[4824]: I1006 10:19:14.825819 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6wx5g\" (UniqueName: \"kubernetes.io/projected/1166f1ef-43be-42e7-9d5f-84252fccfb69-kube-api-access-6wx5g\") pod \"nova-scheduler-0\" (UID: \"1166f1ef-43be-42e7-9d5f-84252fccfb69\") " pod="openstack/nova-scheduler-0" Oct 06 10:19:14 crc kubenswrapper[4824]: I1006 10:19:14.825847 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1166f1ef-43be-42e7-9d5f-84252fccfb69-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"1166f1ef-43be-42e7-9d5f-84252fccfb69\") " pod="openstack/nova-scheduler-0" Oct 06 10:19:14 crc kubenswrapper[4824]: I1006 10:19:14.826471 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2056be6d-e4b2-4f5e-91a7-ac5b2cb357ce-logs" (OuterVolumeSpecName: "logs") pod "2056be6d-e4b2-4f5e-91a7-ac5b2cb357ce" (UID: "2056be6d-e4b2-4f5e-91a7-ac5b2cb357ce"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:19:14 crc kubenswrapper[4824]: I1006 10:19:14.831473 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2056be6d-e4b2-4f5e-91a7-ac5b2cb357ce-kube-api-access-6v8x6" (OuterVolumeSpecName: "kube-api-access-6v8x6") pod "2056be6d-e4b2-4f5e-91a7-ac5b2cb357ce" (UID: "2056be6d-e4b2-4f5e-91a7-ac5b2cb357ce"). InnerVolumeSpecName "kube-api-access-6v8x6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:19:14 crc kubenswrapper[4824]: I1006 10:19:14.861115 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2056be6d-e4b2-4f5e-91a7-ac5b2cb357ce-config-data" (OuterVolumeSpecName: "config-data") pod "2056be6d-e4b2-4f5e-91a7-ac5b2cb357ce" (UID: "2056be6d-e4b2-4f5e-91a7-ac5b2cb357ce"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:19:14 crc kubenswrapper[4824]: I1006 10:19:14.863347 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2056be6d-e4b2-4f5e-91a7-ac5b2cb357ce-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2056be6d-e4b2-4f5e-91a7-ac5b2cb357ce" (UID: "2056be6d-e4b2-4f5e-91a7-ac5b2cb357ce"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:19:14 crc kubenswrapper[4824]: I1006 10:19:14.889914 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2056be6d-e4b2-4f5e-91a7-ac5b2cb357ce-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "2056be6d-e4b2-4f5e-91a7-ac5b2cb357ce" (UID: "2056be6d-e4b2-4f5e-91a7-ac5b2cb357ce"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:19:14 crc kubenswrapper[4824]: I1006 10:19:14.895101 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2056be6d-e4b2-4f5e-91a7-ac5b2cb357ce-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "2056be6d-e4b2-4f5e-91a7-ac5b2cb357ce" (UID: "2056be6d-e4b2-4f5e-91a7-ac5b2cb357ce"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:19:14 crc kubenswrapper[4824]: I1006 10:19:14.928118 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1166f1ef-43be-42e7-9d5f-84252fccfb69-config-data\") pod \"nova-scheduler-0\" (UID: \"1166f1ef-43be-42e7-9d5f-84252fccfb69\") " pod="openstack/nova-scheduler-0" Oct 06 10:19:14 crc kubenswrapper[4824]: I1006 10:19:14.928225 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6wx5g\" (UniqueName: \"kubernetes.io/projected/1166f1ef-43be-42e7-9d5f-84252fccfb69-kube-api-access-6wx5g\") pod \"nova-scheduler-0\" (UID: \"1166f1ef-43be-42e7-9d5f-84252fccfb69\") " pod="openstack/nova-scheduler-0" Oct 06 10:19:14 crc kubenswrapper[4824]: I1006 10:19:14.928271 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1166f1ef-43be-42e7-9d5f-84252fccfb69-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"1166f1ef-43be-42e7-9d5f-84252fccfb69\") " pod="openstack/nova-scheduler-0" Oct 06 10:19:14 crc kubenswrapper[4824]: I1006 10:19:14.928541 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2056be6d-e4b2-4f5e-91a7-ac5b2cb357ce-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 10:19:14 crc kubenswrapper[4824]: I1006 10:19:14.928564 4824 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2056be6d-e4b2-4f5e-91a7-ac5b2cb357ce-logs\") on node \"crc\" DevicePath \"\"" Oct 06 10:19:14 crc kubenswrapper[4824]: I1006 10:19:14.928579 4824 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2056be6d-e4b2-4f5e-91a7-ac5b2cb357ce-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 06 10:19:14 crc kubenswrapper[4824]: I1006 10:19:14.928594 4824 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2056be6d-e4b2-4f5e-91a7-ac5b2cb357ce-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 06 10:19:14 crc kubenswrapper[4824]: I1006 10:19:14.928606 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2056be6d-e4b2-4f5e-91a7-ac5b2cb357ce-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 10:19:14 crc kubenswrapper[4824]: I1006 10:19:14.928617 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6v8x6\" (UniqueName: \"kubernetes.io/projected/2056be6d-e4b2-4f5e-91a7-ac5b2cb357ce-kube-api-access-6v8x6\") on node \"crc\" DevicePath \"\"" Oct 06 10:19:14 crc kubenswrapper[4824]: I1006 10:19:14.934091 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1166f1ef-43be-42e7-9d5f-84252fccfb69-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"1166f1ef-43be-42e7-9d5f-84252fccfb69\") " pod="openstack/nova-scheduler-0" Oct 06 10:19:14 crc kubenswrapper[4824]: I1006 10:19:14.934382 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1166f1ef-43be-42e7-9d5f-84252fccfb69-config-data\") pod \"nova-scheduler-0\" (UID: \"1166f1ef-43be-42e7-9d5f-84252fccfb69\") " pod="openstack/nova-scheduler-0" Oct 06 10:19:14 crc kubenswrapper[4824]: I1006 10:19:14.949532 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6wx5g\" (UniqueName: \"kubernetes.io/projected/1166f1ef-43be-42e7-9d5f-84252fccfb69-kube-api-access-6wx5g\") pod \"nova-scheduler-0\" (UID: \"1166f1ef-43be-42e7-9d5f-84252fccfb69\") " pod="openstack/nova-scheduler-0" Oct 06 10:19:15 crc kubenswrapper[4824]: I1006 10:19:15.041181 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 06 10:19:15 crc kubenswrapper[4824]: I1006 10:19:15.285437 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="38f30d16-bb2b-4a0b-aac6-64078e032eeb" path="/var/lib/kubelet/pods/38f30d16-bb2b-4a0b-aac6-64078e032eeb/volumes" Oct 06 10:19:15 crc kubenswrapper[4824]: I1006 10:19:15.286380 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8958352a-0551-4af4-8e6b-6655ea3490f6" path="/var/lib/kubelet/pods/8958352a-0551-4af4-8e6b-6655ea3490f6/volumes" Oct 06 10:19:15 crc kubenswrapper[4824]: I1006 10:19:15.543952 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 06 10:19:15 crc kubenswrapper[4824]: W1006 10:19:15.551709 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1166f1ef_43be_42e7_9d5f_84252fccfb69.slice/crio-2abda6592d7134dcf3063b2b2588a762a4269537be2f4f8a111b5bcac0879bed WatchSource:0}: Error finding container 2abda6592d7134dcf3063b2b2588a762a4269537be2f4f8a111b5bcac0879bed: Status 404 returned error can't find the container with id 2abda6592d7134dcf3063b2b2588a762a4269537be2f4f8a111b5bcac0879bed Oct 06 10:19:15 crc kubenswrapper[4824]: I1006 10:19:15.647423 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"1166f1ef-43be-42e7-9d5f-84252fccfb69","Type":"ContainerStarted","Data":"2abda6592d7134dcf3063b2b2588a762a4269537be2f4f8a111b5bcac0879bed"} Oct 06 10:19:15 crc kubenswrapper[4824]: I1006 10:19:15.649507 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"754abe5c-3e1b-4a4a-b5a7-619326c06846","Type":"ContainerStarted","Data":"c39c8dc17484bd069c209cfff14b52719ab8944877ebf3dfb6af2d344e08073e"} Oct 06 10:19:15 crc kubenswrapper[4824]: I1006 10:19:15.649567 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"754abe5c-3e1b-4a4a-b5a7-619326c06846","Type":"ContainerStarted","Data":"1440cbc7b62c0b103ce6e36ee5ca1fe0dbc12c8a7ed5ba3de66d76dbaffcffbf"} Oct 06 10:19:15 crc kubenswrapper[4824]: I1006 10:19:15.649587 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"754abe5c-3e1b-4a4a-b5a7-619326c06846","Type":"ContainerStarted","Data":"815e58d1516a4b68f9a0fc4c070dbfeeb71e23c655767fbb525ad2b3bef3324c"} Oct 06 10:19:15 crc kubenswrapper[4824]: I1006 10:19:15.662139 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 06 10:19:15 crc kubenswrapper[4824]: I1006 10:19:15.662147 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2056be6d-e4b2-4f5e-91a7-ac5b2cb357ce","Type":"ContainerDied","Data":"c494e47b029f92169e5a3026fbf8da5e2515ce94d4201282d80a1f48afffecec"} Oct 06 10:19:15 crc kubenswrapper[4824]: I1006 10:19:15.662222 4824 scope.go:117] "RemoveContainer" containerID="ba945d970f2119d285f28c46084b1b59b9567dd3cd65490e0d58c9f33bf072a7" Oct 06 10:19:15 crc kubenswrapper[4824]: I1006 10:19:15.686862 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.686827258 podStartE2EDuration="2.686827258s" podCreationTimestamp="2025-10-06 10:19:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:19:15.677865992 +0000 UTC m=+1265.042288903" watchObservedRunningTime="2025-10-06 10:19:15.686827258 +0000 UTC m=+1265.051250139" Oct 06 10:19:15 crc kubenswrapper[4824]: I1006 10:19:15.751257 4824 scope.go:117] "RemoveContainer" containerID="80156743cc95e672e11830faa0c625af7a756c3dc73c8b5f5c1110102ff75d85" Oct 06 10:19:15 crc kubenswrapper[4824]: I1006 10:19:15.782582 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 06 10:19:15 crc kubenswrapper[4824]: I1006 10:19:15.794499 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 06 10:19:15 crc kubenswrapper[4824]: I1006 10:19:15.804080 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 06 10:19:15 crc kubenswrapper[4824]: E1006 10:19:15.804947 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2056be6d-e4b2-4f5e-91a7-ac5b2cb357ce" containerName="nova-api-api" Oct 06 10:19:15 crc kubenswrapper[4824]: I1006 10:19:15.804971 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="2056be6d-e4b2-4f5e-91a7-ac5b2cb357ce" containerName="nova-api-api" Oct 06 10:19:15 crc kubenswrapper[4824]: E1006 10:19:15.805004 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2056be6d-e4b2-4f5e-91a7-ac5b2cb357ce" containerName="nova-api-log" Oct 06 10:19:15 crc kubenswrapper[4824]: I1006 10:19:15.805014 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="2056be6d-e4b2-4f5e-91a7-ac5b2cb357ce" containerName="nova-api-log" Oct 06 10:19:15 crc kubenswrapper[4824]: I1006 10:19:15.805244 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="2056be6d-e4b2-4f5e-91a7-ac5b2cb357ce" containerName="nova-api-log" Oct 06 10:19:15 crc kubenswrapper[4824]: I1006 10:19:15.805261 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="2056be6d-e4b2-4f5e-91a7-ac5b2cb357ce" containerName="nova-api-api" Oct 06 10:19:15 crc kubenswrapper[4824]: I1006 10:19:15.810428 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 06 10:19:15 crc kubenswrapper[4824]: I1006 10:19:15.813701 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 06 10:19:15 crc kubenswrapper[4824]: I1006 10:19:15.815028 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Oct 06 10:19:15 crc kubenswrapper[4824]: I1006 10:19:15.815261 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 06 10:19:15 crc kubenswrapper[4824]: I1006 10:19:15.815369 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Oct 06 10:19:15 crc kubenswrapper[4824]: I1006 10:19:15.951507 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/20fa260b-d7c3-4f58-82e3-070c45d494f4-public-tls-certs\") pod \"nova-api-0\" (UID: \"20fa260b-d7c3-4f58-82e3-070c45d494f4\") " pod="openstack/nova-api-0" Oct 06 10:19:15 crc kubenswrapper[4824]: I1006 10:19:15.952029 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nb76z\" (UniqueName: \"kubernetes.io/projected/20fa260b-d7c3-4f58-82e3-070c45d494f4-kube-api-access-nb76z\") pod \"nova-api-0\" (UID: \"20fa260b-d7c3-4f58-82e3-070c45d494f4\") " pod="openstack/nova-api-0" Oct 06 10:19:15 crc kubenswrapper[4824]: I1006 10:19:15.952257 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/20fa260b-d7c3-4f58-82e3-070c45d494f4-internal-tls-certs\") pod \"nova-api-0\" (UID: \"20fa260b-d7c3-4f58-82e3-070c45d494f4\") " pod="openstack/nova-api-0" Oct 06 10:19:15 crc kubenswrapper[4824]: I1006 10:19:15.952430 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/20fa260b-d7c3-4f58-82e3-070c45d494f4-logs\") pod \"nova-api-0\" (UID: \"20fa260b-d7c3-4f58-82e3-070c45d494f4\") " pod="openstack/nova-api-0" Oct 06 10:19:15 crc kubenswrapper[4824]: I1006 10:19:15.952585 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/20fa260b-d7c3-4f58-82e3-070c45d494f4-config-data\") pod \"nova-api-0\" (UID: \"20fa260b-d7c3-4f58-82e3-070c45d494f4\") " pod="openstack/nova-api-0" Oct 06 10:19:15 crc kubenswrapper[4824]: I1006 10:19:15.952758 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20fa260b-d7c3-4f58-82e3-070c45d494f4-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"20fa260b-d7c3-4f58-82e3-070c45d494f4\") " pod="openstack/nova-api-0" Oct 06 10:19:16 crc kubenswrapper[4824]: I1006 10:19:16.054819 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/20fa260b-d7c3-4f58-82e3-070c45d494f4-public-tls-certs\") pod \"nova-api-0\" (UID: \"20fa260b-d7c3-4f58-82e3-070c45d494f4\") " pod="openstack/nova-api-0" Oct 06 10:19:16 crc kubenswrapper[4824]: I1006 10:19:16.055630 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nb76z\" (UniqueName: \"kubernetes.io/projected/20fa260b-d7c3-4f58-82e3-070c45d494f4-kube-api-access-nb76z\") pod \"nova-api-0\" (UID: \"20fa260b-d7c3-4f58-82e3-070c45d494f4\") " pod="openstack/nova-api-0" Oct 06 10:19:16 crc kubenswrapper[4824]: I1006 10:19:16.055793 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/20fa260b-d7c3-4f58-82e3-070c45d494f4-internal-tls-certs\") pod \"nova-api-0\" (UID: \"20fa260b-d7c3-4f58-82e3-070c45d494f4\") " pod="openstack/nova-api-0" Oct 06 10:19:16 crc kubenswrapper[4824]: I1006 10:19:16.055916 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/20fa260b-d7c3-4f58-82e3-070c45d494f4-logs\") pod \"nova-api-0\" (UID: \"20fa260b-d7c3-4f58-82e3-070c45d494f4\") " pod="openstack/nova-api-0" Oct 06 10:19:16 crc kubenswrapper[4824]: I1006 10:19:16.056042 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/20fa260b-d7c3-4f58-82e3-070c45d494f4-config-data\") pod \"nova-api-0\" (UID: \"20fa260b-d7c3-4f58-82e3-070c45d494f4\") " pod="openstack/nova-api-0" Oct 06 10:19:16 crc kubenswrapper[4824]: I1006 10:19:16.056183 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20fa260b-d7c3-4f58-82e3-070c45d494f4-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"20fa260b-d7c3-4f58-82e3-070c45d494f4\") " pod="openstack/nova-api-0" Oct 06 10:19:16 crc kubenswrapper[4824]: I1006 10:19:16.056410 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/20fa260b-d7c3-4f58-82e3-070c45d494f4-logs\") pod \"nova-api-0\" (UID: \"20fa260b-d7c3-4f58-82e3-070c45d494f4\") " pod="openstack/nova-api-0" Oct 06 10:19:16 crc kubenswrapper[4824]: I1006 10:19:16.063424 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/20fa260b-d7c3-4f58-82e3-070c45d494f4-public-tls-certs\") pod \"nova-api-0\" (UID: \"20fa260b-d7c3-4f58-82e3-070c45d494f4\") " pod="openstack/nova-api-0" Oct 06 10:19:16 crc kubenswrapper[4824]: I1006 10:19:16.063713 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20fa260b-d7c3-4f58-82e3-070c45d494f4-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"20fa260b-d7c3-4f58-82e3-070c45d494f4\") " pod="openstack/nova-api-0" Oct 06 10:19:16 crc kubenswrapper[4824]: I1006 10:19:16.065667 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/20fa260b-d7c3-4f58-82e3-070c45d494f4-internal-tls-certs\") pod \"nova-api-0\" (UID: \"20fa260b-d7c3-4f58-82e3-070c45d494f4\") " pod="openstack/nova-api-0" Oct 06 10:19:16 crc kubenswrapper[4824]: I1006 10:19:16.065969 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/20fa260b-d7c3-4f58-82e3-070c45d494f4-config-data\") pod \"nova-api-0\" (UID: \"20fa260b-d7c3-4f58-82e3-070c45d494f4\") " pod="openstack/nova-api-0" Oct 06 10:19:16 crc kubenswrapper[4824]: I1006 10:19:16.075012 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nb76z\" (UniqueName: \"kubernetes.io/projected/20fa260b-d7c3-4f58-82e3-070c45d494f4-kube-api-access-nb76z\") pod \"nova-api-0\" (UID: \"20fa260b-d7c3-4f58-82e3-070c45d494f4\") " pod="openstack/nova-api-0" Oct 06 10:19:16 crc kubenswrapper[4824]: I1006 10:19:16.137334 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 06 10:19:16 crc kubenswrapper[4824]: I1006 10:19:16.606585 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 06 10:19:16 crc kubenswrapper[4824]: I1006 10:19:16.702519 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"1166f1ef-43be-42e7-9d5f-84252fccfb69","Type":"ContainerStarted","Data":"d8dd793c895e47a7bcbad8112876e71cdc88fecc45b2291c0198acce9e5d78a2"} Oct 06 10:19:16 crc kubenswrapper[4824]: I1006 10:19:16.703886 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"20fa260b-d7c3-4f58-82e3-070c45d494f4","Type":"ContainerStarted","Data":"01222b00d10c5a4a5e047bdb3257adf0df31a01d5122e11ea0e4dfd98ea9033a"} Oct 06 10:19:16 crc kubenswrapper[4824]: I1006 10:19:16.726715 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.726694082 podStartE2EDuration="2.726694082s" podCreationTimestamp="2025-10-06 10:19:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:19:16.725395799 +0000 UTC m=+1266.089818660" watchObservedRunningTime="2025-10-06 10:19:16.726694082 +0000 UTC m=+1266.091116943" Oct 06 10:19:17 crc kubenswrapper[4824]: I1006 10:19:17.296444 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2056be6d-e4b2-4f5e-91a7-ac5b2cb357ce" path="/var/lib/kubelet/pods/2056be6d-e4b2-4f5e-91a7-ac5b2cb357ce/volumes" Oct 06 10:19:17 crc kubenswrapper[4824]: I1006 10:19:17.734220 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"20fa260b-d7c3-4f58-82e3-070c45d494f4","Type":"ContainerStarted","Data":"6dfbd4d6669a70b2a4ac5e3688a40197d19317800266efd8baf918048eccb5a4"} Oct 06 10:19:17 crc kubenswrapper[4824]: I1006 10:19:17.734704 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"20fa260b-d7c3-4f58-82e3-070c45d494f4","Type":"ContainerStarted","Data":"c6229379627bc825c4da941c7d81ea0e08af6f6e9344d2323a5463d1f74addc4"} Oct 06 10:19:17 crc kubenswrapper[4824]: I1006 10:19:17.777633 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.777594042 podStartE2EDuration="2.777594042s" podCreationTimestamp="2025-10-06 10:19:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:19:17.767812547 +0000 UTC m=+1267.132235418" watchObservedRunningTime="2025-10-06 10:19:17.777594042 +0000 UTC m=+1267.142016913" Oct 06 10:19:18 crc kubenswrapper[4824]: I1006 10:19:18.956791 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 06 10:19:18 crc kubenswrapper[4824]: I1006 10:19:18.957264 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 06 10:19:20 crc kubenswrapper[4824]: I1006 10:19:20.044245 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 06 10:19:23 crc kubenswrapper[4824]: I1006 10:19:23.955878 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 06 10:19:23 crc kubenswrapper[4824]: I1006 10:19:23.956707 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 06 10:19:24 crc kubenswrapper[4824]: I1006 10:19:24.972172 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="754abe5c-3e1b-4a4a-b5a7-619326c06846" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.207:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 06 10:19:24 crc kubenswrapper[4824]: I1006 10:19:24.972228 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="754abe5c-3e1b-4a4a-b5a7-619326c06846" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.207:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 06 10:19:25 crc kubenswrapper[4824]: I1006 10:19:25.043508 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 06 10:19:25 crc kubenswrapper[4824]: I1006 10:19:25.075596 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 06 10:19:25 crc kubenswrapper[4824]: I1006 10:19:25.867527 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 06 10:19:26 crc kubenswrapper[4824]: I1006 10:19:26.137806 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 06 10:19:26 crc kubenswrapper[4824]: I1006 10:19:26.137857 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 06 10:19:27 crc kubenswrapper[4824]: I1006 10:19:27.154299 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="20fa260b-d7c3-4f58-82e3-070c45d494f4" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.209:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 06 10:19:27 crc kubenswrapper[4824]: I1006 10:19:27.155785 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="20fa260b-d7c3-4f58-82e3-070c45d494f4" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.209:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 06 10:19:33 crc kubenswrapper[4824]: I1006 10:19:33.727159 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 06 10:19:33 crc kubenswrapper[4824]: I1006 10:19:33.961094 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 06 10:19:33 crc kubenswrapper[4824]: I1006 10:19:33.965202 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 06 10:19:33 crc kubenswrapper[4824]: I1006 10:19:33.974403 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 06 10:19:34 crc kubenswrapper[4824]: I1006 10:19:34.943342 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 06 10:19:36 crc kubenswrapper[4824]: I1006 10:19:36.144659 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 06 10:19:36 crc kubenswrapper[4824]: I1006 10:19:36.145352 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 06 10:19:36 crc kubenswrapper[4824]: I1006 10:19:36.147208 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 06 10:19:36 crc kubenswrapper[4824]: I1006 10:19:36.153558 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 06 10:19:36 crc kubenswrapper[4824]: I1006 10:19:36.954235 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 06 10:19:36 crc kubenswrapper[4824]: I1006 10:19:36.960904 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 06 10:19:45 crc kubenswrapper[4824]: I1006 10:19:45.719745 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 06 10:19:46 crc kubenswrapper[4824]: I1006 10:19:46.813099 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 06 10:19:50 crc kubenswrapper[4824]: I1006 10:19:50.711609 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="f7d047da-89e3-4cd4-bf02-e6a3e15c585f" containerName="rabbitmq" containerID="cri-o://f8d792090dd667a7c36525cb2a4beb43244cc576a1e4eafe83b7803d2d5cafb8" gracePeriod=604796 Oct 06 10:19:51 crc kubenswrapper[4824]: I1006 10:19:51.384324 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="c8dbad1e-02be-43c8-8f80-ab771ee81742" containerName="rabbitmq" containerID="cri-o://776dc359b5340bb21b168dc69fe78fcefb80aade02e11f0b963bbb9e8b58fb92" gracePeriod=604796 Oct 06 10:19:54 crc kubenswrapper[4824]: I1006 10:19:54.212913 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="f7d047da-89e3-4cd4-bf02-e6a3e15c585f" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.101:5671: connect: connection refused" Oct 06 10:19:54 crc kubenswrapper[4824]: I1006 10:19:54.503704 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="c8dbad1e-02be-43c8-8f80-ab771ee81742" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.102:5671: connect: connection refused" Oct 06 10:19:57 crc kubenswrapper[4824]: I1006 10:19:57.162327 4824 generic.go:334] "Generic (PLEG): container finished" podID="f7d047da-89e3-4cd4-bf02-e6a3e15c585f" containerID="f8d792090dd667a7c36525cb2a4beb43244cc576a1e4eafe83b7803d2d5cafb8" exitCode=0 Oct 06 10:19:57 crc kubenswrapper[4824]: I1006 10:19:57.162408 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"f7d047da-89e3-4cd4-bf02-e6a3e15c585f","Type":"ContainerDied","Data":"f8d792090dd667a7c36525cb2a4beb43244cc576a1e4eafe83b7803d2d5cafb8"} Oct 06 10:19:57 crc kubenswrapper[4824]: I1006 10:19:57.492117 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 06 10:19:57 crc kubenswrapper[4824]: I1006 10:19:57.644217 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8vcmv\" (UniqueName: \"kubernetes.io/projected/f7d047da-89e3-4cd4-bf02-e6a3e15c585f-kube-api-access-8vcmv\") pod \"f7d047da-89e3-4cd4-bf02-e6a3e15c585f\" (UID: \"f7d047da-89e3-4cd4-bf02-e6a3e15c585f\") " Oct 06 10:19:57 crc kubenswrapper[4824]: I1006 10:19:57.644344 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f7d047da-89e3-4cd4-bf02-e6a3e15c585f-rabbitmq-plugins\") pod \"f7d047da-89e3-4cd4-bf02-e6a3e15c585f\" (UID: \"f7d047da-89e3-4cd4-bf02-e6a3e15c585f\") " Oct 06 10:19:57 crc kubenswrapper[4824]: I1006 10:19:57.644380 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f7d047da-89e3-4cd4-bf02-e6a3e15c585f-rabbitmq-confd\") pod \"f7d047da-89e3-4cd4-bf02-e6a3e15c585f\" (UID: \"f7d047da-89e3-4cd4-bf02-e6a3e15c585f\") " Oct 06 10:19:57 crc kubenswrapper[4824]: I1006 10:19:57.644411 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f7d047da-89e3-4cd4-bf02-e6a3e15c585f-pod-info\") pod \"f7d047da-89e3-4cd4-bf02-e6a3e15c585f\" (UID: \"f7d047da-89e3-4cd4-bf02-e6a3e15c585f\") " Oct 06 10:19:57 crc kubenswrapper[4824]: I1006 10:19:57.644437 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f7d047da-89e3-4cd4-bf02-e6a3e15c585f-rabbitmq-erlang-cookie\") pod \"f7d047da-89e3-4cd4-bf02-e6a3e15c585f\" (UID: \"f7d047da-89e3-4cd4-bf02-e6a3e15c585f\") " Oct 06 10:19:57 crc kubenswrapper[4824]: I1006 10:19:57.644485 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f7d047da-89e3-4cd4-bf02-e6a3e15c585f-config-data\") pod \"f7d047da-89e3-4cd4-bf02-e6a3e15c585f\" (UID: \"f7d047da-89e3-4cd4-bf02-e6a3e15c585f\") " Oct 06 10:19:57 crc kubenswrapper[4824]: I1006 10:19:57.644554 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/f7d047da-89e3-4cd4-bf02-e6a3e15c585f-server-conf\") pod \"f7d047da-89e3-4cd4-bf02-e6a3e15c585f\" (UID: \"f7d047da-89e3-4cd4-bf02-e6a3e15c585f\") " Oct 06 10:19:57 crc kubenswrapper[4824]: I1006 10:19:57.644577 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f7d047da-89e3-4cd4-bf02-e6a3e15c585f-erlang-cookie-secret\") pod \"f7d047da-89e3-4cd4-bf02-e6a3e15c585f\" (UID: \"f7d047da-89e3-4cd4-bf02-e6a3e15c585f\") " Oct 06 10:19:57 crc kubenswrapper[4824]: I1006 10:19:57.644606 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"f7d047da-89e3-4cd4-bf02-e6a3e15c585f\" (UID: \"f7d047da-89e3-4cd4-bf02-e6a3e15c585f\") " Oct 06 10:19:57 crc kubenswrapper[4824]: I1006 10:19:57.644672 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f7d047da-89e3-4cd4-bf02-e6a3e15c585f-plugins-conf\") pod \"f7d047da-89e3-4cd4-bf02-e6a3e15c585f\" (UID: \"f7d047da-89e3-4cd4-bf02-e6a3e15c585f\") " Oct 06 10:19:57 crc kubenswrapper[4824]: I1006 10:19:57.644725 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/f7d047da-89e3-4cd4-bf02-e6a3e15c585f-rabbitmq-tls\") pod \"f7d047da-89e3-4cd4-bf02-e6a3e15c585f\" (UID: \"f7d047da-89e3-4cd4-bf02-e6a3e15c585f\") " Oct 06 10:19:57 crc kubenswrapper[4824]: I1006 10:19:57.645929 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f7d047da-89e3-4cd4-bf02-e6a3e15c585f-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "f7d047da-89e3-4cd4-bf02-e6a3e15c585f" (UID: "f7d047da-89e3-4cd4-bf02-e6a3e15c585f"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:19:57 crc kubenswrapper[4824]: I1006 10:19:57.646596 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f7d047da-89e3-4cd4-bf02-e6a3e15c585f-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "f7d047da-89e3-4cd4-bf02-e6a3e15c585f" (UID: "f7d047da-89e3-4cd4-bf02-e6a3e15c585f"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:19:57 crc kubenswrapper[4824]: I1006 10:19:57.646646 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f7d047da-89e3-4cd4-bf02-e6a3e15c585f-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "f7d047da-89e3-4cd4-bf02-e6a3e15c585f" (UID: "f7d047da-89e3-4cd4-bf02-e6a3e15c585f"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:19:57 crc kubenswrapper[4824]: I1006 10:19:57.654349 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/f7d047da-89e3-4cd4-bf02-e6a3e15c585f-pod-info" (OuterVolumeSpecName: "pod-info") pod "f7d047da-89e3-4cd4-bf02-e6a3e15c585f" (UID: "f7d047da-89e3-4cd4-bf02-e6a3e15c585f"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 06 10:19:57 crc kubenswrapper[4824]: I1006 10:19:57.654576 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7d047da-89e3-4cd4-bf02-e6a3e15c585f-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "f7d047da-89e3-4cd4-bf02-e6a3e15c585f" (UID: "f7d047da-89e3-4cd4-bf02-e6a3e15c585f"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:19:57 crc kubenswrapper[4824]: I1006 10:19:57.654735 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f7d047da-89e3-4cd4-bf02-e6a3e15c585f-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "f7d047da-89e3-4cd4-bf02-e6a3e15c585f" (UID: "f7d047da-89e3-4cd4-bf02-e6a3e15c585f"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:19:57 crc kubenswrapper[4824]: I1006 10:19:57.654843 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "persistence") pod "f7d047da-89e3-4cd4-bf02-e6a3e15c585f" (UID: "f7d047da-89e3-4cd4-bf02-e6a3e15c585f"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 06 10:19:57 crc kubenswrapper[4824]: I1006 10:19:57.658341 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f7d047da-89e3-4cd4-bf02-e6a3e15c585f-kube-api-access-8vcmv" (OuterVolumeSpecName: "kube-api-access-8vcmv") pod "f7d047da-89e3-4cd4-bf02-e6a3e15c585f" (UID: "f7d047da-89e3-4cd4-bf02-e6a3e15c585f"). InnerVolumeSpecName "kube-api-access-8vcmv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:19:57 crc kubenswrapper[4824]: I1006 10:19:57.680747 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f7d047da-89e3-4cd4-bf02-e6a3e15c585f-config-data" (OuterVolumeSpecName: "config-data") pod "f7d047da-89e3-4cd4-bf02-e6a3e15c585f" (UID: "f7d047da-89e3-4cd4-bf02-e6a3e15c585f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:19:57 crc kubenswrapper[4824]: I1006 10:19:57.736767 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f7d047da-89e3-4cd4-bf02-e6a3e15c585f-server-conf" (OuterVolumeSpecName: "server-conf") pod "f7d047da-89e3-4cd4-bf02-e6a3e15c585f" (UID: "f7d047da-89e3-4cd4-bf02-e6a3e15c585f"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:19:57 crc kubenswrapper[4824]: I1006 10:19:57.746923 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8vcmv\" (UniqueName: \"kubernetes.io/projected/f7d047da-89e3-4cd4-bf02-e6a3e15c585f-kube-api-access-8vcmv\") on node \"crc\" DevicePath \"\"" Oct 06 10:19:57 crc kubenswrapper[4824]: I1006 10:19:57.746966 4824 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f7d047da-89e3-4cd4-bf02-e6a3e15c585f-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Oct 06 10:19:57 crc kubenswrapper[4824]: I1006 10:19:57.746994 4824 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f7d047da-89e3-4cd4-bf02-e6a3e15c585f-pod-info\") on node \"crc\" DevicePath \"\"" Oct 06 10:19:57 crc kubenswrapper[4824]: I1006 10:19:57.747008 4824 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f7d047da-89e3-4cd4-bf02-e6a3e15c585f-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Oct 06 10:19:57 crc kubenswrapper[4824]: I1006 10:19:57.747020 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f7d047da-89e3-4cd4-bf02-e6a3e15c585f-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 10:19:57 crc kubenswrapper[4824]: I1006 10:19:57.747031 4824 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/f7d047da-89e3-4cd4-bf02-e6a3e15c585f-server-conf\") on node \"crc\" DevicePath \"\"" Oct 06 10:19:57 crc kubenswrapper[4824]: I1006 10:19:57.747041 4824 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f7d047da-89e3-4cd4-bf02-e6a3e15c585f-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Oct 06 10:19:57 crc kubenswrapper[4824]: I1006 10:19:57.747078 4824 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Oct 06 10:19:57 crc kubenswrapper[4824]: I1006 10:19:57.747089 4824 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f7d047da-89e3-4cd4-bf02-e6a3e15c585f-plugins-conf\") on node \"crc\" DevicePath \"\"" Oct 06 10:19:57 crc kubenswrapper[4824]: I1006 10:19:57.747098 4824 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/f7d047da-89e3-4cd4-bf02-e6a3e15c585f-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Oct 06 10:19:57 crc kubenswrapper[4824]: I1006 10:19:57.770590 4824 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Oct 06 10:19:57 crc kubenswrapper[4824]: I1006 10:19:57.810375 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f7d047da-89e3-4cd4-bf02-e6a3e15c585f-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "f7d047da-89e3-4cd4-bf02-e6a3e15c585f" (UID: "f7d047da-89e3-4cd4-bf02-e6a3e15c585f"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:19:57 crc kubenswrapper[4824]: I1006 10:19:57.848742 4824 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f7d047da-89e3-4cd4-bf02-e6a3e15c585f-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Oct 06 10:19:57 crc kubenswrapper[4824]: I1006 10:19:57.848787 4824 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Oct 06 10:19:57 crc kubenswrapper[4824]: I1006 10:19:57.961118 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.153249 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c8dbad1e-02be-43c8-8f80-ab771ee81742-plugins-conf\") pod \"c8dbad1e-02be-43c8-8f80-ab771ee81742\" (UID: \"c8dbad1e-02be-43c8-8f80-ab771ee81742\") " Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.153331 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c8dbad1e-02be-43c8-8f80-ab771ee81742-rabbitmq-erlang-cookie\") pod \"c8dbad1e-02be-43c8-8f80-ab771ee81742\" (UID: \"c8dbad1e-02be-43c8-8f80-ab771ee81742\") " Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.153419 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c8dbad1e-02be-43c8-8f80-ab771ee81742-rabbitmq-plugins\") pod \"c8dbad1e-02be-43c8-8f80-ab771ee81742\" (UID: \"c8dbad1e-02be-43c8-8f80-ab771ee81742\") " Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.153462 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c8dbad1e-02be-43c8-8f80-ab771ee81742-pod-info\") pod \"c8dbad1e-02be-43c8-8f80-ab771ee81742\" (UID: \"c8dbad1e-02be-43c8-8f80-ab771ee81742\") " Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.153515 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c8dbad1e-02be-43c8-8f80-ab771ee81742-rabbitmq-confd\") pod \"c8dbad1e-02be-43c8-8f80-ab771ee81742\" (UID: \"c8dbad1e-02be-43c8-8f80-ab771ee81742\") " Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.153567 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c8dbad1e-02be-43c8-8f80-ab771ee81742-erlang-cookie-secret\") pod \"c8dbad1e-02be-43c8-8f80-ab771ee81742\" (UID: \"c8dbad1e-02be-43c8-8f80-ab771ee81742\") " Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.153603 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c8dbad1e-02be-43c8-8f80-ab771ee81742-config-data\") pod \"c8dbad1e-02be-43c8-8f80-ab771ee81742\" (UID: \"c8dbad1e-02be-43c8-8f80-ab771ee81742\") " Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.153628 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c8dbad1e-02be-43c8-8f80-ab771ee81742-server-conf\") pod \"c8dbad1e-02be-43c8-8f80-ab771ee81742\" (UID: \"c8dbad1e-02be-43c8-8f80-ab771ee81742\") " Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.153649 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c8dbad1e-02be-43c8-8f80-ab771ee81742-rabbitmq-tls\") pod \"c8dbad1e-02be-43c8-8f80-ab771ee81742\" (UID: \"c8dbad1e-02be-43c8-8f80-ab771ee81742\") " Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.153667 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sdw27\" (UniqueName: \"kubernetes.io/projected/c8dbad1e-02be-43c8-8f80-ab771ee81742-kube-api-access-sdw27\") pod \"c8dbad1e-02be-43c8-8f80-ab771ee81742\" (UID: \"c8dbad1e-02be-43c8-8f80-ab771ee81742\") " Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.153696 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"c8dbad1e-02be-43c8-8f80-ab771ee81742\" (UID: \"c8dbad1e-02be-43c8-8f80-ab771ee81742\") " Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.154080 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c8dbad1e-02be-43c8-8f80-ab771ee81742-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "c8dbad1e-02be-43c8-8f80-ab771ee81742" (UID: "c8dbad1e-02be-43c8-8f80-ab771ee81742"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.154141 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c8dbad1e-02be-43c8-8f80-ab771ee81742-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "c8dbad1e-02be-43c8-8f80-ab771ee81742" (UID: "c8dbad1e-02be-43c8-8f80-ab771ee81742"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.154177 4824 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c8dbad1e-02be-43c8-8f80-ab771ee81742-plugins-conf\") on node \"crc\" DevicePath \"\"" Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.156775 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c8dbad1e-02be-43c8-8f80-ab771ee81742-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "c8dbad1e-02be-43c8-8f80-ab771ee81742" (UID: "c8dbad1e-02be-43c8-8f80-ab771ee81742"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.161752 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c8dbad1e-02be-43c8-8f80-ab771ee81742-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "c8dbad1e-02be-43c8-8f80-ab771ee81742" (UID: "c8dbad1e-02be-43c8-8f80-ab771ee81742"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.161762 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "persistence") pod "c8dbad1e-02be-43c8-8f80-ab771ee81742" (UID: "c8dbad1e-02be-43c8-8f80-ab771ee81742"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.162350 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c8dbad1e-02be-43c8-8f80-ab771ee81742-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "c8dbad1e-02be-43c8-8f80-ab771ee81742" (UID: "c8dbad1e-02be-43c8-8f80-ab771ee81742"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.164507 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c8dbad1e-02be-43c8-8f80-ab771ee81742-kube-api-access-sdw27" (OuterVolumeSpecName: "kube-api-access-sdw27") pod "c8dbad1e-02be-43c8-8f80-ab771ee81742" (UID: "c8dbad1e-02be-43c8-8f80-ab771ee81742"). InnerVolumeSpecName "kube-api-access-sdw27". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.165596 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/c8dbad1e-02be-43c8-8f80-ab771ee81742-pod-info" (OuterVolumeSpecName: "pod-info") pod "c8dbad1e-02be-43c8-8f80-ab771ee81742" (UID: "c8dbad1e-02be-43c8-8f80-ab771ee81742"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.180426 4824 generic.go:334] "Generic (PLEG): container finished" podID="c8dbad1e-02be-43c8-8f80-ab771ee81742" containerID="776dc359b5340bb21b168dc69fe78fcefb80aade02e11f0b963bbb9e8b58fb92" exitCode=0 Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.180502 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"c8dbad1e-02be-43c8-8f80-ab771ee81742","Type":"ContainerDied","Data":"776dc359b5340bb21b168dc69fe78fcefb80aade02e11f0b963bbb9e8b58fb92"} Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.180539 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"c8dbad1e-02be-43c8-8f80-ab771ee81742","Type":"ContainerDied","Data":"96dcce3401c3ff21f9fd96a662d06af63598676c2a64ef986e5f4bb7c1d4315d"} Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.180563 4824 scope.go:117] "RemoveContainer" containerID="776dc359b5340bb21b168dc69fe78fcefb80aade02e11f0b963bbb9e8b58fb92" Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.180760 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.186138 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"f7d047da-89e3-4cd4-bf02-e6a3e15c585f","Type":"ContainerDied","Data":"ae8f6eccb88670a7b087c7a985bb0482f09a525c27e50b6cba6d9b58e14310d6"} Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.186274 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.207255 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c8dbad1e-02be-43c8-8f80-ab771ee81742-config-data" (OuterVolumeSpecName: "config-data") pod "c8dbad1e-02be-43c8-8f80-ab771ee81742" (UID: "c8dbad1e-02be-43c8-8f80-ab771ee81742"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.237539 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c8dbad1e-02be-43c8-8f80-ab771ee81742-server-conf" (OuterVolumeSpecName: "server-conf") pod "c8dbad1e-02be-43c8-8f80-ab771ee81742" (UID: "c8dbad1e-02be-43c8-8f80-ab771ee81742"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.256365 4824 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c8dbad1e-02be-43c8-8f80-ab771ee81742-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.256398 4824 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c8dbad1e-02be-43c8-8f80-ab771ee81742-pod-info\") on node \"crc\" DevicePath \"\"" Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.256407 4824 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c8dbad1e-02be-43c8-8f80-ab771ee81742-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.256415 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c8dbad1e-02be-43c8-8f80-ab771ee81742-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.256424 4824 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c8dbad1e-02be-43c8-8f80-ab771ee81742-server-conf\") on node \"crc\" DevicePath \"\"" Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.256431 4824 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c8dbad1e-02be-43c8-8f80-ab771ee81742-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.256440 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sdw27\" (UniqueName: \"kubernetes.io/projected/c8dbad1e-02be-43c8-8f80-ab771ee81742-kube-api-access-sdw27\") on node \"crc\" DevicePath \"\"" Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.256466 4824 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.256475 4824 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c8dbad1e-02be-43c8-8f80-ab771ee81742-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.276824 4824 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.293453 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c8dbad1e-02be-43c8-8f80-ab771ee81742-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "c8dbad1e-02be-43c8-8f80-ab771ee81742" (UID: "c8dbad1e-02be-43c8-8f80-ab771ee81742"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.358311 4824 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c8dbad1e-02be-43c8-8f80-ab771ee81742-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.358345 4824 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.381884 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.386287 4824 scope.go:117] "RemoveContainer" containerID="9e782d23005bd2ab15abfb2afc9b231130f78b0f01da4901f27b58d3b1eb931d" Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.390549 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.410809 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Oct 06 10:19:58 crc kubenswrapper[4824]: E1006 10:19:58.411471 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8dbad1e-02be-43c8-8f80-ab771ee81742" containerName="rabbitmq" Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.411506 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8dbad1e-02be-43c8-8f80-ab771ee81742" containerName="rabbitmq" Oct 06 10:19:58 crc kubenswrapper[4824]: E1006 10:19:58.411519 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7d047da-89e3-4cd4-bf02-e6a3e15c585f" containerName="rabbitmq" Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.411540 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7d047da-89e3-4cd4-bf02-e6a3e15c585f" containerName="rabbitmq" Oct 06 10:19:58 crc kubenswrapper[4824]: E1006 10:19:58.411555 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8dbad1e-02be-43c8-8f80-ab771ee81742" containerName="setup-container" Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.411562 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8dbad1e-02be-43c8-8f80-ab771ee81742" containerName="setup-container" Oct 06 10:19:58 crc kubenswrapper[4824]: E1006 10:19:58.411588 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7d047da-89e3-4cd4-bf02-e6a3e15c585f" containerName="setup-container" Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.411598 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7d047da-89e3-4cd4-bf02-e6a3e15c585f" containerName="setup-container" Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.411830 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7d047da-89e3-4cd4-bf02-e6a3e15c585f" containerName="rabbitmq" Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.411866 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8dbad1e-02be-43c8-8f80-ab771ee81742" containerName="rabbitmq" Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.413224 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.415496 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.415828 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.416078 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-r49v9" Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.416773 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.416836 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.417015 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.417117 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.419260 4824 scope.go:117] "RemoveContainer" containerID="776dc359b5340bb21b168dc69fe78fcefb80aade02e11f0b963bbb9e8b58fb92" Oct 06 10:19:58 crc kubenswrapper[4824]: E1006 10:19:58.421183 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"776dc359b5340bb21b168dc69fe78fcefb80aade02e11f0b963bbb9e8b58fb92\": container with ID starting with 776dc359b5340bb21b168dc69fe78fcefb80aade02e11f0b963bbb9e8b58fb92 not found: ID does not exist" containerID="776dc359b5340bb21b168dc69fe78fcefb80aade02e11f0b963bbb9e8b58fb92" Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.421235 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"776dc359b5340bb21b168dc69fe78fcefb80aade02e11f0b963bbb9e8b58fb92"} err="failed to get container status \"776dc359b5340bb21b168dc69fe78fcefb80aade02e11f0b963bbb9e8b58fb92\": rpc error: code = NotFound desc = could not find container \"776dc359b5340bb21b168dc69fe78fcefb80aade02e11f0b963bbb9e8b58fb92\": container with ID starting with 776dc359b5340bb21b168dc69fe78fcefb80aade02e11f0b963bbb9e8b58fb92 not found: ID does not exist" Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.421264 4824 scope.go:117] "RemoveContainer" containerID="9e782d23005bd2ab15abfb2afc9b231130f78b0f01da4901f27b58d3b1eb931d" Oct 06 10:19:58 crc kubenswrapper[4824]: E1006 10:19:58.421725 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9e782d23005bd2ab15abfb2afc9b231130f78b0f01da4901f27b58d3b1eb931d\": container with ID starting with 9e782d23005bd2ab15abfb2afc9b231130f78b0f01da4901f27b58d3b1eb931d not found: ID does not exist" containerID="9e782d23005bd2ab15abfb2afc9b231130f78b0f01da4901f27b58d3b1eb931d" Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.421764 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9e782d23005bd2ab15abfb2afc9b231130f78b0f01da4901f27b58d3b1eb931d"} err="failed to get container status \"9e782d23005bd2ab15abfb2afc9b231130f78b0f01da4901f27b58d3b1eb931d\": rpc error: code = NotFound desc = could not find container \"9e782d23005bd2ab15abfb2afc9b231130f78b0f01da4901f27b58d3b1eb931d\": container with ID starting with 9e782d23005bd2ab15abfb2afc9b231130f78b0f01da4901f27b58d3b1eb931d not found: ID does not exist" Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.421790 4824 scope.go:117] "RemoveContainer" containerID="f8d792090dd667a7c36525cb2a4beb43244cc576a1e4eafe83b7803d2d5cafb8" Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.434687 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.447695 4824 scope.go:117] "RemoveContainer" containerID="2512559a4666f95711ee0e9d31717a5aa74ce89aab85c9eec580e792833cd962" Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.526034 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.541710 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.553732 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.556993 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.562475 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.562707 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.563084 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-px9kh" Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.563309 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.563725 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.562741 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.562497 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.567412 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.568244 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/9b7a49d4-5482-4da9-aad2-4aabf9584534-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"9b7a49d4-5482-4da9-aad2-4aabf9584534\") " pod="openstack/rabbitmq-server-0" Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.568353 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/9b7a49d4-5482-4da9-aad2-4aabf9584534-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"9b7a49d4-5482-4da9-aad2-4aabf9584534\") " pod="openstack/rabbitmq-server-0" Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.568388 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"9b7a49d4-5482-4da9-aad2-4aabf9584534\") " pod="openstack/rabbitmq-server-0" Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.568441 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/9b7a49d4-5482-4da9-aad2-4aabf9584534-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"9b7a49d4-5482-4da9-aad2-4aabf9584534\") " pod="openstack/rabbitmq-server-0" Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.568705 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/9b7a49d4-5482-4da9-aad2-4aabf9584534-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"9b7a49d4-5482-4da9-aad2-4aabf9584534\") " pod="openstack/rabbitmq-server-0" Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.568734 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/9b7a49d4-5482-4da9-aad2-4aabf9584534-pod-info\") pod \"rabbitmq-server-0\" (UID: \"9b7a49d4-5482-4da9-aad2-4aabf9584534\") " pod="openstack/rabbitmq-server-0" Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.568776 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/9b7a49d4-5482-4da9-aad2-4aabf9584534-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"9b7a49d4-5482-4da9-aad2-4aabf9584534\") " pod="openstack/rabbitmq-server-0" Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.568818 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/9b7a49d4-5482-4da9-aad2-4aabf9584534-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"9b7a49d4-5482-4da9-aad2-4aabf9584534\") " pod="openstack/rabbitmq-server-0" Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.568894 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9b7a49d4-5482-4da9-aad2-4aabf9584534-config-data\") pod \"rabbitmq-server-0\" (UID: \"9b7a49d4-5482-4da9-aad2-4aabf9584534\") " pod="openstack/rabbitmq-server-0" Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.568921 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mbhgc\" (UniqueName: \"kubernetes.io/projected/9b7a49d4-5482-4da9-aad2-4aabf9584534-kube-api-access-mbhgc\") pod \"rabbitmq-server-0\" (UID: \"9b7a49d4-5482-4da9-aad2-4aabf9584534\") " pod="openstack/rabbitmq-server-0" Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.568973 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/9b7a49d4-5482-4da9-aad2-4aabf9584534-server-conf\") pod \"rabbitmq-server-0\" (UID: \"9b7a49d4-5482-4da9-aad2-4aabf9584534\") " pod="openstack/rabbitmq-server-0" Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.670357 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0295abaf-72dd-46d7-9125-543921a6ef61-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"0295abaf-72dd-46d7-9125-543921a6ef61\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.670442 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/9b7a49d4-5482-4da9-aad2-4aabf9584534-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"9b7a49d4-5482-4da9-aad2-4aabf9584534\") " pod="openstack/rabbitmq-server-0" Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.670494 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0295abaf-72dd-46d7-9125-543921a6ef61-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"0295abaf-72dd-46d7-9125-543921a6ef61\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.670532 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0295abaf-72dd-46d7-9125-543921a6ef61-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"0295abaf-72dd-46d7-9125-543921a6ef61\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.670563 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9b7a49d4-5482-4da9-aad2-4aabf9584534-config-data\") pod \"rabbitmq-server-0\" (UID: \"9b7a49d4-5482-4da9-aad2-4aabf9584534\") " pod="openstack/rabbitmq-server-0" Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.670587 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mbhgc\" (UniqueName: \"kubernetes.io/projected/9b7a49d4-5482-4da9-aad2-4aabf9584534-kube-api-access-mbhgc\") pod \"rabbitmq-server-0\" (UID: \"9b7a49d4-5482-4da9-aad2-4aabf9584534\") " pod="openstack/rabbitmq-server-0" Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.670608 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0295abaf-72dd-46d7-9125-543921a6ef61-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"0295abaf-72dd-46d7-9125-543921a6ef61\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.670642 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bccc8\" (UniqueName: \"kubernetes.io/projected/0295abaf-72dd-46d7-9125-543921a6ef61-kube-api-access-bccc8\") pod \"rabbitmq-cell1-server-0\" (UID: \"0295abaf-72dd-46d7-9125-543921a6ef61\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.670670 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/9b7a49d4-5482-4da9-aad2-4aabf9584534-server-conf\") pod \"rabbitmq-server-0\" (UID: \"9b7a49d4-5482-4da9-aad2-4aabf9584534\") " pod="openstack/rabbitmq-server-0" Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.670715 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/9b7a49d4-5482-4da9-aad2-4aabf9584534-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"9b7a49d4-5482-4da9-aad2-4aabf9584534\") " pod="openstack/rabbitmq-server-0" Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.670747 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0295abaf-72dd-46d7-9125-543921a6ef61-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"0295abaf-72dd-46d7-9125-543921a6ef61\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.670769 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0295abaf-72dd-46d7-9125-543921a6ef61-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"0295abaf-72dd-46d7-9125-543921a6ef61\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.670824 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/9b7a49d4-5482-4da9-aad2-4aabf9584534-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"9b7a49d4-5482-4da9-aad2-4aabf9584534\") " pod="openstack/rabbitmq-server-0" Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.670853 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"9b7a49d4-5482-4da9-aad2-4aabf9584534\") " pod="openstack/rabbitmq-server-0" Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.670890 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/9b7a49d4-5482-4da9-aad2-4aabf9584534-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"9b7a49d4-5482-4da9-aad2-4aabf9584534\") " pod="openstack/rabbitmq-server-0" Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.670913 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0295abaf-72dd-46d7-9125-543921a6ef61-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"0295abaf-72dd-46d7-9125-543921a6ef61\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.670948 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0295abaf-72dd-46d7-9125-543921a6ef61-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"0295abaf-72dd-46d7-9125-543921a6ef61\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.671006 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0295abaf-72dd-46d7-9125-543921a6ef61-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"0295abaf-72dd-46d7-9125-543921a6ef61\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.671032 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/9b7a49d4-5482-4da9-aad2-4aabf9584534-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"9b7a49d4-5482-4da9-aad2-4aabf9584534\") " pod="openstack/rabbitmq-server-0" Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.671055 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/9b7a49d4-5482-4da9-aad2-4aabf9584534-pod-info\") pod \"rabbitmq-server-0\" (UID: \"9b7a49d4-5482-4da9-aad2-4aabf9584534\") " pod="openstack/rabbitmq-server-0" Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.671084 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"0295abaf-72dd-46d7-9125-543921a6ef61\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.671120 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/9b7a49d4-5482-4da9-aad2-4aabf9584534-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"9b7a49d4-5482-4da9-aad2-4aabf9584534\") " pod="openstack/rabbitmq-server-0" Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.672324 4824 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"9b7a49d4-5482-4da9-aad2-4aabf9584534\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/rabbitmq-server-0" Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.672429 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/9b7a49d4-5482-4da9-aad2-4aabf9584534-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"9b7a49d4-5482-4da9-aad2-4aabf9584534\") " pod="openstack/rabbitmq-server-0" Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.672846 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/9b7a49d4-5482-4da9-aad2-4aabf9584534-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"9b7a49d4-5482-4da9-aad2-4aabf9584534\") " pod="openstack/rabbitmq-server-0" Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.673352 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9b7a49d4-5482-4da9-aad2-4aabf9584534-config-data\") pod \"rabbitmq-server-0\" (UID: \"9b7a49d4-5482-4da9-aad2-4aabf9584534\") " pod="openstack/rabbitmq-server-0" Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.673353 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/9b7a49d4-5482-4da9-aad2-4aabf9584534-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"9b7a49d4-5482-4da9-aad2-4aabf9584534\") " pod="openstack/rabbitmq-server-0" Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.673406 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/9b7a49d4-5482-4da9-aad2-4aabf9584534-server-conf\") pod \"rabbitmq-server-0\" (UID: \"9b7a49d4-5482-4da9-aad2-4aabf9584534\") " pod="openstack/rabbitmq-server-0" Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.677800 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/9b7a49d4-5482-4da9-aad2-4aabf9584534-pod-info\") pod \"rabbitmq-server-0\" (UID: \"9b7a49d4-5482-4da9-aad2-4aabf9584534\") " pod="openstack/rabbitmq-server-0" Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.678451 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/9b7a49d4-5482-4da9-aad2-4aabf9584534-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"9b7a49d4-5482-4da9-aad2-4aabf9584534\") " pod="openstack/rabbitmq-server-0" Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.678883 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/9b7a49d4-5482-4da9-aad2-4aabf9584534-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"9b7a49d4-5482-4da9-aad2-4aabf9584534\") " pod="openstack/rabbitmq-server-0" Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.679113 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/9b7a49d4-5482-4da9-aad2-4aabf9584534-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"9b7a49d4-5482-4da9-aad2-4aabf9584534\") " pod="openstack/rabbitmq-server-0" Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.693392 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mbhgc\" (UniqueName: \"kubernetes.io/projected/9b7a49d4-5482-4da9-aad2-4aabf9584534-kube-api-access-mbhgc\") pod \"rabbitmq-server-0\" (UID: \"9b7a49d4-5482-4da9-aad2-4aabf9584534\") " pod="openstack/rabbitmq-server-0" Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.717309 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"9b7a49d4-5482-4da9-aad2-4aabf9584534\") " pod="openstack/rabbitmq-server-0" Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.736088 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.772841 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0295abaf-72dd-46d7-9125-543921a6ef61-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"0295abaf-72dd-46d7-9125-543921a6ef61\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.772891 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"0295abaf-72dd-46d7-9125-543921a6ef61\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.772935 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0295abaf-72dd-46d7-9125-543921a6ef61-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"0295abaf-72dd-46d7-9125-543921a6ef61\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.773007 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0295abaf-72dd-46d7-9125-543921a6ef61-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"0295abaf-72dd-46d7-9125-543921a6ef61\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.773045 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0295abaf-72dd-46d7-9125-543921a6ef61-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"0295abaf-72dd-46d7-9125-543921a6ef61\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.773071 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0295abaf-72dd-46d7-9125-543921a6ef61-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"0295abaf-72dd-46d7-9125-543921a6ef61\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.773104 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bccc8\" (UniqueName: \"kubernetes.io/projected/0295abaf-72dd-46d7-9125-543921a6ef61-kube-api-access-bccc8\") pod \"rabbitmq-cell1-server-0\" (UID: \"0295abaf-72dd-46d7-9125-543921a6ef61\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.773164 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0295abaf-72dd-46d7-9125-543921a6ef61-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"0295abaf-72dd-46d7-9125-543921a6ef61\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.773189 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0295abaf-72dd-46d7-9125-543921a6ef61-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"0295abaf-72dd-46d7-9125-543921a6ef61\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.773252 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0295abaf-72dd-46d7-9125-543921a6ef61-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"0295abaf-72dd-46d7-9125-543921a6ef61\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.773282 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0295abaf-72dd-46d7-9125-543921a6ef61-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"0295abaf-72dd-46d7-9125-543921a6ef61\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.773942 4824 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"0295abaf-72dd-46d7-9125-543921a6ef61\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/rabbitmq-cell1-server-0" Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.774189 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0295abaf-72dd-46d7-9125-543921a6ef61-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"0295abaf-72dd-46d7-9125-543921a6ef61\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.774398 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0295abaf-72dd-46d7-9125-543921a6ef61-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"0295abaf-72dd-46d7-9125-543921a6ef61\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.774483 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0295abaf-72dd-46d7-9125-543921a6ef61-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"0295abaf-72dd-46d7-9125-543921a6ef61\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.774791 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0295abaf-72dd-46d7-9125-543921a6ef61-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"0295abaf-72dd-46d7-9125-543921a6ef61\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.775817 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0295abaf-72dd-46d7-9125-543921a6ef61-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"0295abaf-72dd-46d7-9125-543921a6ef61\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.779594 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0295abaf-72dd-46d7-9125-543921a6ef61-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"0295abaf-72dd-46d7-9125-543921a6ef61\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.780196 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0295abaf-72dd-46d7-9125-543921a6ef61-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"0295abaf-72dd-46d7-9125-543921a6ef61\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.780592 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0295abaf-72dd-46d7-9125-543921a6ef61-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"0295abaf-72dd-46d7-9125-543921a6ef61\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.782110 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0295abaf-72dd-46d7-9125-543921a6ef61-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"0295abaf-72dd-46d7-9125-543921a6ef61\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.792935 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bccc8\" (UniqueName: \"kubernetes.io/projected/0295abaf-72dd-46d7-9125-543921a6ef61-kube-api-access-bccc8\") pod \"rabbitmq-cell1-server-0\" (UID: \"0295abaf-72dd-46d7-9125-543921a6ef61\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.819743 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"0295abaf-72dd-46d7-9125-543921a6ef61\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 10:19:58 crc kubenswrapper[4824]: I1006 10:19:58.887743 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 06 10:19:59 crc kubenswrapper[4824]: I1006 10:19:59.214124 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 06 10:19:59 crc kubenswrapper[4824]: W1006 10:19:59.223239 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9b7a49d4_5482_4da9_aad2_4aabf9584534.slice/crio-cd2c0f59761bc4806d8934820b3af43ff3892f5b695dbdd5d56695bd791d960c WatchSource:0}: Error finding container cd2c0f59761bc4806d8934820b3af43ff3892f5b695dbdd5d56695bd791d960c: Status 404 returned error can't find the container with id cd2c0f59761bc4806d8934820b3af43ff3892f5b695dbdd5d56695bd791d960c Oct 06 10:19:59 crc kubenswrapper[4824]: I1006 10:19:59.286531 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c8dbad1e-02be-43c8-8f80-ab771ee81742" path="/var/lib/kubelet/pods/c8dbad1e-02be-43c8-8f80-ab771ee81742/volumes" Oct 06 10:19:59 crc kubenswrapper[4824]: I1006 10:19:59.288483 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f7d047da-89e3-4cd4-bf02-e6a3e15c585f" path="/var/lib/kubelet/pods/f7d047da-89e3-4cd4-bf02-e6a3e15c585f/volumes" Oct 06 10:19:59 crc kubenswrapper[4824]: I1006 10:19:59.457482 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 06 10:19:59 crc kubenswrapper[4824]: W1006 10:19:59.462451 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0295abaf_72dd_46d7_9125_543921a6ef61.slice/crio-5990dcb81406c2d64dfc2a68fb02299ba5171412175b3c601403ac8014a5c84f WatchSource:0}: Error finding container 5990dcb81406c2d64dfc2a68fb02299ba5171412175b3c601403ac8014a5c84f: Status 404 returned error can't find the container with id 5990dcb81406c2d64dfc2a68fb02299ba5171412175b3c601403ac8014a5c84f Oct 06 10:20:00 crc kubenswrapper[4824]: I1006 10:20:00.209517 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"0295abaf-72dd-46d7-9125-543921a6ef61","Type":"ContainerStarted","Data":"5990dcb81406c2d64dfc2a68fb02299ba5171412175b3c601403ac8014a5c84f"} Oct 06 10:20:00 crc kubenswrapper[4824]: I1006 10:20:00.210650 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"9b7a49d4-5482-4da9-aad2-4aabf9584534","Type":"ContainerStarted","Data":"cd2c0f59761bc4806d8934820b3af43ff3892f5b695dbdd5d56695bd791d960c"} Oct 06 10:20:01 crc kubenswrapper[4824]: I1006 10:20:01.094231 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-67b789f86c-kbmnp"] Oct 06 10:20:01 crc kubenswrapper[4824]: I1006 10:20:01.096345 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67b789f86c-kbmnp" Oct 06 10:20:01 crc kubenswrapper[4824]: I1006 10:20:01.099146 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Oct 06 10:20:01 crc kubenswrapper[4824]: I1006 10:20:01.162476 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-67b789f86c-kbmnp"] Oct 06 10:20:01 crc kubenswrapper[4824]: I1006 10:20:01.224098 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ca8ae4b1-bb70-4990-96b2-65be0b3069f8-config\") pod \"dnsmasq-dns-67b789f86c-kbmnp\" (UID: \"ca8ae4b1-bb70-4990-96b2-65be0b3069f8\") " pod="openstack/dnsmasq-dns-67b789f86c-kbmnp" Oct 06 10:20:01 crc kubenswrapper[4824]: I1006 10:20:01.224189 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"9b7a49d4-5482-4da9-aad2-4aabf9584534","Type":"ContainerStarted","Data":"9eadc53b991c6ca7aee566971fb7e3f0f9f8d72fa8ce6ea321deca2e69c85275"} Oct 06 10:20:01 crc kubenswrapper[4824]: I1006 10:20:01.224192 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-swsb9\" (UniqueName: \"kubernetes.io/projected/ca8ae4b1-bb70-4990-96b2-65be0b3069f8-kube-api-access-swsb9\") pod \"dnsmasq-dns-67b789f86c-kbmnp\" (UID: \"ca8ae4b1-bb70-4990-96b2-65be0b3069f8\") " pod="openstack/dnsmasq-dns-67b789f86c-kbmnp" Oct 06 10:20:01 crc kubenswrapper[4824]: I1006 10:20:01.224294 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ca8ae4b1-bb70-4990-96b2-65be0b3069f8-ovsdbserver-sb\") pod \"dnsmasq-dns-67b789f86c-kbmnp\" (UID: \"ca8ae4b1-bb70-4990-96b2-65be0b3069f8\") " pod="openstack/dnsmasq-dns-67b789f86c-kbmnp" Oct 06 10:20:01 crc kubenswrapper[4824]: I1006 10:20:01.224490 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/ca8ae4b1-bb70-4990-96b2-65be0b3069f8-openstack-edpm-ipam\") pod \"dnsmasq-dns-67b789f86c-kbmnp\" (UID: \"ca8ae4b1-bb70-4990-96b2-65be0b3069f8\") " pod="openstack/dnsmasq-dns-67b789f86c-kbmnp" Oct 06 10:20:01 crc kubenswrapper[4824]: I1006 10:20:01.225814 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ca8ae4b1-bb70-4990-96b2-65be0b3069f8-ovsdbserver-nb\") pod \"dnsmasq-dns-67b789f86c-kbmnp\" (UID: \"ca8ae4b1-bb70-4990-96b2-65be0b3069f8\") " pod="openstack/dnsmasq-dns-67b789f86c-kbmnp" Oct 06 10:20:01 crc kubenswrapper[4824]: I1006 10:20:01.226006 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ca8ae4b1-bb70-4990-96b2-65be0b3069f8-dns-swift-storage-0\") pod \"dnsmasq-dns-67b789f86c-kbmnp\" (UID: \"ca8ae4b1-bb70-4990-96b2-65be0b3069f8\") " pod="openstack/dnsmasq-dns-67b789f86c-kbmnp" Oct 06 10:20:01 crc kubenswrapper[4824]: I1006 10:20:01.226246 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ca8ae4b1-bb70-4990-96b2-65be0b3069f8-dns-svc\") pod \"dnsmasq-dns-67b789f86c-kbmnp\" (UID: \"ca8ae4b1-bb70-4990-96b2-65be0b3069f8\") " pod="openstack/dnsmasq-dns-67b789f86c-kbmnp" Oct 06 10:20:01 crc kubenswrapper[4824]: I1006 10:20:01.328915 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/ca8ae4b1-bb70-4990-96b2-65be0b3069f8-openstack-edpm-ipam\") pod \"dnsmasq-dns-67b789f86c-kbmnp\" (UID: \"ca8ae4b1-bb70-4990-96b2-65be0b3069f8\") " pod="openstack/dnsmasq-dns-67b789f86c-kbmnp" Oct 06 10:20:01 crc kubenswrapper[4824]: I1006 10:20:01.329064 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ca8ae4b1-bb70-4990-96b2-65be0b3069f8-ovsdbserver-nb\") pod \"dnsmasq-dns-67b789f86c-kbmnp\" (UID: \"ca8ae4b1-bb70-4990-96b2-65be0b3069f8\") " pod="openstack/dnsmasq-dns-67b789f86c-kbmnp" Oct 06 10:20:01 crc kubenswrapper[4824]: I1006 10:20:01.329169 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ca8ae4b1-bb70-4990-96b2-65be0b3069f8-dns-swift-storage-0\") pod \"dnsmasq-dns-67b789f86c-kbmnp\" (UID: \"ca8ae4b1-bb70-4990-96b2-65be0b3069f8\") " pod="openstack/dnsmasq-dns-67b789f86c-kbmnp" Oct 06 10:20:01 crc kubenswrapper[4824]: I1006 10:20:01.329315 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ca8ae4b1-bb70-4990-96b2-65be0b3069f8-dns-svc\") pod \"dnsmasq-dns-67b789f86c-kbmnp\" (UID: \"ca8ae4b1-bb70-4990-96b2-65be0b3069f8\") " pod="openstack/dnsmasq-dns-67b789f86c-kbmnp" Oct 06 10:20:01 crc kubenswrapper[4824]: I1006 10:20:01.330080 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/ca8ae4b1-bb70-4990-96b2-65be0b3069f8-openstack-edpm-ipam\") pod \"dnsmasq-dns-67b789f86c-kbmnp\" (UID: \"ca8ae4b1-bb70-4990-96b2-65be0b3069f8\") " pod="openstack/dnsmasq-dns-67b789f86c-kbmnp" Oct 06 10:20:01 crc kubenswrapper[4824]: I1006 10:20:01.330226 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ca8ae4b1-bb70-4990-96b2-65be0b3069f8-ovsdbserver-nb\") pod \"dnsmasq-dns-67b789f86c-kbmnp\" (UID: \"ca8ae4b1-bb70-4990-96b2-65be0b3069f8\") " pod="openstack/dnsmasq-dns-67b789f86c-kbmnp" Oct 06 10:20:01 crc kubenswrapper[4824]: I1006 10:20:01.330638 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ca8ae4b1-bb70-4990-96b2-65be0b3069f8-dns-swift-storage-0\") pod \"dnsmasq-dns-67b789f86c-kbmnp\" (UID: \"ca8ae4b1-bb70-4990-96b2-65be0b3069f8\") " pod="openstack/dnsmasq-dns-67b789f86c-kbmnp" Oct 06 10:20:01 crc kubenswrapper[4824]: I1006 10:20:01.331183 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ca8ae4b1-bb70-4990-96b2-65be0b3069f8-config\") pod \"dnsmasq-dns-67b789f86c-kbmnp\" (UID: \"ca8ae4b1-bb70-4990-96b2-65be0b3069f8\") " pod="openstack/dnsmasq-dns-67b789f86c-kbmnp" Oct 06 10:20:01 crc kubenswrapper[4824]: I1006 10:20:01.331313 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-swsb9\" (UniqueName: \"kubernetes.io/projected/ca8ae4b1-bb70-4990-96b2-65be0b3069f8-kube-api-access-swsb9\") pod \"dnsmasq-dns-67b789f86c-kbmnp\" (UID: \"ca8ae4b1-bb70-4990-96b2-65be0b3069f8\") " pod="openstack/dnsmasq-dns-67b789f86c-kbmnp" Oct 06 10:20:01 crc kubenswrapper[4824]: I1006 10:20:01.331346 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ca8ae4b1-bb70-4990-96b2-65be0b3069f8-ovsdbserver-sb\") pod \"dnsmasq-dns-67b789f86c-kbmnp\" (UID: \"ca8ae4b1-bb70-4990-96b2-65be0b3069f8\") " pod="openstack/dnsmasq-dns-67b789f86c-kbmnp" Oct 06 10:20:01 crc kubenswrapper[4824]: I1006 10:20:01.331773 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ca8ae4b1-bb70-4990-96b2-65be0b3069f8-dns-svc\") pod \"dnsmasq-dns-67b789f86c-kbmnp\" (UID: \"ca8ae4b1-bb70-4990-96b2-65be0b3069f8\") " pod="openstack/dnsmasq-dns-67b789f86c-kbmnp" Oct 06 10:20:01 crc kubenswrapper[4824]: I1006 10:20:01.331926 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ca8ae4b1-bb70-4990-96b2-65be0b3069f8-config\") pod \"dnsmasq-dns-67b789f86c-kbmnp\" (UID: \"ca8ae4b1-bb70-4990-96b2-65be0b3069f8\") " pod="openstack/dnsmasq-dns-67b789f86c-kbmnp" Oct 06 10:20:01 crc kubenswrapper[4824]: I1006 10:20:01.332098 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ca8ae4b1-bb70-4990-96b2-65be0b3069f8-ovsdbserver-sb\") pod \"dnsmasq-dns-67b789f86c-kbmnp\" (UID: \"ca8ae4b1-bb70-4990-96b2-65be0b3069f8\") " pod="openstack/dnsmasq-dns-67b789f86c-kbmnp" Oct 06 10:20:01 crc kubenswrapper[4824]: I1006 10:20:01.353180 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-swsb9\" (UniqueName: \"kubernetes.io/projected/ca8ae4b1-bb70-4990-96b2-65be0b3069f8-kube-api-access-swsb9\") pod \"dnsmasq-dns-67b789f86c-kbmnp\" (UID: \"ca8ae4b1-bb70-4990-96b2-65be0b3069f8\") " pod="openstack/dnsmasq-dns-67b789f86c-kbmnp" Oct 06 10:20:01 crc kubenswrapper[4824]: I1006 10:20:01.420481 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67b789f86c-kbmnp" Oct 06 10:20:01 crc kubenswrapper[4824]: I1006 10:20:01.943521 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-67b789f86c-kbmnp"] Oct 06 10:20:01 crc kubenswrapper[4824]: W1006 10:20:01.951048 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podca8ae4b1_bb70_4990_96b2_65be0b3069f8.slice/crio-9a575afbd0194d8f94f2b3945f886e0104a02a14daa48eb335b5d5a95a03b64a WatchSource:0}: Error finding container 9a575afbd0194d8f94f2b3945f886e0104a02a14daa48eb335b5d5a95a03b64a: Status 404 returned error can't find the container with id 9a575afbd0194d8f94f2b3945f886e0104a02a14daa48eb335b5d5a95a03b64a Oct 06 10:20:02 crc kubenswrapper[4824]: I1006 10:20:02.234627 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"0295abaf-72dd-46d7-9125-543921a6ef61","Type":"ContainerStarted","Data":"e1d8499b9c4e9af73a75992df9e12c0941c1611bf57972474f0c399b39b189d2"} Oct 06 10:20:02 crc kubenswrapper[4824]: I1006 10:20:02.236719 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67b789f86c-kbmnp" event={"ID":"ca8ae4b1-bb70-4990-96b2-65be0b3069f8","Type":"ContainerStarted","Data":"9a575afbd0194d8f94f2b3945f886e0104a02a14daa48eb335b5d5a95a03b64a"} Oct 06 10:20:03 crc kubenswrapper[4824]: I1006 10:20:03.249067 4824 generic.go:334] "Generic (PLEG): container finished" podID="ca8ae4b1-bb70-4990-96b2-65be0b3069f8" containerID="e0ebbf37409574207b093c1fa522feee73803fc6b72dd3a9dcf69678bbd02d03" exitCode=0 Oct 06 10:20:03 crc kubenswrapper[4824]: I1006 10:20:03.249124 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67b789f86c-kbmnp" event={"ID":"ca8ae4b1-bb70-4990-96b2-65be0b3069f8","Type":"ContainerDied","Data":"e0ebbf37409574207b093c1fa522feee73803fc6b72dd3a9dcf69678bbd02d03"} Oct 06 10:20:04 crc kubenswrapper[4824]: I1006 10:20:04.262703 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67b789f86c-kbmnp" event={"ID":"ca8ae4b1-bb70-4990-96b2-65be0b3069f8","Type":"ContainerStarted","Data":"4783d4a5bc2c9004bae3fb840826b0abe9f2bbf6be58653959700e160720fc27"} Oct 06 10:20:04 crc kubenswrapper[4824]: I1006 10:20:04.263299 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-67b789f86c-kbmnp" Oct 06 10:20:04 crc kubenswrapper[4824]: I1006 10:20:04.302299 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-67b789f86c-kbmnp" podStartSLOduration=3.302275598 podStartE2EDuration="3.302275598s" podCreationTimestamp="2025-10-06 10:20:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:20:04.286521092 +0000 UTC m=+1313.650943953" watchObservedRunningTime="2025-10-06 10:20:04.302275598 +0000 UTC m=+1313.666698479" Oct 06 10:20:11 crc kubenswrapper[4824]: I1006 10:20:11.427374 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-67b789f86c-kbmnp" Oct 06 10:20:11 crc kubenswrapper[4824]: I1006 10:20:11.512037 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-pk86t"] Oct 06 10:20:11 crc kubenswrapper[4824]: I1006 10:20:11.512345 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-59cf4bdb65-pk86t" podUID="b940b381-928a-4f5d-b91e-0da628631f7c" containerName="dnsmasq-dns" containerID="cri-o://8071c18c77c91063dea49b3439242174955cb8078a65deffe8537ccc795d0394" gracePeriod=10 Oct 06 10:20:11 crc kubenswrapper[4824]: I1006 10:20:11.672766 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-cb6ffcf87-xt4l8"] Oct 06 10:20:11 crc kubenswrapper[4824]: I1006 10:20:11.674888 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cb6ffcf87-xt4l8" Oct 06 10:20:11 crc kubenswrapper[4824]: I1006 10:20:11.686049 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cb6ffcf87-xt4l8"] Oct 06 10:20:11 crc kubenswrapper[4824]: I1006 10:20:11.788475 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7648fd80-2fcf-4ddd-bb32-1dd33c86a0fb-ovsdbserver-nb\") pod \"dnsmasq-dns-cb6ffcf87-xt4l8\" (UID: \"7648fd80-2fcf-4ddd-bb32-1dd33c86a0fb\") " pod="openstack/dnsmasq-dns-cb6ffcf87-xt4l8" Oct 06 10:20:11 crc kubenswrapper[4824]: I1006 10:20:11.788870 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7rgk6\" (UniqueName: \"kubernetes.io/projected/7648fd80-2fcf-4ddd-bb32-1dd33c86a0fb-kube-api-access-7rgk6\") pod \"dnsmasq-dns-cb6ffcf87-xt4l8\" (UID: \"7648fd80-2fcf-4ddd-bb32-1dd33c86a0fb\") " pod="openstack/dnsmasq-dns-cb6ffcf87-xt4l8" Oct 06 10:20:11 crc kubenswrapper[4824]: I1006 10:20:11.788962 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7648fd80-2fcf-4ddd-bb32-1dd33c86a0fb-config\") pod \"dnsmasq-dns-cb6ffcf87-xt4l8\" (UID: \"7648fd80-2fcf-4ddd-bb32-1dd33c86a0fb\") " pod="openstack/dnsmasq-dns-cb6ffcf87-xt4l8" Oct 06 10:20:11 crc kubenswrapper[4824]: I1006 10:20:11.789026 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7648fd80-2fcf-4ddd-bb32-1dd33c86a0fb-ovsdbserver-sb\") pod \"dnsmasq-dns-cb6ffcf87-xt4l8\" (UID: \"7648fd80-2fcf-4ddd-bb32-1dd33c86a0fb\") " pod="openstack/dnsmasq-dns-cb6ffcf87-xt4l8" Oct 06 10:20:11 crc kubenswrapper[4824]: I1006 10:20:11.789057 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7648fd80-2fcf-4ddd-bb32-1dd33c86a0fb-dns-svc\") pod \"dnsmasq-dns-cb6ffcf87-xt4l8\" (UID: \"7648fd80-2fcf-4ddd-bb32-1dd33c86a0fb\") " pod="openstack/dnsmasq-dns-cb6ffcf87-xt4l8" Oct 06 10:20:11 crc kubenswrapper[4824]: I1006 10:20:11.789229 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/7648fd80-2fcf-4ddd-bb32-1dd33c86a0fb-openstack-edpm-ipam\") pod \"dnsmasq-dns-cb6ffcf87-xt4l8\" (UID: \"7648fd80-2fcf-4ddd-bb32-1dd33c86a0fb\") " pod="openstack/dnsmasq-dns-cb6ffcf87-xt4l8" Oct 06 10:20:11 crc kubenswrapper[4824]: I1006 10:20:11.789278 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7648fd80-2fcf-4ddd-bb32-1dd33c86a0fb-dns-swift-storage-0\") pod \"dnsmasq-dns-cb6ffcf87-xt4l8\" (UID: \"7648fd80-2fcf-4ddd-bb32-1dd33c86a0fb\") " pod="openstack/dnsmasq-dns-cb6ffcf87-xt4l8" Oct 06 10:20:11 crc kubenswrapper[4824]: I1006 10:20:11.892092 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7648fd80-2fcf-4ddd-bb32-1dd33c86a0fb-ovsdbserver-nb\") pod \"dnsmasq-dns-cb6ffcf87-xt4l8\" (UID: \"7648fd80-2fcf-4ddd-bb32-1dd33c86a0fb\") " pod="openstack/dnsmasq-dns-cb6ffcf87-xt4l8" Oct 06 10:20:11 crc kubenswrapper[4824]: I1006 10:20:11.892162 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7rgk6\" (UniqueName: \"kubernetes.io/projected/7648fd80-2fcf-4ddd-bb32-1dd33c86a0fb-kube-api-access-7rgk6\") pod \"dnsmasq-dns-cb6ffcf87-xt4l8\" (UID: \"7648fd80-2fcf-4ddd-bb32-1dd33c86a0fb\") " pod="openstack/dnsmasq-dns-cb6ffcf87-xt4l8" Oct 06 10:20:11 crc kubenswrapper[4824]: I1006 10:20:11.892225 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7648fd80-2fcf-4ddd-bb32-1dd33c86a0fb-config\") pod \"dnsmasq-dns-cb6ffcf87-xt4l8\" (UID: \"7648fd80-2fcf-4ddd-bb32-1dd33c86a0fb\") " pod="openstack/dnsmasq-dns-cb6ffcf87-xt4l8" Oct 06 10:20:11 crc kubenswrapper[4824]: I1006 10:20:11.892266 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7648fd80-2fcf-4ddd-bb32-1dd33c86a0fb-ovsdbserver-sb\") pod \"dnsmasq-dns-cb6ffcf87-xt4l8\" (UID: \"7648fd80-2fcf-4ddd-bb32-1dd33c86a0fb\") " pod="openstack/dnsmasq-dns-cb6ffcf87-xt4l8" Oct 06 10:20:11 crc kubenswrapper[4824]: I1006 10:20:11.892301 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7648fd80-2fcf-4ddd-bb32-1dd33c86a0fb-dns-svc\") pod \"dnsmasq-dns-cb6ffcf87-xt4l8\" (UID: \"7648fd80-2fcf-4ddd-bb32-1dd33c86a0fb\") " pod="openstack/dnsmasq-dns-cb6ffcf87-xt4l8" Oct 06 10:20:11 crc kubenswrapper[4824]: I1006 10:20:11.892400 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/7648fd80-2fcf-4ddd-bb32-1dd33c86a0fb-openstack-edpm-ipam\") pod \"dnsmasq-dns-cb6ffcf87-xt4l8\" (UID: \"7648fd80-2fcf-4ddd-bb32-1dd33c86a0fb\") " pod="openstack/dnsmasq-dns-cb6ffcf87-xt4l8" Oct 06 10:20:11 crc kubenswrapper[4824]: I1006 10:20:11.892440 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7648fd80-2fcf-4ddd-bb32-1dd33c86a0fb-dns-swift-storage-0\") pod \"dnsmasq-dns-cb6ffcf87-xt4l8\" (UID: \"7648fd80-2fcf-4ddd-bb32-1dd33c86a0fb\") " pod="openstack/dnsmasq-dns-cb6ffcf87-xt4l8" Oct 06 10:20:11 crc kubenswrapper[4824]: I1006 10:20:11.893633 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7648fd80-2fcf-4ddd-bb32-1dd33c86a0fb-dns-swift-storage-0\") pod \"dnsmasq-dns-cb6ffcf87-xt4l8\" (UID: \"7648fd80-2fcf-4ddd-bb32-1dd33c86a0fb\") " pod="openstack/dnsmasq-dns-cb6ffcf87-xt4l8" Oct 06 10:20:11 crc kubenswrapper[4824]: I1006 10:20:11.893667 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7648fd80-2fcf-4ddd-bb32-1dd33c86a0fb-ovsdbserver-nb\") pod \"dnsmasq-dns-cb6ffcf87-xt4l8\" (UID: \"7648fd80-2fcf-4ddd-bb32-1dd33c86a0fb\") " pod="openstack/dnsmasq-dns-cb6ffcf87-xt4l8" Oct 06 10:20:11 crc kubenswrapper[4824]: I1006 10:20:11.893878 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7648fd80-2fcf-4ddd-bb32-1dd33c86a0fb-dns-svc\") pod \"dnsmasq-dns-cb6ffcf87-xt4l8\" (UID: \"7648fd80-2fcf-4ddd-bb32-1dd33c86a0fb\") " pod="openstack/dnsmasq-dns-cb6ffcf87-xt4l8" Oct 06 10:20:11 crc kubenswrapper[4824]: I1006 10:20:11.894034 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/7648fd80-2fcf-4ddd-bb32-1dd33c86a0fb-openstack-edpm-ipam\") pod \"dnsmasq-dns-cb6ffcf87-xt4l8\" (UID: \"7648fd80-2fcf-4ddd-bb32-1dd33c86a0fb\") " pod="openstack/dnsmasq-dns-cb6ffcf87-xt4l8" Oct 06 10:20:11 crc kubenswrapper[4824]: I1006 10:20:11.893967 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7648fd80-2fcf-4ddd-bb32-1dd33c86a0fb-config\") pod \"dnsmasq-dns-cb6ffcf87-xt4l8\" (UID: \"7648fd80-2fcf-4ddd-bb32-1dd33c86a0fb\") " pod="openstack/dnsmasq-dns-cb6ffcf87-xt4l8" Oct 06 10:20:11 crc kubenswrapper[4824]: I1006 10:20:11.894609 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7648fd80-2fcf-4ddd-bb32-1dd33c86a0fb-ovsdbserver-sb\") pod \"dnsmasq-dns-cb6ffcf87-xt4l8\" (UID: \"7648fd80-2fcf-4ddd-bb32-1dd33c86a0fb\") " pod="openstack/dnsmasq-dns-cb6ffcf87-xt4l8" Oct 06 10:20:11 crc kubenswrapper[4824]: I1006 10:20:11.917264 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7rgk6\" (UniqueName: \"kubernetes.io/projected/7648fd80-2fcf-4ddd-bb32-1dd33c86a0fb-kube-api-access-7rgk6\") pod \"dnsmasq-dns-cb6ffcf87-xt4l8\" (UID: \"7648fd80-2fcf-4ddd-bb32-1dd33c86a0fb\") " pod="openstack/dnsmasq-dns-cb6ffcf87-xt4l8" Oct 06 10:20:12 crc kubenswrapper[4824]: I1006 10:20:12.044448 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cb6ffcf87-xt4l8" Oct 06 10:20:12 crc kubenswrapper[4824]: I1006 10:20:12.362751 4824 generic.go:334] "Generic (PLEG): container finished" podID="b940b381-928a-4f5d-b91e-0da628631f7c" containerID="8071c18c77c91063dea49b3439242174955cb8078a65deffe8537ccc795d0394" exitCode=0 Oct 06 10:20:12 crc kubenswrapper[4824]: I1006 10:20:12.362861 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cf4bdb65-pk86t" event={"ID":"b940b381-928a-4f5d-b91e-0da628631f7c","Type":"ContainerDied","Data":"8071c18c77c91063dea49b3439242174955cb8078a65deffe8537ccc795d0394"} Oct 06 10:20:12 crc kubenswrapper[4824]: I1006 10:20:12.512545 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cb6ffcf87-xt4l8"] Oct 06 10:20:12 crc kubenswrapper[4824]: W1006 10:20:12.522115 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7648fd80_2fcf_4ddd_bb32_1dd33c86a0fb.slice/crio-1c82c350f55f1a51e276fae781f94fbb3f11f184a35fbe81aef1bb83be26627b WatchSource:0}: Error finding container 1c82c350f55f1a51e276fae781f94fbb3f11f184a35fbe81aef1bb83be26627b: Status 404 returned error can't find the container with id 1c82c350f55f1a51e276fae781f94fbb3f11f184a35fbe81aef1bb83be26627b Oct 06 10:20:12 crc kubenswrapper[4824]: I1006 10:20:12.561230 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59cf4bdb65-pk86t" Oct 06 10:20:12 crc kubenswrapper[4824]: I1006 10:20:12.723243 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b940b381-928a-4f5d-b91e-0da628631f7c-dns-svc\") pod \"b940b381-928a-4f5d-b91e-0da628631f7c\" (UID: \"b940b381-928a-4f5d-b91e-0da628631f7c\") " Oct 06 10:20:12 crc kubenswrapper[4824]: I1006 10:20:12.723872 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b940b381-928a-4f5d-b91e-0da628631f7c-ovsdbserver-sb\") pod \"b940b381-928a-4f5d-b91e-0da628631f7c\" (UID: \"b940b381-928a-4f5d-b91e-0da628631f7c\") " Oct 06 10:20:12 crc kubenswrapper[4824]: I1006 10:20:12.723905 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b940b381-928a-4f5d-b91e-0da628631f7c-ovsdbserver-nb\") pod \"b940b381-928a-4f5d-b91e-0da628631f7c\" (UID: \"b940b381-928a-4f5d-b91e-0da628631f7c\") " Oct 06 10:20:12 crc kubenswrapper[4824]: I1006 10:20:12.723929 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b940b381-928a-4f5d-b91e-0da628631f7c-config\") pod \"b940b381-928a-4f5d-b91e-0da628631f7c\" (UID: \"b940b381-928a-4f5d-b91e-0da628631f7c\") " Oct 06 10:20:12 crc kubenswrapper[4824]: I1006 10:20:12.724001 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b940b381-928a-4f5d-b91e-0da628631f7c-dns-swift-storage-0\") pod \"b940b381-928a-4f5d-b91e-0da628631f7c\" (UID: \"b940b381-928a-4f5d-b91e-0da628631f7c\") " Oct 06 10:20:12 crc kubenswrapper[4824]: I1006 10:20:12.724192 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hdrlh\" (UniqueName: \"kubernetes.io/projected/b940b381-928a-4f5d-b91e-0da628631f7c-kube-api-access-hdrlh\") pod \"b940b381-928a-4f5d-b91e-0da628631f7c\" (UID: \"b940b381-928a-4f5d-b91e-0da628631f7c\") " Oct 06 10:20:12 crc kubenswrapper[4824]: I1006 10:20:12.749035 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b940b381-928a-4f5d-b91e-0da628631f7c-kube-api-access-hdrlh" (OuterVolumeSpecName: "kube-api-access-hdrlh") pod "b940b381-928a-4f5d-b91e-0da628631f7c" (UID: "b940b381-928a-4f5d-b91e-0da628631f7c"). InnerVolumeSpecName "kube-api-access-hdrlh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:20:12 crc kubenswrapper[4824]: I1006 10:20:12.830392 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hdrlh\" (UniqueName: \"kubernetes.io/projected/b940b381-928a-4f5d-b91e-0da628631f7c-kube-api-access-hdrlh\") on node \"crc\" DevicePath \"\"" Oct 06 10:20:12 crc kubenswrapper[4824]: I1006 10:20:12.853483 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b940b381-928a-4f5d-b91e-0da628631f7c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b940b381-928a-4f5d-b91e-0da628631f7c" (UID: "b940b381-928a-4f5d-b91e-0da628631f7c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:20:12 crc kubenswrapper[4824]: I1006 10:20:12.868203 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b940b381-928a-4f5d-b91e-0da628631f7c-config" (OuterVolumeSpecName: "config") pod "b940b381-928a-4f5d-b91e-0da628631f7c" (UID: "b940b381-928a-4f5d-b91e-0da628631f7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:20:12 crc kubenswrapper[4824]: I1006 10:20:12.876264 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b940b381-928a-4f5d-b91e-0da628631f7c-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "b940b381-928a-4f5d-b91e-0da628631f7c" (UID: "b940b381-928a-4f5d-b91e-0da628631f7c"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:20:12 crc kubenswrapper[4824]: I1006 10:20:12.880368 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b940b381-928a-4f5d-b91e-0da628631f7c-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "b940b381-928a-4f5d-b91e-0da628631f7c" (UID: "b940b381-928a-4f5d-b91e-0da628631f7c"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:20:12 crc kubenswrapper[4824]: I1006 10:20:12.885542 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b940b381-928a-4f5d-b91e-0da628631f7c-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "b940b381-928a-4f5d-b91e-0da628631f7c" (UID: "b940b381-928a-4f5d-b91e-0da628631f7c"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:20:12 crc kubenswrapper[4824]: I1006 10:20:12.935459 4824 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b940b381-928a-4f5d-b91e-0da628631f7c-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 06 10:20:12 crc kubenswrapper[4824]: I1006 10:20:12.935494 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b940b381-928a-4f5d-b91e-0da628631f7c-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 06 10:20:12 crc kubenswrapper[4824]: I1006 10:20:12.935504 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b940b381-928a-4f5d-b91e-0da628631f7c-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 06 10:20:12 crc kubenswrapper[4824]: I1006 10:20:12.935512 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b940b381-928a-4f5d-b91e-0da628631f7c-config\") on node \"crc\" DevicePath \"\"" Oct 06 10:20:12 crc kubenswrapper[4824]: I1006 10:20:12.935522 4824 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b940b381-928a-4f5d-b91e-0da628631f7c-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 06 10:20:13 crc kubenswrapper[4824]: I1006 10:20:13.375106 4824 generic.go:334] "Generic (PLEG): container finished" podID="7648fd80-2fcf-4ddd-bb32-1dd33c86a0fb" containerID="233ace83f02b302178a09b8d951b7c581e39cfcc93aad78a86bafa71ab683f77" exitCode=0 Oct 06 10:20:13 crc kubenswrapper[4824]: I1006 10:20:13.375184 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cb6ffcf87-xt4l8" event={"ID":"7648fd80-2fcf-4ddd-bb32-1dd33c86a0fb","Type":"ContainerDied","Data":"233ace83f02b302178a09b8d951b7c581e39cfcc93aad78a86bafa71ab683f77"} Oct 06 10:20:13 crc kubenswrapper[4824]: I1006 10:20:13.375849 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cb6ffcf87-xt4l8" event={"ID":"7648fd80-2fcf-4ddd-bb32-1dd33c86a0fb","Type":"ContainerStarted","Data":"1c82c350f55f1a51e276fae781f94fbb3f11f184a35fbe81aef1bb83be26627b"} Oct 06 10:20:13 crc kubenswrapper[4824]: I1006 10:20:13.378717 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cf4bdb65-pk86t" event={"ID":"b940b381-928a-4f5d-b91e-0da628631f7c","Type":"ContainerDied","Data":"06242b7e1651eb30be31bba669ae59149898c09bd9ae14975275ecc623b0fddf"} Oct 06 10:20:13 crc kubenswrapper[4824]: I1006 10:20:13.378789 4824 scope.go:117] "RemoveContainer" containerID="8071c18c77c91063dea49b3439242174955cb8078a65deffe8537ccc795d0394" Oct 06 10:20:13 crc kubenswrapper[4824]: I1006 10:20:13.378999 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59cf4bdb65-pk86t" Oct 06 10:20:13 crc kubenswrapper[4824]: I1006 10:20:13.449290 4824 scope.go:117] "RemoveContainer" containerID="10ca06c8cefd1e8577231fd8670f5e205b5d1475d91bb02a7c3168387210b59a" Oct 06 10:20:13 crc kubenswrapper[4824]: I1006 10:20:13.458489 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-pk86t"] Oct 06 10:20:13 crc kubenswrapper[4824]: I1006 10:20:13.487620 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-pk86t"] Oct 06 10:20:14 crc kubenswrapper[4824]: I1006 10:20:14.393414 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cb6ffcf87-xt4l8" event={"ID":"7648fd80-2fcf-4ddd-bb32-1dd33c86a0fb","Type":"ContainerStarted","Data":"71d1b716847def962b3db93c03186ac987b02481256b2ec27e5d9cfc12d3e111"} Oct 06 10:20:14 crc kubenswrapper[4824]: I1006 10:20:14.394141 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-cb6ffcf87-xt4l8" Oct 06 10:20:14 crc kubenswrapper[4824]: I1006 10:20:14.416687 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-cb6ffcf87-xt4l8" podStartSLOduration=3.416664601 podStartE2EDuration="3.416664601s" podCreationTimestamp="2025-10-06 10:20:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:20:14.412878797 +0000 UTC m=+1323.777301668" watchObservedRunningTime="2025-10-06 10:20:14.416664601 +0000 UTC m=+1323.781087472" Oct 06 10:20:15 crc kubenswrapper[4824]: I1006 10:20:15.293062 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b940b381-928a-4f5d-b91e-0da628631f7c" path="/var/lib/kubelet/pods/b940b381-928a-4f5d-b91e-0da628631f7c/volumes" Oct 06 10:20:22 crc kubenswrapper[4824]: I1006 10:20:22.046134 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-cb6ffcf87-xt4l8" Oct 06 10:20:22 crc kubenswrapper[4824]: I1006 10:20:22.113938 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-67b789f86c-kbmnp"] Oct 06 10:20:22 crc kubenswrapper[4824]: I1006 10:20:22.114274 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-67b789f86c-kbmnp" podUID="ca8ae4b1-bb70-4990-96b2-65be0b3069f8" containerName="dnsmasq-dns" containerID="cri-o://4783d4a5bc2c9004bae3fb840826b0abe9f2bbf6be58653959700e160720fc27" gracePeriod=10 Oct 06 10:20:22 crc kubenswrapper[4824]: I1006 10:20:22.484721 4824 generic.go:334] "Generic (PLEG): container finished" podID="ca8ae4b1-bb70-4990-96b2-65be0b3069f8" containerID="4783d4a5bc2c9004bae3fb840826b0abe9f2bbf6be58653959700e160720fc27" exitCode=0 Oct 06 10:20:22 crc kubenswrapper[4824]: I1006 10:20:22.484962 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67b789f86c-kbmnp" event={"ID":"ca8ae4b1-bb70-4990-96b2-65be0b3069f8","Type":"ContainerDied","Data":"4783d4a5bc2c9004bae3fb840826b0abe9f2bbf6be58653959700e160720fc27"} Oct 06 10:20:22 crc kubenswrapper[4824]: I1006 10:20:22.668604 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67b789f86c-kbmnp" Oct 06 10:20:22 crc kubenswrapper[4824]: I1006 10:20:22.762811 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ca8ae4b1-bb70-4990-96b2-65be0b3069f8-dns-svc\") pod \"ca8ae4b1-bb70-4990-96b2-65be0b3069f8\" (UID: \"ca8ae4b1-bb70-4990-96b2-65be0b3069f8\") " Oct 06 10:20:22 crc kubenswrapper[4824]: I1006 10:20:22.763386 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ca8ae4b1-bb70-4990-96b2-65be0b3069f8-ovsdbserver-sb\") pod \"ca8ae4b1-bb70-4990-96b2-65be0b3069f8\" (UID: \"ca8ae4b1-bb70-4990-96b2-65be0b3069f8\") " Oct 06 10:20:22 crc kubenswrapper[4824]: I1006 10:20:22.763440 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/ca8ae4b1-bb70-4990-96b2-65be0b3069f8-openstack-edpm-ipam\") pod \"ca8ae4b1-bb70-4990-96b2-65be0b3069f8\" (UID: \"ca8ae4b1-bb70-4990-96b2-65be0b3069f8\") " Oct 06 10:20:22 crc kubenswrapper[4824]: I1006 10:20:22.763495 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ca8ae4b1-bb70-4990-96b2-65be0b3069f8-dns-swift-storage-0\") pod \"ca8ae4b1-bb70-4990-96b2-65be0b3069f8\" (UID: \"ca8ae4b1-bb70-4990-96b2-65be0b3069f8\") " Oct 06 10:20:22 crc kubenswrapper[4824]: I1006 10:20:22.763569 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ca8ae4b1-bb70-4990-96b2-65be0b3069f8-config\") pod \"ca8ae4b1-bb70-4990-96b2-65be0b3069f8\" (UID: \"ca8ae4b1-bb70-4990-96b2-65be0b3069f8\") " Oct 06 10:20:22 crc kubenswrapper[4824]: I1006 10:20:22.763631 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-swsb9\" (UniqueName: \"kubernetes.io/projected/ca8ae4b1-bb70-4990-96b2-65be0b3069f8-kube-api-access-swsb9\") pod \"ca8ae4b1-bb70-4990-96b2-65be0b3069f8\" (UID: \"ca8ae4b1-bb70-4990-96b2-65be0b3069f8\") " Oct 06 10:20:22 crc kubenswrapper[4824]: I1006 10:20:22.763667 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ca8ae4b1-bb70-4990-96b2-65be0b3069f8-ovsdbserver-nb\") pod \"ca8ae4b1-bb70-4990-96b2-65be0b3069f8\" (UID: \"ca8ae4b1-bb70-4990-96b2-65be0b3069f8\") " Oct 06 10:20:22 crc kubenswrapper[4824]: I1006 10:20:22.770592 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ca8ae4b1-bb70-4990-96b2-65be0b3069f8-kube-api-access-swsb9" (OuterVolumeSpecName: "kube-api-access-swsb9") pod "ca8ae4b1-bb70-4990-96b2-65be0b3069f8" (UID: "ca8ae4b1-bb70-4990-96b2-65be0b3069f8"). InnerVolumeSpecName "kube-api-access-swsb9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:20:22 crc kubenswrapper[4824]: I1006 10:20:22.832699 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ca8ae4b1-bb70-4990-96b2-65be0b3069f8-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "ca8ae4b1-bb70-4990-96b2-65be0b3069f8" (UID: "ca8ae4b1-bb70-4990-96b2-65be0b3069f8"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:20:22 crc kubenswrapper[4824]: I1006 10:20:22.837009 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ca8ae4b1-bb70-4990-96b2-65be0b3069f8-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "ca8ae4b1-bb70-4990-96b2-65be0b3069f8" (UID: "ca8ae4b1-bb70-4990-96b2-65be0b3069f8"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:20:22 crc kubenswrapper[4824]: I1006 10:20:22.842719 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ca8ae4b1-bb70-4990-96b2-65be0b3069f8-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ca8ae4b1-bb70-4990-96b2-65be0b3069f8" (UID: "ca8ae4b1-bb70-4990-96b2-65be0b3069f8"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:20:22 crc kubenswrapper[4824]: I1006 10:20:22.844478 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ca8ae4b1-bb70-4990-96b2-65be0b3069f8-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "ca8ae4b1-bb70-4990-96b2-65be0b3069f8" (UID: "ca8ae4b1-bb70-4990-96b2-65be0b3069f8"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:20:22 crc kubenswrapper[4824]: I1006 10:20:22.849461 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ca8ae4b1-bb70-4990-96b2-65be0b3069f8-config" (OuterVolumeSpecName: "config") pod "ca8ae4b1-bb70-4990-96b2-65be0b3069f8" (UID: "ca8ae4b1-bb70-4990-96b2-65be0b3069f8"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:20:22 crc kubenswrapper[4824]: I1006 10:20:22.849569 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ca8ae4b1-bb70-4990-96b2-65be0b3069f8-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "ca8ae4b1-bb70-4990-96b2-65be0b3069f8" (UID: "ca8ae4b1-bb70-4990-96b2-65be0b3069f8"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:20:22 crc kubenswrapper[4824]: I1006 10:20:22.867349 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ca8ae4b1-bb70-4990-96b2-65be0b3069f8-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 06 10:20:22 crc kubenswrapper[4824]: I1006 10:20:22.867383 4824 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/ca8ae4b1-bb70-4990-96b2-65be0b3069f8-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Oct 06 10:20:22 crc kubenswrapper[4824]: I1006 10:20:22.867393 4824 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ca8ae4b1-bb70-4990-96b2-65be0b3069f8-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 06 10:20:22 crc kubenswrapper[4824]: I1006 10:20:22.867403 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ca8ae4b1-bb70-4990-96b2-65be0b3069f8-config\") on node \"crc\" DevicePath \"\"" Oct 06 10:20:22 crc kubenswrapper[4824]: I1006 10:20:22.867414 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-swsb9\" (UniqueName: \"kubernetes.io/projected/ca8ae4b1-bb70-4990-96b2-65be0b3069f8-kube-api-access-swsb9\") on node \"crc\" DevicePath \"\"" Oct 06 10:20:22 crc kubenswrapper[4824]: I1006 10:20:22.867425 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ca8ae4b1-bb70-4990-96b2-65be0b3069f8-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 06 10:20:22 crc kubenswrapper[4824]: I1006 10:20:22.867433 4824 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ca8ae4b1-bb70-4990-96b2-65be0b3069f8-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 06 10:20:23 crc kubenswrapper[4824]: I1006 10:20:23.497748 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67b789f86c-kbmnp" event={"ID":"ca8ae4b1-bb70-4990-96b2-65be0b3069f8","Type":"ContainerDied","Data":"9a575afbd0194d8f94f2b3945f886e0104a02a14daa48eb335b5d5a95a03b64a"} Oct 06 10:20:23 crc kubenswrapper[4824]: I1006 10:20:23.498523 4824 scope.go:117] "RemoveContainer" containerID="4783d4a5bc2c9004bae3fb840826b0abe9f2bbf6be58653959700e160720fc27" Oct 06 10:20:23 crc kubenswrapper[4824]: I1006 10:20:23.498127 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67b789f86c-kbmnp" Oct 06 10:20:23 crc kubenswrapper[4824]: I1006 10:20:23.538390 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-67b789f86c-kbmnp"] Oct 06 10:20:23 crc kubenswrapper[4824]: I1006 10:20:23.544800 4824 scope.go:117] "RemoveContainer" containerID="e0ebbf37409574207b093c1fa522feee73803fc6b72dd3a9dcf69678bbd02d03" Oct 06 10:20:23 crc kubenswrapper[4824]: I1006 10:20:23.549279 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-67b789f86c-kbmnp"] Oct 06 10:20:25 crc kubenswrapper[4824]: I1006 10:20:25.292161 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ca8ae4b1-bb70-4990-96b2-65be0b3069f8" path="/var/lib/kubelet/pods/ca8ae4b1-bb70-4990-96b2-65be0b3069f8/volumes" Oct 06 10:20:33 crc kubenswrapper[4824]: I1006 10:20:33.627342 4824 generic.go:334] "Generic (PLEG): container finished" podID="9b7a49d4-5482-4da9-aad2-4aabf9584534" containerID="9eadc53b991c6ca7aee566971fb7e3f0f9f8d72fa8ce6ea321deca2e69c85275" exitCode=0 Oct 06 10:20:33 crc kubenswrapper[4824]: I1006 10:20:33.627373 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"9b7a49d4-5482-4da9-aad2-4aabf9584534","Type":"ContainerDied","Data":"9eadc53b991c6ca7aee566971fb7e3f0f9f8d72fa8ce6ea321deca2e69c85275"} Oct 06 10:20:34 crc kubenswrapper[4824]: I1006 10:20:34.642424 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"9b7a49d4-5482-4da9-aad2-4aabf9584534","Type":"ContainerStarted","Data":"5ee1305e0f3fbf7b4533557bda5b509e2c6ecc89862a09c083a8874928940d83"} Oct 06 10:20:34 crc kubenswrapper[4824]: I1006 10:20:34.643770 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Oct 06 10:20:34 crc kubenswrapper[4824]: I1006 10:20:34.644554 4824 generic.go:334] "Generic (PLEG): container finished" podID="0295abaf-72dd-46d7-9125-543921a6ef61" containerID="e1d8499b9c4e9af73a75992df9e12c0941c1611bf57972474f0c399b39b189d2" exitCode=0 Oct 06 10:20:34 crc kubenswrapper[4824]: I1006 10:20:34.644595 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"0295abaf-72dd-46d7-9125-543921a6ef61","Type":"ContainerDied","Data":"e1d8499b9c4e9af73a75992df9e12c0941c1611bf57972474f0c399b39b189d2"} Oct 06 10:20:34 crc kubenswrapper[4824]: I1006 10:20:34.675409 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=36.675391944 podStartE2EDuration="36.675391944s" podCreationTimestamp="2025-10-06 10:19:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:20:34.673406013 +0000 UTC m=+1344.037828874" watchObservedRunningTime="2025-10-06 10:20:34.675391944 +0000 UTC m=+1344.039814805" Oct 06 10:20:35 crc kubenswrapper[4824]: I1006 10:20:35.226136 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9kpm9"] Oct 06 10:20:35 crc kubenswrapper[4824]: E1006 10:20:35.226923 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca8ae4b1-bb70-4990-96b2-65be0b3069f8" containerName="init" Oct 06 10:20:35 crc kubenswrapper[4824]: I1006 10:20:35.226941 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca8ae4b1-bb70-4990-96b2-65be0b3069f8" containerName="init" Oct 06 10:20:35 crc kubenswrapper[4824]: E1006 10:20:35.226952 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b940b381-928a-4f5d-b91e-0da628631f7c" containerName="init" Oct 06 10:20:35 crc kubenswrapper[4824]: I1006 10:20:35.226961 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="b940b381-928a-4f5d-b91e-0da628631f7c" containerName="init" Oct 06 10:20:35 crc kubenswrapper[4824]: E1006 10:20:35.227005 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b940b381-928a-4f5d-b91e-0da628631f7c" containerName="dnsmasq-dns" Oct 06 10:20:35 crc kubenswrapper[4824]: I1006 10:20:35.227013 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="b940b381-928a-4f5d-b91e-0da628631f7c" containerName="dnsmasq-dns" Oct 06 10:20:35 crc kubenswrapper[4824]: E1006 10:20:35.227048 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca8ae4b1-bb70-4990-96b2-65be0b3069f8" containerName="dnsmasq-dns" Oct 06 10:20:35 crc kubenswrapper[4824]: I1006 10:20:35.227054 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca8ae4b1-bb70-4990-96b2-65be0b3069f8" containerName="dnsmasq-dns" Oct 06 10:20:35 crc kubenswrapper[4824]: I1006 10:20:35.227266 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="b940b381-928a-4f5d-b91e-0da628631f7c" containerName="dnsmasq-dns" Oct 06 10:20:35 crc kubenswrapper[4824]: I1006 10:20:35.227283 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="ca8ae4b1-bb70-4990-96b2-65be0b3069f8" containerName="dnsmasq-dns" Oct 06 10:20:35 crc kubenswrapper[4824]: I1006 10:20:35.228163 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9kpm9" Oct 06 10:20:35 crc kubenswrapper[4824]: I1006 10:20:35.238963 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 06 10:20:35 crc kubenswrapper[4824]: I1006 10:20:35.238963 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 06 10:20:35 crc kubenswrapper[4824]: I1006 10:20:35.239140 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 06 10:20:35 crc kubenswrapper[4824]: I1006 10:20:35.243361 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9kpm9"] Oct 06 10:20:35 crc kubenswrapper[4824]: I1006 10:20:35.249798 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-f886w" Oct 06 10:20:35 crc kubenswrapper[4824]: I1006 10:20:35.290734 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e90bc2c-77ea-40b7-87b1-82cd30dbb082-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-9kpm9\" (UID: \"4e90bc2c-77ea-40b7-87b1-82cd30dbb082\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9kpm9" Oct 06 10:20:35 crc kubenswrapper[4824]: I1006 10:20:35.290792 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-swkk8\" (UniqueName: \"kubernetes.io/projected/4e90bc2c-77ea-40b7-87b1-82cd30dbb082-kube-api-access-swkk8\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-9kpm9\" (UID: \"4e90bc2c-77ea-40b7-87b1-82cd30dbb082\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9kpm9" Oct 06 10:20:35 crc kubenswrapper[4824]: I1006 10:20:35.290853 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4e90bc2c-77ea-40b7-87b1-82cd30dbb082-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-9kpm9\" (UID: \"4e90bc2c-77ea-40b7-87b1-82cd30dbb082\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9kpm9" Oct 06 10:20:35 crc kubenswrapper[4824]: I1006 10:20:35.290884 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4e90bc2c-77ea-40b7-87b1-82cd30dbb082-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-9kpm9\" (UID: \"4e90bc2c-77ea-40b7-87b1-82cd30dbb082\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9kpm9" Oct 06 10:20:35 crc kubenswrapper[4824]: I1006 10:20:35.392636 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4e90bc2c-77ea-40b7-87b1-82cd30dbb082-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-9kpm9\" (UID: \"4e90bc2c-77ea-40b7-87b1-82cd30dbb082\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9kpm9" Oct 06 10:20:35 crc kubenswrapper[4824]: I1006 10:20:35.393887 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e90bc2c-77ea-40b7-87b1-82cd30dbb082-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-9kpm9\" (UID: \"4e90bc2c-77ea-40b7-87b1-82cd30dbb082\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9kpm9" Oct 06 10:20:35 crc kubenswrapper[4824]: I1006 10:20:35.393960 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-swkk8\" (UniqueName: \"kubernetes.io/projected/4e90bc2c-77ea-40b7-87b1-82cd30dbb082-kube-api-access-swkk8\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-9kpm9\" (UID: \"4e90bc2c-77ea-40b7-87b1-82cd30dbb082\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9kpm9" Oct 06 10:20:35 crc kubenswrapper[4824]: I1006 10:20:35.394725 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4e90bc2c-77ea-40b7-87b1-82cd30dbb082-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-9kpm9\" (UID: \"4e90bc2c-77ea-40b7-87b1-82cd30dbb082\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9kpm9" Oct 06 10:20:35 crc kubenswrapper[4824]: I1006 10:20:35.399936 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4e90bc2c-77ea-40b7-87b1-82cd30dbb082-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-9kpm9\" (UID: \"4e90bc2c-77ea-40b7-87b1-82cd30dbb082\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9kpm9" Oct 06 10:20:35 crc kubenswrapper[4824]: I1006 10:20:35.400167 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e90bc2c-77ea-40b7-87b1-82cd30dbb082-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-9kpm9\" (UID: \"4e90bc2c-77ea-40b7-87b1-82cd30dbb082\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9kpm9" Oct 06 10:20:35 crc kubenswrapper[4824]: I1006 10:20:35.401161 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4e90bc2c-77ea-40b7-87b1-82cd30dbb082-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-9kpm9\" (UID: \"4e90bc2c-77ea-40b7-87b1-82cd30dbb082\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9kpm9" Oct 06 10:20:35 crc kubenswrapper[4824]: I1006 10:20:35.417605 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-swkk8\" (UniqueName: \"kubernetes.io/projected/4e90bc2c-77ea-40b7-87b1-82cd30dbb082-kube-api-access-swkk8\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-9kpm9\" (UID: \"4e90bc2c-77ea-40b7-87b1-82cd30dbb082\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9kpm9" Oct 06 10:20:35 crc kubenswrapper[4824]: I1006 10:20:35.548686 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9kpm9" Oct 06 10:20:35 crc kubenswrapper[4824]: I1006 10:20:35.673516 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"0295abaf-72dd-46d7-9125-543921a6ef61","Type":"ContainerStarted","Data":"eeac78b02e4aa0014c22f8cf3609c18929ab388bf1f5c358487b514fb0e95619"} Oct 06 10:20:35 crc kubenswrapper[4824]: I1006 10:20:35.674099 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Oct 06 10:20:35 crc kubenswrapper[4824]: I1006 10:20:35.706371 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=37.706349473 podStartE2EDuration="37.706349473s" podCreationTimestamp="2025-10-06 10:19:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:20:35.704195989 +0000 UTC m=+1345.068618850" watchObservedRunningTime="2025-10-06 10:20:35.706349473 +0000 UTC m=+1345.070772334" Oct 06 10:20:35 crc kubenswrapper[4824]: I1006 10:20:35.983432 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9kpm9"] Oct 06 10:20:36 crc kubenswrapper[4824]: I1006 10:20:36.689324 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9kpm9" event={"ID":"4e90bc2c-77ea-40b7-87b1-82cd30dbb082","Type":"ContainerStarted","Data":"ed455dfdfd2261a8ca12f9089336c4c4c1b59381bbe7bd48266586873ad3acac"} Oct 06 10:20:45 crc kubenswrapper[4824]: I1006 10:20:45.802372 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9kpm9" event={"ID":"4e90bc2c-77ea-40b7-87b1-82cd30dbb082","Type":"ContainerStarted","Data":"47945cf3ab7249aeac1009479285e33f46ab45f55379ce89542e943b226f0afe"} Oct 06 10:20:45 crc kubenswrapper[4824]: I1006 10:20:45.831650 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9kpm9" podStartSLOduration=1.669914172 podStartE2EDuration="10.831624641s" podCreationTimestamp="2025-10-06 10:20:35 +0000 UTC" firstStartedPulling="2025-10-06 10:20:35.992513581 +0000 UTC m=+1345.356936442" lastFinishedPulling="2025-10-06 10:20:45.15422406 +0000 UTC m=+1354.518646911" observedRunningTime="2025-10-06 10:20:45.824380379 +0000 UTC m=+1355.188803250" watchObservedRunningTime="2025-10-06 10:20:45.831624641 +0000 UTC m=+1355.196047502" Oct 06 10:20:48 crc kubenswrapper[4824]: I1006 10:20:48.744370 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Oct 06 10:20:48 crc kubenswrapper[4824]: I1006 10:20:48.901230 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Oct 06 10:20:57 crc kubenswrapper[4824]: I1006 10:20:57.953043 4824 generic.go:334] "Generic (PLEG): container finished" podID="4e90bc2c-77ea-40b7-87b1-82cd30dbb082" containerID="47945cf3ab7249aeac1009479285e33f46ab45f55379ce89542e943b226f0afe" exitCode=0 Oct 06 10:20:57 crc kubenswrapper[4824]: I1006 10:20:57.953134 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9kpm9" event={"ID":"4e90bc2c-77ea-40b7-87b1-82cd30dbb082","Type":"ContainerDied","Data":"47945cf3ab7249aeac1009479285e33f46ab45f55379ce89542e943b226f0afe"} Oct 06 10:20:59 crc kubenswrapper[4824]: I1006 10:20:59.440835 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9kpm9" Oct 06 10:20:59 crc kubenswrapper[4824]: I1006 10:20:59.568200 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-swkk8\" (UniqueName: \"kubernetes.io/projected/4e90bc2c-77ea-40b7-87b1-82cd30dbb082-kube-api-access-swkk8\") pod \"4e90bc2c-77ea-40b7-87b1-82cd30dbb082\" (UID: \"4e90bc2c-77ea-40b7-87b1-82cd30dbb082\") " Oct 06 10:20:59 crc kubenswrapper[4824]: I1006 10:20:59.568471 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4e90bc2c-77ea-40b7-87b1-82cd30dbb082-ssh-key\") pod \"4e90bc2c-77ea-40b7-87b1-82cd30dbb082\" (UID: \"4e90bc2c-77ea-40b7-87b1-82cd30dbb082\") " Oct 06 10:20:59 crc kubenswrapper[4824]: I1006 10:20:59.568555 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e90bc2c-77ea-40b7-87b1-82cd30dbb082-repo-setup-combined-ca-bundle\") pod \"4e90bc2c-77ea-40b7-87b1-82cd30dbb082\" (UID: \"4e90bc2c-77ea-40b7-87b1-82cd30dbb082\") " Oct 06 10:20:59 crc kubenswrapper[4824]: I1006 10:20:59.568619 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4e90bc2c-77ea-40b7-87b1-82cd30dbb082-inventory\") pod \"4e90bc2c-77ea-40b7-87b1-82cd30dbb082\" (UID: \"4e90bc2c-77ea-40b7-87b1-82cd30dbb082\") " Oct 06 10:20:59 crc kubenswrapper[4824]: I1006 10:20:59.588184 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4e90bc2c-77ea-40b7-87b1-82cd30dbb082-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "4e90bc2c-77ea-40b7-87b1-82cd30dbb082" (UID: "4e90bc2c-77ea-40b7-87b1-82cd30dbb082"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:20:59 crc kubenswrapper[4824]: I1006 10:20:59.598658 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4e90bc2c-77ea-40b7-87b1-82cd30dbb082-kube-api-access-swkk8" (OuterVolumeSpecName: "kube-api-access-swkk8") pod "4e90bc2c-77ea-40b7-87b1-82cd30dbb082" (UID: "4e90bc2c-77ea-40b7-87b1-82cd30dbb082"). InnerVolumeSpecName "kube-api-access-swkk8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:20:59 crc kubenswrapper[4824]: I1006 10:20:59.639898 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4e90bc2c-77ea-40b7-87b1-82cd30dbb082-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "4e90bc2c-77ea-40b7-87b1-82cd30dbb082" (UID: "4e90bc2c-77ea-40b7-87b1-82cd30dbb082"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:20:59 crc kubenswrapper[4824]: I1006 10:20:59.647372 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4e90bc2c-77ea-40b7-87b1-82cd30dbb082-inventory" (OuterVolumeSpecName: "inventory") pod "4e90bc2c-77ea-40b7-87b1-82cd30dbb082" (UID: "4e90bc2c-77ea-40b7-87b1-82cd30dbb082"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:20:59 crc kubenswrapper[4824]: I1006 10:20:59.671625 4824 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4e90bc2c-77ea-40b7-87b1-82cd30dbb082-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 06 10:20:59 crc kubenswrapper[4824]: I1006 10:20:59.671665 4824 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e90bc2c-77ea-40b7-87b1-82cd30dbb082-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 10:20:59 crc kubenswrapper[4824]: I1006 10:20:59.671680 4824 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4e90bc2c-77ea-40b7-87b1-82cd30dbb082-inventory\") on node \"crc\" DevicePath \"\"" Oct 06 10:20:59 crc kubenswrapper[4824]: I1006 10:20:59.671691 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-swkk8\" (UniqueName: \"kubernetes.io/projected/4e90bc2c-77ea-40b7-87b1-82cd30dbb082-kube-api-access-swkk8\") on node \"crc\" DevicePath \"\"" Oct 06 10:20:59 crc kubenswrapper[4824]: I1006 10:20:59.978295 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9kpm9" event={"ID":"4e90bc2c-77ea-40b7-87b1-82cd30dbb082","Type":"ContainerDied","Data":"ed455dfdfd2261a8ca12f9089336c4c4c1b59381bbe7bd48266586873ad3acac"} Oct 06 10:20:59 crc kubenswrapper[4824]: I1006 10:20:59.978683 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ed455dfdfd2261a8ca12f9089336c4c4c1b59381bbe7bd48266586873ad3acac" Oct 06 10:20:59 crc kubenswrapper[4824]: I1006 10:20:59.978358 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9kpm9" Oct 06 10:21:00 crc kubenswrapper[4824]: I1006 10:21:00.059826 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-pqwkw"] Oct 06 10:21:00 crc kubenswrapper[4824]: E1006 10:21:00.060433 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e90bc2c-77ea-40b7-87b1-82cd30dbb082" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Oct 06 10:21:00 crc kubenswrapper[4824]: I1006 10:21:00.060463 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e90bc2c-77ea-40b7-87b1-82cd30dbb082" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Oct 06 10:21:00 crc kubenswrapper[4824]: I1006 10:21:00.060706 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="4e90bc2c-77ea-40b7-87b1-82cd30dbb082" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Oct 06 10:21:00 crc kubenswrapper[4824]: I1006 10:21:00.061592 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-pqwkw" Oct 06 10:21:00 crc kubenswrapper[4824]: I1006 10:21:00.068035 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 06 10:21:00 crc kubenswrapper[4824]: I1006 10:21:00.068281 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 06 10:21:00 crc kubenswrapper[4824]: I1006 10:21:00.068445 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-f886w" Oct 06 10:21:00 crc kubenswrapper[4824]: I1006 10:21:00.068516 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 06 10:21:00 crc kubenswrapper[4824]: I1006 10:21:00.068561 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-pqwkw"] Oct 06 10:21:00 crc kubenswrapper[4824]: I1006 10:21:00.181126 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a043cdf1-f56a-415d-8914-dc689aca5a9f-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-pqwkw\" (UID: \"a043cdf1-f56a-415d-8914-dc689aca5a9f\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-pqwkw" Oct 06 10:21:00 crc kubenswrapper[4824]: I1006 10:21:00.181364 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xp4wh\" (UniqueName: \"kubernetes.io/projected/a043cdf1-f56a-415d-8914-dc689aca5a9f-kube-api-access-xp4wh\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-pqwkw\" (UID: \"a043cdf1-f56a-415d-8914-dc689aca5a9f\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-pqwkw" Oct 06 10:21:00 crc kubenswrapper[4824]: I1006 10:21:00.181812 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a043cdf1-f56a-415d-8914-dc689aca5a9f-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-pqwkw\" (UID: \"a043cdf1-f56a-415d-8914-dc689aca5a9f\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-pqwkw" Oct 06 10:21:00 crc kubenswrapper[4824]: I1006 10:21:00.283298 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a043cdf1-f56a-415d-8914-dc689aca5a9f-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-pqwkw\" (UID: \"a043cdf1-f56a-415d-8914-dc689aca5a9f\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-pqwkw" Oct 06 10:21:00 crc kubenswrapper[4824]: I1006 10:21:00.283377 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a043cdf1-f56a-415d-8914-dc689aca5a9f-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-pqwkw\" (UID: \"a043cdf1-f56a-415d-8914-dc689aca5a9f\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-pqwkw" Oct 06 10:21:00 crc kubenswrapper[4824]: I1006 10:21:00.283459 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xp4wh\" (UniqueName: \"kubernetes.io/projected/a043cdf1-f56a-415d-8914-dc689aca5a9f-kube-api-access-xp4wh\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-pqwkw\" (UID: \"a043cdf1-f56a-415d-8914-dc689aca5a9f\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-pqwkw" Oct 06 10:21:00 crc kubenswrapper[4824]: I1006 10:21:00.287607 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a043cdf1-f56a-415d-8914-dc689aca5a9f-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-pqwkw\" (UID: \"a043cdf1-f56a-415d-8914-dc689aca5a9f\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-pqwkw" Oct 06 10:21:00 crc kubenswrapper[4824]: I1006 10:21:00.287905 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a043cdf1-f56a-415d-8914-dc689aca5a9f-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-pqwkw\" (UID: \"a043cdf1-f56a-415d-8914-dc689aca5a9f\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-pqwkw" Oct 06 10:21:00 crc kubenswrapper[4824]: I1006 10:21:00.301842 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xp4wh\" (UniqueName: \"kubernetes.io/projected/a043cdf1-f56a-415d-8914-dc689aca5a9f-kube-api-access-xp4wh\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-pqwkw\" (UID: \"a043cdf1-f56a-415d-8914-dc689aca5a9f\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-pqwkw" Oct 06 10:21:00 crc kubenswrapper[4824]: I1006 10:21:00.392558 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-pqwkw" Oct 06 10:21:00 crc kubenswrapper[4824]: I1006 10:21:00.934714 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-pqwkw"] Oct 06 10:21:00 crc kubenswrapper[4824]: I1006 10:21:00.988286 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-pqwkw" event={"ID":"a043cdf1-f56a-415d-8914-dc689aca5a9f","Type":"ContainerStarted","Data":"146f9e312d000c72f907b0aa318ebb0ad4506e208520f15f091cd7820a46a9c5"} Oct 06 10:21:03 crc kubenswrapper[4824]: I1006 10:21:03.006802 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-pqwkw" event={"ID":"a043cdf1-f56a-415d-8914-dc689aca5a9f","Type":"ContainerStarted","Data":"ab183c56459c7405442362e0d24d322cfc66d99b0980ab3555156eea552c343d"} Oct 06 10:21:03 crc kubenswrapper[4824]: I1006 10:21:03.035761 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-pqwkw" podStartSLOduration=2.102381749 podStartE2EDuration="3.035740054s" podCreationTimestamp="2025-10-06 10:21:00 +0000 UTC" firstStartedPulling="2025-10-06 10:21:00.936456546 +0000 UTC m=+1370.300879407" lastFinishedPulling="2025-10-06 10:21:01.869814851 +0000 UTC m=+1371.234237712" observedRunningTime="2025-10-06 10:21:03.020799116 +0000 UTC m=+1372.385221987" watchObservedRunningTime="2025-10-06 10:21:03.035740054 +0000 UTC m=+1372.400162915" Oct 06 10:21:06 crc kubenswrapper[4824]: I1006 10:21:06.041175 4824 generic.go:334] "Generic (PLEG): container finished" podID="a043cdf1-f56a-415d-8914-dc689aca5a9f" containerID="ab183c56459c7405442362e0d24d322cfc66d99b0980ab3555156eea552c343d" exitCode=0 Oct 06 10:21:06 crc kubenswrapper[4824]: I1006 10:21:06.041769 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-pqwkw" event={"ID":"a043cdf1-f56a-415d-8914-dc689aca5a9f","Type":"ContainerDied","Data":"ab183c56459c7405442362e0d24d322cfc66d99b0980ab3555156eea552c343d"} Oct 06 10:21:07 crc kubenswrapper[4824]: I1006 10:21:07.631575 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-pqwkw" Oct 06 10:21:07 crc kubenswrapper[4824]: I1006 10:21:07.802655 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a043cdf1-f56a-415d-8914-dc689aca5a9f-ssh-key\") pod \"a043cdf1-f56a-415d-8914-dc689aca5a9f\" (UID: \"a043cdf1-f56a-415d-8914-dc689aca5a9f\") " Oct 06 10:21:07 crc kubenswrapper[4824]: I1006 10:21:07.802745 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xp4wh\" (UniqueName: \"kubernetes.io/projected/a043cdf1-f56a-415d-8914-dc689aca5a9f-kube-api-access-xp4wh\") pod \"a043cdf1-f56a-415d-8914-dc689aca5a9f\" (UID: \"a043cdf1-f56a-415d-8914-dc689aca5a9f\") " Oct 06 10:21:07 crc kubenswrapper[4824]: I1006 10:21:07.802797 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a043cdf1-f56a-415d-8914-dc689aca5a9f-inventory\") pod \"a043cdf1-f56a-415d-8914-dc689aca5a9f\" (UID: \"a043cdf1-f56a-415d-8914-dc689aca5a9f\") " Oct 06 10:21:07 crc kubenswrapper[4824]: I1006 10:21:07.809354 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a043cdf1-f56a-415d-8914-dc689aca5a9f-kube-api-access-xp4wh" (OuterVolumeSpecName: "kube-api-access-xp4wh") pod "a043cdf1-f56a-415d-8914-dc689aca5a9f" (UID: "a043cdf1-f56a-415d-8914-dc689aca5a9f"). InnerVolumeSpecName "kube-api-access-xp4wh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:21:07 crc kubenswrapper[4824]: I1006 10:21:07.834479 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a043cdf1-f56a-415d-8914-dc689aca5a9f-inventory" (OuterVolumeSpecName: "inventory") pod "a043cdf1-f56a-415d-8914-dc689aca5a9f" (UID: "a043cdf1-f56a-415d-8914-dc689aca5a9f"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:21:07 crc kubenswrapper[4824]: I1006 10:21:07.838023 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a043cdf1-f56a-415d-8914-dc689aca5a9f-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "a043cdf1-f56a-415d-8914-dc689aca5a9f" (UID: "a043cdf1-f56a-415d-8914-dc689aca5a9f"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:21:07 crc kubenswrapper[4824]: I1006 10:21:07.905687 4824 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a043cdf1-f56a-415d-8914-dc689aca5a9f-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 06 10:21:07 crc kubenswrapper[4824]: I1006 10:21:07.905727 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xp4wh\" (UniqueName: \"kubernetes.io/projected/a043cdf1-f56a-415d-8914-dc689aca5a9f-kube-api-access-xp4wh\") on node \"crc\" DevicePath \"\"" Oct 06 10:21:07 crc kubenswrapper[4824]: I1006 10:21:07.905742 4824 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a043cdf1-f56a-415d-8914-dc689aca5a9f-inventory\") on node \"crc\" DevicePath \"\"" Oct 06 10:21:08 crc kubenswrapper[4824]: I1006 10:21:08.061510 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-pqwkw" event={"ID":"a043cdf1-f56a-415d-8914-dc689aca5a9f","Type":"ContainerDied","Data":"146f9e312d000c72f907b0aa318ebb0ad4506e208520f15f091cd7820a46a9c5"} Oct 06 10:21:08 crc kubenswrapper[4824]: I1006 10:21:08.061865 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="146f9e312d000c72f907b0aa318ebb0ad4506e208520f15f091cd7820a46a9c5" Oct 06 10:21:08 crc kubenswrapper[4824]: I1006 10:21:08.061656 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-pqwkw" Oct 06 10:21:08 crc kubenswrapper[4824]: I1006 10:21:08.202605 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-mlckl"] Oct 06 10:21:08 crc kubenswrapper[4824]: E1006 10:21:08.207014 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a043cdf1-f56a-415d-8914-dc689aca5a9f" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Oct 06 10:21:08 crc kubenswrapper[4824]: I1006 10:21:08.207175 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="a043cdf1-f56a-415d-8914-dc689aca5a9f" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Oct 06 10:21:08 crc kubenswrapper[4824]: I1006 10:21:08.208487 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="a043cdf1-f56a-415d-8914-dc689aca5a9f" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Oct 06 10:21:08 crc kubenswrapper[4824]: I1006 10:21:08.210264 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-mlckl" Oct 06 10:21:08 crc kubenswrapper[4824]: I1006 10:21:08.214997 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 06 10:21:08 crc kubenswrapper[4824]: I1006 10:21:08.217413 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-f886w" Oct 06 10:21:08 crc kubenswrapper[4824]: I1006 10:21:08.218067 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-mlckl"] Oct 06 10:21:08 crc kubenswrapper[4824]: I1006 10:21:08.218281 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 06 10:21:08 crc kubenswrapper[4824]: I1006 10:21:08.221567 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c995dce1-6fa5-456b-b984-b397dcc9f9dc-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-mlckl\" (UID: \"c995dce1-6fa5-456b-b984-b397dcc9f9dc\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-mlckl" Oct 06 10:21:08 crc kubenswrapper[4824]: I1006 10:21:08.221914 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5x8hs\" (UniqueName: \"kubernetes.io/projected/c995dce1-6fa5-456b-b984-b397dcc9f9dc-kube-api-access-5x8hs\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-mlckl\" (UID: \"c995dce1-6fa5-456b-b984-b397dcc9f9dc\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-mlckl" Oct 06 10:21:08 crc kubenswrapper[4824]: I1006 10:21:08.221963 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c995dce1-6fa5-456b-b984-b397dcc9f9dc-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-mlckl\" (UID: \"c995dce1-6fa5-456b-b984-b397dcc9f9dc\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-mlckl" Oct 06 10:21:08 crc kubenswrapper[4824]: I1006 10:21:08.222364 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c995dce1-6fa5-456b-b984-b397dcc9f9dc-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-mlckl\" (UID: \"c995dce1-6fa5-456b-b984-b397dcc9f9dc\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-mlckl" Oct 06 10:21:08 crc kubenswrapper[4824]: I1006 10:21:08.223581 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 06 10:21:08 crc kubenswrapper[4824]: I1006 10:21:08.323183 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c995dce1-6fa5-456b-b984-b397dcc9f9dc-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-mlckl\" (UID: \"c995dce1-6fa5-456b-b984-b397dcc9f9dc\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-mlckl" Oct 06 10:21:08 crc kubenswrapper[4824]: I1006 10:21:08.323280 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c995dce1-6fa5-456b-b984-b397dcc9f9dc-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-mlckl\" (UID: \"c995dce1-6fa5-456b-b984-b397dcc9f9dc\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-mlckl" Oct 06 10:21:08 crc kubenswrapper[4824]: I1006 10:21:08.323404 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5x8hs\" (UniqueName: \"kubernetes.io/projected/c995dce1-6fa5-456b-b984-b397dcc9f9dc-kube-api-access-5x8hs\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-mlckl\" (UID: \"c995dce1-6fa5-456b-b984-b397dcc9f9dc\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-mlckl" Oct 06 10:21:08 crc kubenswrapper[4824]: I1006 10:21:08.323434 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c995dce1-6fa5-456b-b984-b397dcc9f9dc-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-mlckl\" (UID: \"c995dce1-6fa5-456b-b984-b397dcc9f9dc\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-mlckl" Oct 06 10:21:08 crc kubenswrapper[4824]: I1006 10:21:08.329256 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c995dce1-6fa5-456b-b984-b397dcc9f9dc-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-mlckl\" (UID: \"c995dce1-6fa5-456b-b984-b397dcc9f9dc\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-mlckl" Oct 06 10:21:08 crc kubenswrapper[4824]: I1006 10:21:08.330809 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c995dce1-6fa5-456b-b984-b397dcc9f9dc-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-mlckl\" (UID: \"c995dce1-6fa5-456b-b984-b397dcc9f9dc\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-mlckl" Oct 06 10:21:08 crc kubenswrapper[4824]: I1006 10:21:08.343605 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c995dce1-6fa5-456b-b984-b397dcc9f9dc-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-mlckl\" (UID: \"c995dce1-6fa5-456b-b984-b397dcc9f9dc\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-mlckl" Oct 06 10:21:08 crc kubenswrapper[4824]: I1006 10:21:08.344459 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5x8hs\" (UniqueName: \"kubernetes.io/projected/c995dce1-6fa5-456b-b984-b397dcc9f9dc-kube-api-access-5x8hs\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-mlckl\" (UID: \"c995dce1-6fa5-456b-b984-b397dcc9f9dc\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-mlckl" Oct 06 10:21:08 crc kubenswrapper[4824]: I1006 10:21:08.533953 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-mlckl" Oct 06 10:21:09 crc kubenswrapper[4824]: I1006 10:21:09.253584 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-mlckl"] Oct 06 10:21:10 crc kubenswrapper[4824]: I1006 10:21:10.088101 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-mlckl" event={"ID":"c995dce1-6fa5-456b-b984-b397dcc9f9dc","Type":"ContainerStarted","Data":"8b605b390d04aeadf5dda98fad576bb262657d9a1a7c9087513b0de8ca2b18a5"} Oct 06 10:21:11 crc kubenswrapper[4824]: I1006 10:21:11.124185 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-mlckl" event={"ID":"c995dce1-6fa5-456b-b984-b397dcc9f9dc","Type":"ContainerStarted","Data":"550b5638ac37f01b97333520fd691b2cef12d4f0439856452d41c374c9d08517"} Oct 06 10:21:11 crc kubenswrapper[4824]: I1006 10:21:11.151774 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-mlckl" podStartSLOduration=2.595840849 podStartE2EDuration="3.151753081s" podCreationTimestamp="2025-10-06 10:21:08 +0000 UTC" firstStartedPulling="2025-10-06 10:21:09.264386707 +0000 UTC m=+1378.628809568" lastFinishedPulling="2025-10-06 10:21:09.820298949 +0000 UTC m=+1379.184721800" observedRunningTime="2025-10-06 10:21:11.144157379 +0000 UTC m=+1380.508580260" watchObservedRunningTime="2025-10-06 10:21:11.151753081 +0000 UTC m=+1380.516175962" Oct 06 10:21:30 crc kubenswrapper[4824]: I1006 10:21:30.307695 4824 scope.go:117] "RemoveContainer" containerID="aff700ad3efac962ccd117169b4ca28a7581f213e1f027952ab79d16a6167ae0" Oct 06 10:21:30 crc kubenswrapper[4824]: I1006 10:21:30.337115 4824 scope.go:117] "RemoveContainer" containerID="e17f227e1abc0d9b1729053104cdd67ed12db77ffc53e5658e7c5da09ebc3383" Oct 06 10:21:30 crc kubenswrapper[4824]: I1006 10:21:30.387702 4824 scope.go:117] "RemoveContainer" containerID="c57b55dd4e9a6b22fc4938cdb0ab8c79b1aec852bc7fe714759a448e5e41a598" Oct 06 10:21:43 crc kubenswrapper[4824]: I1006 10:21:43.915998 4824 patch_prober.go:28] interesting pod/machine-config-daemon-khgzw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 10:21:43 crc kubenswrapper[4824]: I1006 10:21:43.916828 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 10:21:56 crc kubenswrapper[4824]: I1006 10:21:56.175945 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-mqtfm"] Oct 06 10:21:56 crc kubenswrapper[4824]: I1006 10:21:56.179373 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mqtfm" Oct 06 10:21:56 crc kubenswrapper[4824]: I1006 10:21:56.193327 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-mqtfm"] Oct 06 10:21:56 crc kubenswrapper[4824]: I1006 10:21:56.351263 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8dc08944-150b-42f6-afb9-161300ac8f96-catalog-content\") pod \"redhat-marketplace-mqtfm\" (UID: \"8dc08944-150b-42f6-afb9-161300ac8f96\") " pod="openshift-marketplace/redhat-marketplace-mqtfm" Oct 06 10:21:56 crc kubenswrapper[4824]: I1006 10:21:56.351324 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gf5gs\" (UniqueName: \"kubernetes.io/projected/8dc08944-150b-42f6-afb9-161300ac8f96-kube-api-access-gf5gs\") pod \"redhat-marketplace-mqtfm\" (UID: \"8dc08944-150b-42f6-afb9-161300ac8f96\") " pod="openshift-marketplace/redhat-marketplace-mqtfm" Oct 06 10:21:56 crc kubenswrapper[4824]: I1006 10:21:56.351437 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8dc08944-150b-42f6-afb9-161300ac8f96-utilities\") pod \"redhat-marketplace-mqtfm\" (UID: \"8dc08944-150b-42f6-afb9-161300ac8f96\") " pod="openshift-marketplace/redhat-marketplace-mqtfm" Oct 06 10:21:56 crc kubenswrapper[4824]: I1006 10:21:56.453782 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8dc08944-150b-42f6-afb9-161300ac8f96-utilities\") pod \"redhat-marketplace-mqtfm\" (UID: \"8dc08944-150b-42f6-afb9-161300ac8f96\") " pod="openshift-marketplace/redhat-marketplace-mqtfm" Oct 06 10:21:56 crc kubenswrapper[4824]: I1006 10:21:56.454067 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8dc08944-150b-42f6-afb9-161300ac8f96-catalog-content\") pod \"redhat-marketplace-mqtfm\" (UID: \"8dc08944-150b-42f6-afb9-161300ac8f96\") " pod="openshift-marketplace/redhat-marketplace-mqtfm" Oct 06 10:21:56 crc kubenswrapper[4824]: I1006 10:21:56.454140 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gf5gs\" (UniqueName: \"kubernetes.io/projected/8dc08944-150b-42f6-afb9-161300ac8f96-kube-api-access-gf5gs\") pod \"redhat-marketplace-mqtfm\" (UID: \"8dc08944-150b-42f6-afb9-161300ac8f96\") " pod="openshift-marketplace/redhat-marketplace-mqtfm" Oct 06 10:21:56 crc kubenswrapper[4824]: I1006 10:21:56.454398 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8dc08944-150b-42f6-afb9-161300ac8f96-utilities\") pod \"redhat-marketplace-mqtfm\" (UID: \"8dc08944-150b-42f6-afb9-161300ac8f96\") " pod="openshift-marketplace/redhat-marketplace-mqtfm" Oct 06 10:21:56 crc kubenswrapper[4824]: I1006 10:21:56.454742 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8dc08944-150b-42f6-afb9-161300ac8f96-catalog-content\") pod \"redhat-marketplace-mqtfm\" (UID: \"8dc08944-150b-42f6-afb9-161300ac8f96\") " pod="openshift-marketplace/redhat-marketplace-mqtfm" Oct 06 10:21:56 crc kubenswrapper[4824]: I1006 10:21:56.477127 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gf5gs\" (UniqueName: \"kubernetes.io/projected/8dc08944-150b-42f6-afb9-161300ac8f96-kube-api-access-gf5gs\") pod \"redhat-marketplace-mqtfm\" (UID: \"8dc08944-150b-42f6-afb9-161300ac8f96\") " pod="openshift-marketplace/redhat-marketplace-mqtfm" Oct 06 10:21:56 crc kubenswrapper[4824]: I1006 10:21:56.510842 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mqtfm" Oct 06 10:21:56 crc kubenswrapper[4824]: I1006 10:21:56.984878 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-mqtfm"] Oct 06 10:21:57 crc kubenswrapper[4824]: I1006 10:21:57.639007 4824 generic.go:334] "Generic (PLEG): container finished" podID="8dc08944-150b-42f6-afb9-161300ac8f96" containerID="7dfba2d431bafef1e99b057b25a4cf6fe1bdbe419827c978bdab1f589b8cbec8" exitCode=0 Oct 06 10:21:57 crc kubenswrapper[4824]: I1006 10:21:57.639171 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mqtfm" event={"ID":"8dc08944-150b-42f6-afb9-161300ac8f96","Type":"ContainerDied","Data":"7dfba2d431bafef1e99b057b25a4cf6fe1bdbe419827c978bdab1f589b8cbec8"} Oct 06 10:21:57 crc kubenswrapper[4824]: I1006 10:21:57.639407 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mqtfm" event={"ID":"8dc08944-150b-42f6-afb9-161300ac8f96","Type":"ContainerStarted","Data":"f4afba96c9b072c938a7ea0210f59a1204a98b7d6242364197f0bc43d9ce93fb"} Oct 06 10:21:59 crc kubenswrapper[4824]: I1006 10:21:59.539421 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-gw8b9"] Oct 06 10:21:59 crc kubenswrapper[4824]: I1006 10:21:59.542122 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gw8b9" Oct 06 10:21:59 crc kubenswrapper[4824]: I1006 10:21:59.554558 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gw8b9"] Oct 06 10:21:59 crc kubenswrapper[4824]: I1006 10:21:59.626291 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fc0ce670-992e-4cb5-a283-4835be8d85e3-utilities\") pod \"community-operators-gw8b9\" (UID: \"fc0ce670-992e-4cb5-a283-4835be8d85e3\") " pod="openshift-marketplace/community-operators-gw8b9" Oct 06 10:21:59 crc kubenswrapper[4824]: I1006 10:21:59.626660 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fc0ce670-992e-4cb5-a283-4835be8d85e3-catalog-content\") pod \"community-operators-gw8b9\" (UID: \"fc0ce670-992e-4cb5-a283-4835be8d85e3\") " pod="openshift-marketplace/community-operators-gw8b9" Oct 06 10:21:59 crc kubenswrapper[4824]: I1006 10:21:59.626797 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hlzgf\" (UniqueName: \"kubernetes.io/projected/fc0ce670-992e-4cb5-a283-4835be8d85e3-kube-api-access-hlzgf\") pod \"community-operators-gw8b9\" (UID: \"fc0ce670-992e-4cb5-a283-4835be8d85e3\") " pod="openshift-marketplace/community-operators-gw8b9" Oct 06 10:21:59 crc kubenswrapper[4824]: I1006 10:21:59.659227 4824 generic.go:334] "Generic (PLEG): container finished" podID="8dc08944-150b-42f6-afb9-161300ac8f96" containerID="b8ddc7e0f78b29f1ca08ef3ca4e18d072013cffe66648c3266da13f35bc0ec98" exitCode=0 Oct 06 10:21:59 crc kubenswrapper[4824]: I1006 10:21:59.659282 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mqtfm" event={"ID":"8dc08944-150b-42f6-afb9-161300ac8f96","Type":"ContainerDied","Data":"b8ddc7e0f78b29f1ca08ef3ca4e18d072013cffe66648c3266da13f35bc0ec98"} Oct 06 10:21:59 crc kubenswrapper[4824]: I1006 10:21:59.728507 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fc0ce670-992e-4cb5-a283-4835be8d85e3-utilities\") pod \"community-operators-gw8b9\" (UID: \"fc0ce670-992e-4cb5-a283-4835be8d85e3\") " pod="openshift-marketplace/community-operators-gw8b9" Oct 06 10:21:59 crc kubenswrapper[4824]: I1006 10:21:59.728765 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fc0ce670-992e-4cb5-a283-4835be8d85e3-catalog-content\") pod \"community-operators-gw8b9\" (UID: \"fc0ce670-992e-4cb5-a283-4835be8d85e3\") " pod="openshift-marketplace/community-operators-gw8b9" Oct 06 10:21:59 crc kubenswrapper[4824]: I1006 10:21:59.728924 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hlzgf\" (UniqueName: \"kubernetes.io/projected/fc0ce670-992e-4cb5-a283-4835be8d85e3-kube-api-access-hlzgf\") pod \"community-operators-gw8b9\" (UID: \"fc0ce670-992e-4cb5-a283-4835be8d85e3\") " pod="openshift-marketplace/community-operators-gw8b9" Oct 06 10:21:59 crc kubenswrapper[4824]: I1006 10:21:59.729166 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fc0ce670-992e-4cb5-a283-4835be8d85e3-utilities\") pod \"community-operators-gw8b9\" (UID: \"fc0ce670-992e-4cb5-a283-4835be8d85e3\") " pod="openshift-marketplace/community-operators-gw8b9" Oct 06 10:21:59 crc kubenswrapper[4824]: I1006 10:21:59.729344 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fc0ce670-992e-4cb5-a283-4835be8d85e3-catalog-content\") pod \"community-operators-gw8b9\" (UID: \"fc0ce670-992e-4cb5-a283-4835be8d85e3\") " pod="openshift-marketplace/community-operators-gw8b9" Oct 06 10:21:59 crc kubenswrapper[4824]: I1006 10:21:59.753024 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hlzgf\" (UniqueName: \"kubernetes.io/projected/fc0ce670-992e-4cb5-a283-4835be8d85e3-kube-api-access-hlzgf\") pod \"community-operators-gw8b9\" (UID: \"fc0ce670-992e-4cb5-a283-4835be8d85e3\") " pod="openshift-marketplace/community-operators-gw8b9" Oct 06 10:21:59 crc kubenswrapper[4824]: I1006 10:21:59.866460 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gw8b9" Oct 06 10:22:00 crc kubenswrapper[4824]: I1006 10:22:00.430882 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gw8b9"] Oct 06 10:22:00 crc kubenswrapper[4824]: W1006 10:22:00.443335 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfc0ce670_992e_4cb5_a283_4835be8d85e3.slice/crio-7852484ae03772cd8e3577fd88d54c822c9f48be6abcfe811d7abc03ad85cbde WatchSource:0}: Error finding container 7852484ae03772cd8e3577fd88d54c822c9f48be6abcfe811d7abc03ad85cbde: Status 404 returned error can't find the container with id 7852484ae03772cd8e3577fd88d54c822c9f48be6abcfe811d7abc03ad85cbde Oct 06 10:22:00 crc kubenswrapper[4824]: I1006 10:22:00.671698 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mqtfm" event={"ID":"8dc08944-150b-42f6-afb9-161300ac8f96","Type":"ContainerStarted","Data":"f984dc43b1fb449079cba9ff4bdbb8b6e133cfd31f0fece81786bdc693f93e5a"} Oct 06 10:22:00 crc kubenswrapper[4824]: I1006 10:22:00.674263 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gw8b9" event={"ID":"fc0ce670-992e-4cb5-a283-4835be8d85e3","Type":"ContainerStarted","Data":"4da16e1eaad2f68741cd855e8b4df79ae0c91c8dcefa450922f8565b677e5d93"} Oct 06 10:22:00 crc kubenswrapper[4824]: I1006 10:22:00.674626 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gw8b9" event={"ID":"fc0ce670-992e-4cb5-a283-4835be8d85e3","Type":"ContainerStarted","Data":"7852484ae03772cd8e3577fd88d54c822c9f48be6abcfe811d7abc03ad85cbde"} Oct 06 10:22:00 crc kubenswrapper[4824]: I1006 10:22:00.699440 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-mqtfm" podStartSLOduration=2.157055171 podStartE2EDuration="4.699422984s" podCreationTimestamp="2025-10-06 10:21:56 +0000 UTC" firstStartedPulling="2025-10-06 10:21:57.640953446 +0000 UTC m=+1427.005376357" lastFinishedPulling="2025-10-06 10:22:00.183321309 +0000 UTC m=+1429.547744170" observedRunningTime="2025-10-06 10:22:00.692098509 +0000 UTC m=+1430.056521400" watchObservedRunningTime="2025-10-06 10:22:00.699422984 +0000 UTC m=+1430.063845845" Oct 06 10:22:01 crc kubenswrapper[4824]: I1006 10:22:01.693014 4824 generic.go:334] "Generic (PLEG): container finished" podID="fc0ce670-992e-4cb5-a283-4835be8d85e3" containerID="4da16e1eaad2f68741cd855e8b4df79ae0c91c8dcefa450922f8565b677e5d93" exitCode=0 Oct 06 10:22:01 crc kubenswrapper[4824]: I1006 10:22:01.694112 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gw8b9" event={"ID":"fc0ce670-992e-4cb5-a283-4835be8d85e3","Type":"ContainerDied","Data":"4da16e1eaad2f68741cd855e8b4df79ae0c91c8dcefa450922f8565b677e5d93"} Oct 06 10:22:03 crc kubenswrapper[4824]: I1006 10:22:03.719327 4824 generic.go:334] "Generic (PLEG): container finished" podID="fc0ce670-992e-4cb5-a283-4835be8d85e3" containerID="a987a1b1f80c06dc0e77cd4aae5dc583599e873bf656dab04854d16445a0ecf8" exitCode=0 Oct 06 10:22:03 crc kubenswrapper[4824]: I1006 10:22:03.719543 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gw8b9" event={"ID":"fc0ce670-992e-4cb5-a283-4835be8d85e3","Type":"ContainerDied","Data":"a987a1b1f80c06dc0e77cd4aae5dc583599e873bf656dab04854d16445a0ecf8"} Oct 06 10:22:04 crc kubenswrapper[4824]: I1006 10:22:04.734101 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gw8b9" event={"ID":"fc0ce670-992e-4cb5-a283-4835be8d85e3","Type":"ContainerStarted","Data":"5d5f01e73b86f8456bc4489546399c1bd7f987b6b7c9d8721cc5755dc9419d52"} Oct 06 10:22:04 crc kubenswrapper[4824]: I1006 10:22:04.759218 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-gw8b9" podStartSLOduration=3.113454177 podStartE2EDuration="5.759189348s" podCreationTimestamp="2025-10-06 10:21:59 +0000 UTC" firstStartedPulling="2025-10-06 10:22:01.696298788 +0000 UTC m=+1431.060721639" lastFinishedPulling="2025-10-06 10:22:04.342033939 +0000 UTC m=+1433.706456810" observedRunningTime="2025-10-06 10:22:04.755101095 +0000 UTC m=+1434.119523966" watchObservedRunningTime="2025-10-06 10:22:04.759189348 +0000 UTC m=+1434.123612249" Oct 06 10:22:06 crc kubenswrapper[4824]: I1006 10:22:06.511421 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-mqtfm" Oct 06 10:22:06 crc kubenswrapper[4824]: I1006 10:22:06.511810 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-mqtfm" Oct 06 10:22:06 crc kubenswrapper[4824]: I1006 10:22:06.581935 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-mqtfm" Oct 06 10:22:06 crc kubenswrapper[4824]: I1006 10:22:06.810508 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-mqtfm" Oct 06 10:22:07 crc kubenswrapper[4824]: I1006 10:22:07.729259 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-mqtfm"] Oct 06 10:22:08 crc kubenswrapper[4824]: I1006 10:22:08.777998 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-mqtfm" podUID="8dc08944-150b-42f6-afb9-161300ac8f96" containerName="registry-server" containerID="cri-o://f984dc43b1fb449079cba9ff4bdbb8b6e133cfd31f0fece81786bdc693f93e5a" gracePeriod=2 Oct 06 10:22:09 crc kubenswrapper[4824]: I1006 10:22:09.245107 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mqtfm" Oct 06 10:22:09 crc kubenswrapper[4824]: I1006 10:22:09.371183 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8dc08944-150b-42f6-afb9-161300ac8f96-catalog-content\") pod \"8dc08944-150b-42f6-afb9-161300ac8f96\" (UID: \"8dc08944-150b-42f6-afb9-161300ac8f96\") " Oct 06 10:22:09 crc kubenswrapper[4824]: I1006 10:22:09.371618 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8dc08944-150b-42f6-afb9-161300ac8f96-utilities\") pod \"8dc08944-150b-42f6-afb9-161300ac8f96\" (UID: \"8dc08944-150b-42f6-afb9-161300ac8f96\") " Oct 06 10:22:09 crc kubenswrapper[4824]: I1006 10:22:09.371655 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf5gs\" (UniqueName: \"kubernetes.io/projected/8dc08944-150b-42f6-afb9-161300ac8f96-kube-api-access-gf5gs\") pod \"8dc08944-150b-42f6-afb9-161300ac8f96\" (UID: \"8dc08944-150b-42f6-afb9-161300ac8f96\") " Oct 06 10:22:09 crc kubenswrapper[4824]: I1006 10:22:09.372713 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8dc08944-150b-42f6-afb9-161300ac8f96-utilities" (OuterVolumeSpecName: "utilities") pod "8dc08944-150b-42f6-afb9-161300ac8f96" (UID: "8dc08944-150b-42f6-afb9-161300ac8f96"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:22:09 crc kubenswrapper[4824]: I1006 10:22:09.379320 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8dc08944-150b-42f6-afb9-161300ac8f96-kube-api-access-gf5gs" (OuterVolumeSpecName: "kube-api-access-gf5gs") pod "8dc08944-150b-42f6-afb9-161300ac8f96" (UID: "8dc08944-150b-42f6-afb9-161300ac8f96"). InnerVolumeSpecName "kube-api-access-gf5gs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:22:09 crc kubenswrapper[4824]: I1006 10:22:09.386409 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8dc08944-150b-42f6-afb9-161300ac8f96-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8dc08944-150b-42f6-afb9-161300ac8f96" (UID: "8dc08944-150b-42f6-afb9-161300ac8f96"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:22:09 crc kubenswrapper[4824]: I1006 10:22:09.474479 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8dc08944-150b-42f6-afb9-161300ac8f96-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 06 10:22:09 crc kubenswrapper[4824]: I1006 10:22:09.474538 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8dc08944-150b-42f6-afb9-161300ac8f96-utilities\") on node \"crc\" DevicePath \"\"" Oct 06 10:22:09 crc kubenswrapper[4824]: I1006 10:22:09.474548 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf5gs\" (UniqueName: \"kubernetes.io/projected/8dc08944-150b-42f6-afb9-161300ac8f96-kube-api-access-gf5gs\") on node \"crc\" DevicePath \"\"" Oct 06 10:22:09 crc kubenswrapper[4824]: I1006 10:22:09.790742 4824 generic.go:334] "Generic (PLEG): container finished" podID="8dc08944-150b-42f6-afb9-161300ac8f96" containerID="f984dc43b1fb449079cba9ff4bdbb8b6e133cfd31f0fece81786bdc693f93e5a" exitCode=0 Oct 06 10:22:09 crc kubenswrapper[4824]: I1006 10:22:09.790790 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mqtfm" event={"ID":"8dc08944-150b-42f6-afb9-161300ac8f96","Type":"ContainerDied","Data":"f984dc43b1fb449079cba9ff4bdbb8b6e133cfd31f0fece81786bdc693f93e5a"} Oct 06 10:22:09 crc kubenswrapper[4824]: I1006 10:22:09.790819 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mqtfm" Oct 06 10:22:09 crc kubenswrapper[4824]: I1006 10:22:09.790840 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mqtfm" event={"ID":"8dc08944-150b-42f6-afb9-161300ac8f96","Type":"ContainerDied","Data":"f4afba96c9b072c938a7ea0210f59a1204a98b7d6242364197f0bc43d9ce93fb"} Oct 06 10:22:09 crc kubenswrapper[4824]: I1006 10:22:09.790861 4824 scope.go:117] "RemoveContainer" containerID="f984dc43b1fb449079cba9ff4bdbb8b6e133cfd31f0fece81786bdc693f93e5a" Oct 06 10:22:09 crc kubenswrapper[4824]: I1006 10:22:09.829946 4824 scope.go:117] "RemoveContainer" containerID="b8ddc7e0f78b29f1ca08ef3ca4e18d072013cffe66648c3266da13f35bc0ec98" Oct 06 10:22:09 crc kubenswrapper[4824]: I1006 10:22:09.833931 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-mqtfm"] Oct 06 10:22:09 crc kubenswrapper[4824]: I1006 10:22:09.844759 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-mqtfm"] Oct 06 10:22:09 crc kubenswrapper[4824]: I1006 10:22:09.859292 4824 scope.go:117] "RemoveContainer" containerID="7dfba2d431bafef1e99b057b25a4cf6fe1bdbe419827c978bdab1f589b8cbec8" Oct 06 10:22:09 crc kubenswrapper[4824]: I1006 10:22:09.866622 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-gw8b9" Oct 06 10:22:09 crc kubenswrapper[4824]: I1006 10:22:09.867199 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-gw8b9" Oct 06 10:22:09 crc kubenswrapper[4824]: I1006 10:22:09.904161 4824 scope.go:117] "RemoveContainer" containerID="f984dc43b1fb449079cba9ff4bdbb8b6e133cfd31f0fece81786bdc693f93e5a" Oct 06 10:22:09 crc kubenswrapper[4824]: E1006 10:22:09.904686 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f984dc43b1fb449079cba9ff4bdbb8b6e133cfd31f0fece81786bdc693f93e5a\": container with ID starting with f984dc43b1fb449079cba9ff4bdbb8b6e133cfd31f0fece81786bdc693f93e5a not found: ID does not exist" containerID="f984dc43b1fb449079cba9ff4bdbb8b6e133cfd31f0fece81786bdc693f93e5a" Oct 06 10:22:09 crc kubenswrapper[4824]: I1006 10:22:09.904728 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f984dc43b1fb449079cba9ff4bdbb8b6e133cfd31f0fece81786bdc693f93e5a"} err="failed to get container status \"f984dc43b1fb449079cba9ff4bdbb8b6e133cfd31f0fece81786bdc693f93e5a\": rpc error: code = NotFound desc = could not find container \"f984dc43b1fb449079cba9ff4bdbb8b6e133cfd31f0fece81786bdc693f93e5a\": container with ID starting with f984dc43b1fb449079cba9ff4bdbb8b6e133cfd31f0fece81786bdc693f93e5a not found: ID does not exist" Oct 06 10:22:09 crc kubenswrapper[4824]: I1006 10:22:09.904755 4824 scope.go:117] "RemoveContainer" containerID="b8ddc7e0f78b29f1ca08ef3ca4e18d072013cffe66648c3266da13f35bc0ec98" Oct 06 10:22:09 crc kubenswrapper[4824]: E1006 10:22:09.905577 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b8ddc7e0f78b29f1ca08ef3ca4e18d072013cffe66648c3266da13f35bc0ec98\": container with ID starting with b8ddc7e0f78b29f1ca08ef3ca4e18d072013cffe66648c3266da13f35bc0ec98 not found: ID does not exist" containerID="b8ddc7e0f78b29f1ca08ef3ca4e18d072013cffe66648c3266da13f35bc0ec98" Oct 06 10:22:09 crc kubenswrapper[4824]: I1006 10:22:09.905638 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b8ddc7e0f78b29f1ca08ef3ca4e18d072013cffe66648c3266da13f35bc0ec98"} err="failed to get container status \"b8ddc7e0f78b29f1ca08ef3ca4e18d072013cffe66648c3266da13f35bc0ec98\": rpc error: code = NotFound desc = could not find container \"b8ddc7e0f78b29f1ca08ef3ca4e18d072013cffe66648c3266da13f35bc0ec98\": container with ID starting with b8ddc7e0f78b29f1ca08ef3ca4e18d072013cffe66648c3266da13f35bc0ec98 not found: ID does not exist" Oct 06 10:22:09 crc kubenswrapper[4824]: I1006 10:22:09.905674 4824 scope.go:117] "RemoveContainer" containerID="7dfba2d431bafef1e99b057b25a4cf6fe1bdbe419827c978bdab1f589b8cbec8" Oct 06 10:22:09 crc kubenswrapper[4824]: E1006 10:22:09.906198 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7dfba2d431bafef1e99b057b25a4cf6fe1bdbe419827c978bdab1f589b8cbec8\": container with ID starting with 7dfba2d431bafef1e99b057b25a4cf6fe1bdbe419827c978bdab1f589b8cbec8 not found: ID does not exist" containerID="7dfba2d431bafef1e99b057b25a4cf6fe1bdbe419827c978bdab1f589b8cbec8" Oct 06 10:22:09 crc kubenswrapper[4824]: I1006 10:22:09.906257 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7dfba2d431bafef1e99b057b25a4cf6fe1bdbe419827c978bdab1f589b8cbec8"} err="failed to get container status \"7dfba2d431bafef1e99b057b25a4cf6fe1bdbe419827c978bdab1f589b8cbec8\": rpc error: code = NotFound desc = could not find container \"7dfba2d431bafef1e99b057b25a4cf6fe1bdbe419827c978bdab1f589b8cbec8\": container with ID starting with 7dfba2d431bafef1e99b057b25a4cf6fe1bdbe419827c978bdab1f589b8cbec8 not found: ID does not exist" Oct 06 10:22:09 crc kubenswrapper[4824]: I1006 10:22:09.920486 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-gw8b9" Oct 06 10:22:10 crc kubenswrapper[4824]: I1006 10:22:10.871278 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-gw8b9" Oct 06 10:22:11 crc kubenswrapper[4824]: I1006 10:22:11.294287 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8dc08944-150b-42f6-afb9-161300ac8f96" path="/var/lib/kubelet/pods/8dc08944-150b-42f6-afb9-161300ac8f96/volumes" Oct 06 10:22:12 crc kubenswrapper[4824]: I1006 10:22:12.126323 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-gw8b9"] Oct 06 10:22:13 crc kubenswrapper[4824]: I1006 10:22:13.846427 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-gw8b9" podUID="fc0ce670-992e-4cb5-a283-4835be8d85e3" containerName="registry-server" containerID="cri-o://5d5f01e73b86f8456bc4489546399c1bd7f987b6b7c9d8721cc5755dc9419d52" gracePeriod=2 Oct 06 10:22:13 crc kubenswrapper[4824]: I1006 10:22:13.915327 4824 patch_prober.go:28] interesting pod/machine-config-daemon-khgzw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 10:22:13 crc kubenswrapper[4824]: I1006 10:22:13.915382 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 10:22:14 crc kubenswrapper[4824]: I1006 10:22:14.321705 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gw8b9" Oct 06 10:22:14 crc kubenswrapper[4824]: I1006 10:22:14.476997 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fc0ce670-992e-4cb5-a283-4835be8d85e3-catalog-content\") pod \"fc0ce670-992e-4cb5-a283-4835be8d85e3\" (UID: \"fc0ce670-992e-4cb5-a283-4835be8d85e3\") " Oct 06 10:22:14 crc kubenswrapper[4824]: I1006 10:22:14.477146 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fc0ce670-992e-4cb5-a283-4835be8d85e3-utilities\") pod \"fc0ce670-992e-4cb5-a283-4835be8d85e3\" (UID: \"fc0ce670-992e-4cb5-a283-4835be8d85e3\") " Oct 06 10:22:14 crc kubenswrapper[4824]: I1006 10:22:14.477317 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hlzgf\" (UniqueName: \"kubernetes.io/projected/fc0ce670-992e-4cb5-a283-4835be8d85e3-kube-api-access-hlzgf\") pod \"fc0ce670-992e-4cb5-a283-4835be8d85e3\" (UID: \"fc0ce670-992e-4cb5-a283-4835be8d85e3\") " Oct 06 10:22:14 crc kubenswrapper[4824]: I1006 10:22:14.478202 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fc0ce670-992e-4cb5-a283-4835be8d85e3-utilities" (OuterVolumeSpecName: "utilities") pod "fc0ce670-992e-4cb5-a283-4835be8d85e3" (UID: "fc0ce670-992e-4cb5-a283-4835be8d85e3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:22:14 crc kubenswrapper[4824]: I1006 10:22:14.483780 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fc0ce670-992e-4cb5-a283-4835be8d85e3-kube-api-access-hlzgf" (OuterVolumeSpecName: "kube-api-access-hlzgf") pod "fc0ce670-992e-4cb5-a283-4835be8d85e3" (UID: "fc0ce670-992e-4cb5-a283-4835be8d85e3"). InnerVolumeSpecName "kube-api-access-hlzgf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:22:14 crc kubenswrapper[4824]: I1006 10:22:14.525489 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fc0ce670-992e-4cb5-a283-4835be8d85e3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fc0ce670-992e-4cb5-a283-4835be8d85e3" (UID: "fc0ce670-992e-4cb5-a283-4835be8d85e3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:22:14 crc kubenswrapper[4824]: I1006 10:22:14.579435 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hlzgf\" (UniqueName: \"kubernetes.io/projected/fc0ce670-992e-4cb5-a283-4835be8d85e3-kube-api-access-hlzgf\") on node \"crc\" DevicePath \"\"" Oct 06 10:22:14 crc kubenswrapper[4824]: I1006 10:22:14.579808 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fc0ce670-992e-4cb5-a283-4835be8d85e3-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 06 10:22:14 crc kubenswrapper[4824]: I1006 10:22:14.579876 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fc0ce670-992e-4cb5-a283-4835be8d85e3-utilities\") on node \"crc\" DevicePath \"\"" Oct 06 10:22:14 crc kubenswrapper[4824]: I1006 10:22:14.858271 4824 generic.go:334] "Generic (PLEG): container finished" podID="fc0ce670-992e-4cb5-a283-4835be8d85e3" containerID="5d5f01e73b86f8456bc4489546399c1bd7f987b6b7c9d8721cc5755dc9419d52" exitCode=0 Oct 06 10:22:14 crc kubenswrapper[4824]: I1006 10:22:14.858320 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gw8b9" event={"ID":"fc0ce670-992e-4cb5-a283-4835be8d85e3","Type":"ContainerDied","Data":"5d5f01e73b86f8456bc4489546399c1bd7f987b6b7c9d8721cc5755dc9419d52"} Oct 06 10:22:14 crc kubenswrapper[4824]: I1006 10:22:14.858353 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gw8b9" Oct 06 10:22:14 crc kubenswrapper[4824]: I1006 10:22:14.858393 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gw8b9" event={"ID":"fc0ce670-992e-4cb5-a283-4835be8d85e3","Type":"ContainerDied","Data":"7852484ae03772cd8e3577fd88d54c822c9f48be6abcfe811d7abc03ad85cbde"} Oct 06 10:22:14 crc kubenswrapper[4824]: I1006 10:22:14.858419 4824 scope.go:117] "RemoveContainer" containerID="5d5f01e73b86f8456bc4489546399c1bd7f987b6b7c9d8721cc5755dc9419d52" Oct 06 10:22:14 crc kubenswrapper[4824]: I1006 10:22:14.893329 4824 scope.go:117] "RemoveContainer" containerID="a987a1b1f80c06dc0e77cd4aae5dc583599e873bf656dab04854d16445a0ecf8" Oct 06 10:22:14 crc kubenswrapper[4824]: I1006 10:22:14.896490 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-gw8b9"] Oct 06 10:22:14 crc kubenswrapper[4824]: I1006 10:22:14.904013 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-gw8b9"] Oct 06 10:22:14 crc kubenswrapper[4824]: I1006 10:22:14.922815 4824 scope.go:117] "RemoveContainer" containerID="4da16e1eaad2f68741cd855e8b4df79ae0c91c8dcefa450922f8565b677e5d93" Oct 06 10:22:14 crc kubenswrapper[4824]: I1006 10:22:14.964687 4824 scope.go:117] "RemoveContainer" containerID="5d5f01e73b86f8456bc4489546399c1bd7f987b6b7c9d8721cc5755dc9419d52" Oct 06 10:22:14 crc kubenswrapper[4824]: E1006 10:22:14.965364 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5d5f01e73b86f8456bc4489546399c1bd7f987b6b7c9d8721cc5755dc9419d52\": container with ID starting with 5d5f01e73b86f8456bc4489546399c1bd7f987b6b7c9d8721cc5755dc9419d52 not found: ID does not exist" containerID="5d5f01e73b86f8456bc4489546399c1bd7f987b6b7c9d8721cc5755dc9419d52" Oct 06 10:22:14 crc kubenswrapper[4824]: I1006 10:22:14.965433 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5d5f01e73b86f8456bc4489546399c1bd7f987b6b7c9d8721cc5755dc9419d52"} err="failed to get container status \"5d5f01e73b86f8456bc4489546399c1bd7f987b6b7c9d8721cc5755dc9419d52\": rpc error: code = NotFound desc = could not find container \"5d5f01e73b86f8456bc4489546399c1bd7f987b6b7c9d8721cc5755dc9419d52\": container with ID starting with 5d5f01e73b86f8456bc4489546399c1bd7f987b6b7c9d8721cc5755dc9419d52 not found: ID does not exist" Oct 06 10:22:14 crc kubenswrapper[4824]: I1006 10:22:14.965472 4824 scope.go:117] "RemoveContainer" containerID="a987a1b1f80c06dc0e77cd4aae5dc583599e873bf656dab04854d16445a0ecf8" Oct 06 10:22:14 crc kubenswrapper[4824]: E1006 10:22:14.965935 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a987a1b1f80c06dc0e77cd4aae5dc583599e873bf656dab04854d16445a0ecf8\": container with ID starting with a987a1b1f80c06dc0e77cd4aae5dc583599e873bf656dab04854d16445a0ecf8 not found: ID does not exist" containerID="a987a1b1f80c06dc0e77cd4aae5dc583599e873bf656dab04854d16445a0ecf8" Oct 06 10:22:14 crc kubenswrapper[4824]: I1006 10:22:14.965999 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a987a1b1f80c06dc0e77cd4aae5dc583599e873bf656dab04854d16445a0ecf8"} err="failed to get container status \"a987a1b1f80c06dc0e77cd4aae5dc583599e873bf656dab04854d16445a0ecf8\": rpc error: code = NotFound desc = could not find container \"a987a1b1f80c06dc0e77cd4aae5dc583599e873bf656dab04854d16445a0ecf8\": container with ID starting with a987a1b1f80c06dc0e77cd4aae5dc583599e873bf656dab04854d16445a0ecf8 not found: ID does not exist" Oct 06 10:22:14 crc kubenswrapper[4824]: I1006 10:22:14.966021 4824 scope.go:117] "RemoveContainer" containerID="4da16e1eaad2f68741cd855e8b4df79ae0c91c8dcefa450922f8565b677e5d93" Oct 06 10:22:14 crc kubenswrapper[4824]: E1006 10:22:14.966289 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4da16e1eaad2f68741cd855e8b4df79ae0c91c8dcefa450922f8565b677e5d93\": container with ID starting with 4da16e1eaad2f68741cd855e8b4df79ae0c91c8dcefa450922f8565b677e5d93 not found: ID does not exist" containerID="4da16e1eaad2f68741cd855e8b4df79ae0c91c8dcefa450922f8565b677e5d93" Oct 06 10:22:14 crc kubenswrapper[4824]: I1006 10:22:14.966315 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4da16e1eaad2f68741cd855e8b4df79ae0c91c8dcefa450922f8565b677e5d93"} err="failed to get container status \"4da16e1eaad2f68741cd855e8b4df79ae0c91c8dcefa450922f8565b677e5d93\": rpc error: code = NotFound desc = could not find container \"4da16e1eaad2f68741cd855e8b4df79ae0c91c8dcefa450922f8565b677e5d93\": container with ID starting with 4da16e1eaad2f68741cd855e8b4df79ae0c91c8dcefa450922f8565b677e5d93 not found: ID does not exist" Oct 06 10:22:15 crc kubenswrapper[4824]: I1006 10:22:15.140958 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-rgtl6"] Oct 06 10:22:15 crc kubenswrapper[4824]: E1006 10:22:15.141472 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8dc08944-150b-42f6-afb9-161300ac8f96" containerName="extract-utilities" Oct 06 10:22:15 crc kubenswrapper[4824]: I1006 10:22:15.141495 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="8dc08944-150b-42f6-afb9-161300ac8f96" containerName="extract-utilities" Oct 06 10:22:15 crc kubenswrapper[4824]: E1006 10:22:15.141524 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc0ce670-992e-4cb5-a283-4835be8d85e3" containerName="registry-server" Oct 06 10:22:15 crc kubenswrapper[4824]: I1006 10:22:15.141532 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc0ce670-992e-4cb5-a283-4835be8d85e3" containerName="registry-server" Oct 06 10:22:15 crc kubenswrapper[4824]: E1006 10:22:15.141546 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8dc08944-150b-42f6-afb9-161300ac8f96" containerName="registry-server" Oct 06 10:22:15 crc kubenswrapper[4824]: I1006 10:22:15.141566 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="8dc08944-150b-42f6-afb9-161300ac8f96" containerName="registry-server" Oct 06 10:22:15 crc kubenswrapper[4824]: E1006 10:22:15.141579 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc0ce670-992e-4cb5-a283-4835be8d85e3" containerName="extract-content" Oct 06 10:22:15 crc kubenswrapper[4824]: I1006 10:22:15.141585 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc0ce670-992e-4cb5-a283-4835be8d85e3" containerName="extract-content" Oct 06 10:22:15 crc kubenswrapper[4824]: E1006 10:22:15.141595 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8dc08944-150b-42f6-afb9-161300ac8f96" containerName="extract-content" Oct 06 10:22:15 crc kubenswrapper[4824]: I1006 10:22:15.141601 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="8dc08944-150b-42f6-afb9-161300ac8f96" containerName="extract-content" Oct 06 10:22:15 crc kubenswrapper[4824]: E1006 10:22:15.141613 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc0ce670-992e-4cb5-a283-4835be8d85e3" containerName="extract-utilities" Oct 06 10:22:15 crc kubenswrapper[4824]: I1006 10:22:15.141619 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc0ce670-992e-4cb5-a283-4835be8d85e3" containerName="extract-utilities" Oct 06 10:22:15 crc kubenswrapper[4824]: I1006 10:22:15.141821 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="fc0ce670-992e-4cb5-a283-4835be8d85e3" containerName="registry-server" Oct 06 10:22:15 crc kubenswrapper[4824]: I1006 10:22:15.141842 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="8dc08944-150b-42f6-afb9-161300ac8f96" containerName="registry-server" Oct 06 10:22:15 crc kubenswrapper[4824]: I1006 10:22:15.143290 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rgtl6" Oct 06 10:22:15 crc kubenswrapper[4824]: I1006 10:22:15.178988 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rgtl6"] Oct 06 10:22:15 crc kubenswrapper[4824]: I1006 10:22:15.286739 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fc0ce670-992e-4cb5-a283-4835be8d85e3" path="/var/lib/kubelet/pods/fc0ce670-992e-4cb5-a283-4835be8d85e3/volumes" Oct 06 10:22:15 crc kubenswrapper[4824]: I1006 10:22:15.303518 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nzwdq\" (UniqueName: \"kubernetes.io/projected/73ac5c84-1924-49ec-9182-20c2d196e0ec-kube-api-access-nzwdq\") pod \"certified-operators-rgtl6\" (UID: \"73ac5c84-1924-49ec-9182-20c2d196e0ec\") " pod="openshift-marketplace/certified-operators-rgtl6" Oct 06 10:22:15 crc kubenswrapper[4824]: I1006 10:22:15.305261 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/73ac5c84-1924-49ec-9182-20c2d196e0ec-catalog-content\") pod \"certified-operators-rgtl6\" (UID: \"73ac5c84-1924-49ec-9182-20c2d196e0ec\") " pod="openshift-marketplace/certified-operators-rgtl6" Oct 06 10:22:15 crc kubenswrapper[4824]: I1006 10:22:15.305713 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/73ac5c84-1924-49ec-9182-20c2d196e0ec-utilities\") pod \"certified-operators-rgtl6\" (UID: \"73ac5c84-1924-49ec-9182-20c2d196e0ec\") " pod="openshift-marketplace/certified-operators-rgtl6" Oct 06 10:22:15 crc kubenswrapper[4824]: I1006 10:22:15.410353 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/73ac5c84-1924-49ec-9182-20c2d196e0ec-utilities\") pod \"certified-operators-rgtl6\" (UID: \"73ac5c84-1924-49ec-9182-20c2d196e0ec\") " pod="openshift-marketplace/certified-operators-rgtl6" Oct 06 10:22:15 crc kubenswrapper[4824]: I1006 10:22:15.410711 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nzwdq\" (UniqueName: \"kubernetes.io/projected/73ac5c84-1924-49ec-9182-20c2d196e0ec-kube-api-access-nzwdq\") pod \"certified-operators-rgtl6\" (UID: \"73ac5c84-1924-49ec-9182-20c2d196e0ec\") " pod="openshift-marketplace/certified-operators-rgtl6" Oct 06 10:22:15 crc kubenswrapper[4824]: I1006 10:22:15.410809 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/73ac5c84-1924-49ec-9182-20c2d196e0ec-catalog-content\") pod \"certified-operators-rgtl6\" (UID: \"73ac5c84-1924-49ec-9182-20c2d196e0ec\") " pod="openshift-marketplace/certified-operators-rgtl6" Oct 06 10:22:15 crc kubenswrapper[4824]: I1006 10:22:15.410891 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/73ac5c84-1924-49ec-9182-20c2d196e0ec-utilities\") pod \"certified-operators-rgtl6\" (UID: \"73ac5c84-1924-49ec-9182-20c2d196e0ec\") " pod="openshift-marketplace/certified-operators-rgtl6" Oct 06 10:22:15 crc kubenswrapper[4824]: I1006 10:22:15.411173 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/73ac5c84-1924-49ec-9182-20c2d196e0ec-catalog-content\") pod \"certified-operators-rgtl6\" (UID: \"73ac5c84-1924-49ec-9182-20c2d196e0ec\") " pod="openshift-marketplace/certified-operators-rgtl6" Oct 06 10:22:15 crc kubenswrapper[4824]: I1006 10:22:15.434148 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nzwdq\" (UniqueName: \"kubernetes.io/projected/73ac5c84-1924-49ec-9182-20c2d196e0ec-kube-api-access-nzwdq\") pod \"certified-operators-rgtl6\" (UID: \"73ac5c84-1924-49ec-9182-20c2d196e0ec\") " pod="openshift-marketplace/certified-operators-rgtl6" Oct 06 10:22:15 crc kubenswrapper[4824]: I1006 10:22:15.475323 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rgtl6" Oct 06 10:22:15 crc kubenswrapper[4824]: I1006 10:22:15.976652 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rgtl6"] Oct 06 10:22:15 crc kubenswrapper[4824]: W1006 10:22:15.987814 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod73ac5c84_1924_49ec_9182_20c2d196e0ec.slice/crio-44fa62f25ea757180b60534f0b99cc858b83ed1dc07d75c3226cdf6b017e3346 WatchSource:0}: Error finding container 44fa62f25ea757180b60534f0b99cc858b83ed1dc07d75c3226cdf6b017e3346: Status 404 returned error can't find the container with id 44fa62f25ea757180b60534f0b99cc858b83ed1dc07d75c3226cdf6b017e3346 Oct 06 10:22:16 crc kubenswrapper[4824]: I1006 10:22:16.886645 4824 generic.go:334] "Generic (PLEG): container finished" podID="73ac5c84-1924-49ec-9182-20c2d196e0ec" containerID="3d30bb06e7b7a32c9b4de56584675693c534d298c20c55859abfb4ac298ce977" exitCode=0 Oct 06 10:22:16 crc kubenswrapper[4824]: I1006 10:22:16.886770 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rgtl6" event={"ID":"73ac5c84-1924-49ec-9182-20c2d196e0ec","Type":"ContainerDied","Data":"3d30bb06e7b7a32c9b4de56584675693c534d298c20c55859abfb4ac298ce977"} Oct 06 10:22:16 crc kubenswrapper[4824]: I1006 10:22:16.887343 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rgtl6" event={"ID":"73ac5c84-1924-49ec-9182-20c2d196e0ec","Type":"ContainerStarted","Data":"44fa62f25ea757180b60534f0b99cc858b83ed1dc07d75c3226cdf6b017e3346"} Oct 06 10:22:17 crc kubenswrapper[4824]: I1006 10:22:17.910320 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rgtl6" event={"ID":"73ac5c84-1924-49ec-9182-20c2d196e0ec","Type":"ContainerStarted","Data":"5973aaeb458d45c346cc248b3ecae4a382aa63b77a7c0b51b1b3a833a0edca21"} Oct 06 10:22:18 crc kubenswrapper[4824]: I1006 10:22:18.928274 4824 generic.go:334] "Generic (PLEG): container finished" podID="73ac5c84-1924-49ec-9182-20c2d196e0ec" containerID="5973aaeb458d45c346cc248b3ecae4a382aa63b77a7c0b51b1b3a833a0edca21" exitCode=0 Oct 06 10:22:18 crc kubenswrapper[4824]: I1006 10:22:18.928333 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rgtl6" event={"ID":"73ac5c84-1924-49ec-9182-20c2d196e0ec","Type":"ContainerDied","Data":"5973aaeb458d45c346cc248b3ecae4a382aa63b77a7c0b51b1b3a833a0edca21"} Oct 06 10:22:19 crc kubenswrapper[4824]: I1006 10:22:19.944490 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rgtl6" event={"ID":"73ac5c84-1924-49ec-9182-20c2d196e0ec","Type":"ContainerStarted","Data":"728e39401219dc3fc9598496c6feafb3c8848247b21a9af63d980ed183ecea92"} Oct 06 10:22:19 crc kubenswrapper[4824]: I1006 10:22:19.973285 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-rgtl6" podStartSLOduration=2.377244236 podStartE2EDuration="4.973257608s" podCreationTimestamp="2025-10-06 10:22:15 +0000 UTC" firstStartedPulling="2025-10-06 10:22:16.891185322 +0000 UTC m=+1446.255608223" lastFinishedPulling="2025-10-06 10:22:19.487198734 +0000 UTC m=+1448.851621595" observedRunningTime="2025-10-06 10:22:19.966433745 +0000 UTC m=+1449.330856606" watchObservedRunningTime="2025-10-06 10:22:19.973257608 +0000 UTC m=+1449.337680489" Oct 06 10:22:25 crc kubenswrapper[4824]: I1006 10:22:25.476513 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-rgtl6" Oct 06 10:22:25 crc kubenswrapper[4824]: I1006 10:22:25.477398 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-rgtl6" Oct 06 10:22:25 crc kubenswrapper[4824]: I1006 10:22:25.540175 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-rgtl6" Oct 06 10:22:26 crc kubenswrapper[4824]: I1006 10:22:26.089159 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-rgtl6" Oct 06 10:22:26 crc kubenswrapper[4824]: I1006 10:22:26.170307 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rgtl6"] Oct 06 10:22:28 crc kubenswrapper[4824]: I1006 10:22:28.041813 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-rgtl6" podUID="73ac5c84-1924-49ec-9182-20c2d196e0ec" containerName="registry-server" containerID="cri-o://728e39401219dc3fc9598496c6feafb3c8848247b21a9af63d980ed183ecea92" gracePeriod=2 Oct 06 10:22:28 crc kubenswrapper[4824]: I1006 10:22:28.551080 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rgtl6" Oct 06 10:22:28 crc kubenswrapper[4824]: I1006 10:22:28.754087 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/73ac5c84-1924-49ec-9182-20c2d196e0ec-catalog-content\") pod \"73ac5c84-1924-49ec-9182-20c2d196e0ec\" (UID: \"73ac5c84-1924-49ec-9182-20c2d196e0ec\") " Oct 06 10:22:28 crc kubenswrapper[4824]: I1006 10:22:28.754548 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/73ac5c84-1924-49ec-9182-20c2d196e0ec-utilities\") pod \"73ac5c84-1924-49ec-9182-20c2d196e0ec\" (UID: \"73ac5c84-1924-49ec-9182-20c2d196e0ec\") " Oct 06 10:22:28 crc kubenswrapper[4824]: I1006 10:22:28.756384 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwdq\" (UniqueName: \"kubernetes.io/projected/73ac5c84-1924-49ec-9182-20c2d196e0ec-kube-api-access-nzwdq\") pod \"73ac5c84-1924-49ec-9182-20c2d196e0ec\" (UID: \"73ac5c84-1924-49ec-9182-20c2d196e0ec\") " Oct 06 10:22:28 crc kubenswrapper[4824]: I1006 10:22:28.756536 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/73ac5c84-1924-49ec-9182-20c2d196e0ec-utilities" (OuterVolumeSpecName: "utilities") pod "73ac5c84-1924-49ec-9182-20c2d196e0ec" (UID: "73ac5c84-1924-49ec-9182-20c2d196e0ec"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:22:28 crc kubenswrapper[4824]: I1006 10:22:28.757801 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/73ac5c84-1924-49ec-9182-20c2d196e0ec-utilities\") on node \"crc\" DevicePath \"\"" Oct 06 10:22:28 crc kubenswrapper[4824]: I1006 10:22:28.770848 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/73ac5c84-1924-49ec-9182-20c2d196e0ec-kube-api-access-nzwdq" (OuterVolumeSpecName: "kube-api-access-nzwdq") pod "73ac5c84-1924-49ec-9182-20c2d196e0ec" (UID: "73ac5c84-1924-49ec-9182-20c2d196e0ec"). InnerVolumeSpecName "kube-api-access-nzwdq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:22:28 crc kubenswrapper[4824]: I1006 10:22:28.832940 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/73ac5c84-1924-49ec-9182-20c2d196e0ec-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "73ac5c84-1924-49ec-9182-20c2d196e0ec" (UID: "73ac5c84-1924-49ec-9182-20c2d196e0ec"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:22:28 crc kubenswrapper[4824]: I1006 10:22:28.859418 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwdq\" (UniqueName: \"kubernetes.io/projected/73ac5c84-1924-49ec-9182-20c2d196e0ec-kube-api-access-nzwdq\") on node \"crc\" DevicePath \"\"" Oct 06 10:22:28 crc kubenswrapper[4824]: I1006 10:22:28.859478 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/73ac5c84-1924-49ec-9182-20c2d196e0ec-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 06 10:22:29 crc kubenswrapper[4824]: I1006 10:22:29.059495 4824 generic.go:334] "Generic (PLEG): container finished" podID="73ac5c84-1924-49ec-9182-20c2d196e0ec" containerID="728e39401219dc3fc9598496c6feafb3c8848247b21a9af63d980ed183ecea92" exitCode=0 Oct 06 10:22:29 crc kubenswrapper[4824]: I1006 10:22:29.059550 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rgtl6" event={"ID":"73ac5c84-1924-49ec-9182-20c2d196e0ec","Type":"ContainerDied","Data":"728e39401219dc3fc9598496c6feafb3c8848247b21a9af63d980ed183ecea92"} Oct 06 10:22:29 crc kubenswrapper[4824]: I1006 10:22:29.059592 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rgtl6" event={"ID":"73ac5c84-1924-49ec-9182-20c2d196e0ec","Type":"ContainerDied","Data":"44fa62f25ea757180b60534f0b99cc858b83ed1dc07d75c3226cdf6b017e3346"} Oct 06 10:22:29 crc kubenswrapper[4824]: I1006 10:22:29.059613 4824 scope.go:117] "RemoveContainer" containerID="728e39401219dc3fc9598496c6feafb3c8848247b21a9af63d980ed183ecea92" Oct 06 10:22:29 crc kubenswrapper[4824]: I1006 10:22:29.059692 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rgtl6" Oct 06 10:22:29 crc kubenswrapper[4824]: I1006 10:22:29.092805 4824 scope.go:117] "RemoveContainer" containerID="5973aaeb458d45c346cc248b3ecae4a382aa63b77a7c0b51b1b3a833a0edca21" Oct 06 10:22:29 crc kubenswrapper[4824]: I1006 10:22:29.128070 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rgtl6"] Oct 06 10:22:29 crc kubenswrapper[4824]: I1006 10:22:29.135455 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-rgtl6"] Oct 06 10:22:29 crc kubenswrapper[4824]: I1006 10:22:29.146498 4824 scope.go:117] "RemoveContainer" containerID="3d30bb06e7b7a32c9b4de56584675693c534d298c20c55859abfb4ac298ce977" Oct 06 10:22:29 crc kubenswrapper[4824]: I1006 10:22:29.201831 4824 scope.go:117] "RemoveContainer" containerID="728e39401219dc3fc9598496c6feafb3c8848247b21a9af63d980ed183ecea92" Oct 06 10:22:29 crc kubenswrapper[4824]: E1006 10:22:29.202728 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"728e39401219dc3fc9598496c6feafb3c8848247b21a9af63d980ed183ecea92\": container with ID starting with 728e39401219dc3fc9598496c6feafb3c8848247b21a9af63d980ed183ecea92 not found: ID does not exist" containerID="728e39401219dc3fc9598496c6feafb3c8848247b21a9af63d980ed183ecea92" Oct 06 10:22:29 crc kubenswrapper[4824]: I1006 10:22:29.202786 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"728e39401219dc3fc9598496c6feafb3c8848247b21a9af63d980ed183ecea92"} err="failed to get container status \"728e39401219dc3fc9598496c6feafb3c8848247b21a9af63d980ed183ecea92\": rpc error: code = NotFound desc = could not find container \"728e39401219dc3fc9598496c6feafb3c8848247b21a9af63d980ed183ecea92\": container with ID starting with 728e39401219dc3fc9598496c6feafb3c8848247b21a9af63d980ed183ecea92 not found: ID does not exist" Oct 06 10:22:29 crc kubenswrapper[4824]: I1006 10:22:29.202818 4824 scope.go:117] "RemoveContainer" containerID="5973aaeb458d45c346cc248b3ecae4a382aa63b77a7c0b51b1b3a833a0edca21" Oct 06 10:22:29 crc kubenswrapper[4824]: E1006 10:22:29.203387 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5973aaeb458d45c346cc248b3ecae4a382aa63b77a7c0b51b1b3a833a0edca21\": container with ID starting with 5973aaeb458d45c346cc248b3ecae4a382aa63b77a7c0b51b1b3a833a0edca21 not found: ID does not exist" containerID="5973aaeb458d45c346cc248b3ecae4a382aa63b77a7c0b51b1b3a833a0edca21" Oct 06 10:22:29 crc kubenswrapper[4824]: I1006 10:22:29.203430 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5973aaeb458d45c346cc248b3ecae4a382aa63b77a7c0b51b1b3a833a0edca21"} err="failed to get container status \"5973aaeb458d45c346cc248b3ecae4a382aa63b77a7c0b51b1b3a833a0edca21\": rpc error: code = NotFound desc = could not find container \"5973aaeb458d45c346cc248b3ecae4a382aa63b77a7c0b51b1b3a833a0edca21\": container with ID starting with 5973aaeb458d45c346cc248b3ecae4a382aa63b77a7c0b51b1b3a833a0edca21 not found: ID does not exist" Oct 06 10:22:29 crc kubenswrapper[4824]: I1006 10:22:29.203459 4824 scope.go:117] "RemoveContainer" containerID="3d30bb06e7b7a32c9b4de56584675693c534d298c20c55859abfb4ac298ce977" Oct 06 10:22:29 crc kubenswrapper[4824]: E1006 10:22:29.204217 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3d30bb06e7b7a32c9b4de56584675693c534d298c20c55859abfb4ac298ce977\": container with ID starting with 3d30bb06e7b7a32c9b4de56584675693c534d298c20c55859abfb4ac298ce977 not found: ID does not exist" containerID="3d30bb06e7b7a32c9b4de56584675693c534d298c20c55859abfb4ac298ce977" Oct 06 10:22:29 crc kubenswrapper[4824]: I1006 10:22:29.204289 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3d30bb06e7b7a32c9b4de56584675693c534d298c20c55859abfb4ac298ce977"} err="failed to get container status \"3d30bb06e7b7a32c9b4de56584675693c534d298c20c55859abfb4ac298ce977\": rpc error: code = NotFound desc = could not find container \"3d30bb06e7b7a32c9b4de56584675693c534d298c20c55859abfb4ac298ce977\": container with ID starting with 3d30bb06e7b7a32c9b4de56584675693c534d298c20c55859abfb4ac298ce977 not found: ID does not exist" Oct 06 10:22:29 crc kubenswrapper[4824]: I1006 10:22:29.290921 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="73ac5c84-1924-49ec-9182-20c2d196e0ec" path="/var/lib/kubelet/pods/73ac5c84-1924-49ec-9182-20c2d196e0ec/volumes" Oct 06 10:22:30 crc kubenswrapper[4824]: I1006 10:22:30.475318 4824 scope.go:117] "RemoveContainer" containerID="e21b3838b55eab22121026e4bcc94e71b1ddcd8a052894021b1b38e01c5f0f50" Oct 06 10:22:30 crc kubenswrapper[4824]: I1006 10:22:30.547081 4824 scope.go:117] "RemoveContainer" containerID="d09824b8fc32e72c2d6d30bfea36532762fa0afa48b2d6885a6612dd89035ad0" Oct 06 10:22:30 crc kubenswrapper[4824]: I1006 10:22:30.589366 4824 scope.go:117] "RemoveContainer" containerID="51657330175f5797f5507a724e5f16d8bc068fe1ab011ba8ea18d9f8f4298e0c" Oct 06 10:22:30 crc kubenswrapper[4824]: I1006 10:22:30.626078 4824 scope.go:117] "RemoveContainer" containerID="4b4abcc8eb3bb8c1ba125783f2c3e87528d26f2c2a2fae76b8c8c4f23d84ff1f" Oct 06 10:22:30 crc kubenswrapper[4824]: I1006 10:22:30.652096 4824 scope.go:117] "RemoveContainer" containerID="bd755795208f73b11771157f14d20e980d843dd07c235a308a1c5ef8ace0f43f" Oct 06 10:22:43 crc kubenswrapper[4824]: I1006 10:22:43.916937 4824 patch_prober.go:28] interesting pod/machine-config-daemon-khgzw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 10:22:43 crc kubenswrapper[4824]: I1006 10:22:43.917622 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 10:22:43 crc kubenswrapper[4824]: I1006 10:22:43.917692 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" Oct 06 10:22:43 crc kubenswrapper[4824]: I1006 10:22:43.918710 4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7f680663f7fe4c1e2c1c711b5dc3a94dfad5110e11cc00fd374e4d6a79007f09"} pod="openshift-machine-config-operator/machine-config-daemon-khgzw" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 06 10:22:43 crc kubenswrapper[4824]: I1006 10:22:43.918778 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" containerName="machine-config-daemon" containerID="cri-o://7f680663f7fe4c1e2c1c711b5dc3a94dfad5110e11cc00fd374e4d6a79007f09" gracePeriod=600 Oct 06 10:22:44 crc kubenswrapper[4824]: I1006 10:22:44.233297 4824 generic.go:334] "Generic (PLEG): container finished" podID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" containerID="7f680663f7fe4c1e2c1c711b5dc3a94dfad5110e11cc00fd374e4d6a79007f09" exitCode=0 Oct 06 10:22:44 crc kubenswrapper[4824]: I1006 10:22:44.233388 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" event={"ID":"e1140c8c-93f1-4fce-8c06-a69561ff8a6d","Type":"ContainerDied","Data":"7f680663f7fe4c1e2c1c711b5dc3a94dfad5110e11cc00fd374e4d6a79007f09"} Oct 06 10:22:44 crc kubenswrapper[4824]: I1006 10:22:44.233825 4824 scope.go:117] "RemoveContainer" containerID="ad63fa5818247f7e35384bdeb2bf047f095a2d6eed77601da04ba8f26adf2f37" Oct 06 10:22:45 crc kubenswrapper[4824]: I1006 10:22:45.247908 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" event={"ID":"e1140c8c-93f1-4fce-8c06-a69561ff8a6d","Type":"ContainerStarted","Data":"9de43c7f48ed9ad071a24adf937829ab92276bd505bb72d5b89bac4922aac3b0"} Oct 06 10:24:13 crc kubenswrapper[4824]: I1006 10:24:13.211234 4824 generic.go:334] "Generic (PLEG): container finished" podID="c995dce1-6fa5-456b-b984-b397dcc9f9dc" containerID="550b5638ac37f01b97333520fd691b2cef12d4f0439856452d41c374c9d08517" exitCode=0 Oct 06 10:24:13 crc kubenswrapper[4824]: I1006 10:24:13.211396 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-mlckl" event={"ID":"c995dce1-6fa5-456b-b984-b397dcc9f9dc","Type":"ContainerDied","Data":"550b5638ac37f01b97333520fd691b2cef12d4f0439856452d41c374c9d08517"} Oct 06 10:24:14 crc kubenswrapper[4824]: I1006 10:24:14.717925 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-mlckl" Oct 06 10:24:14 crc kubenswrapper[4824]: I1006 10:24:14.885933 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c995dce1-6fa5-456b-b984-b397dcc9f9dc-bootstrap-combined-ca-bundle\") pod \"c995dce1-6fa5-456b-b984-b397dcc9f9dc\" (UID: \"c995dce1-6fa5-456b-b984-b397dcc9f9dc\") " Oct 06 10:24:14 crc kubenswrapper[4824]: I1006 10:24:14.886422 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c995dce1-6fa5-456b-b984-b397dcc9f9dc-ssh-key\") pod \"c995dce1-6fa5-456b-b984-b397dcc9f9dc\" (UID: \"c995dce1-6fa5-456b-b984-b397dcc9f9dc\") " Oct 06 10:24:14 crc kubenswrapper[4824]: I1006 10:24:14.886530 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5x8hs\" (UniqueName: \"kubernetes.io/projected/c995dce1-6fa5-456b-b984-b397dcc9f9dc-kube-api-access-5x8hs\") pod \"c995dce1-6fa5-456b-b984-b397dcc9f9dc\" (UID: \"c995dce1-6fa5-456b-b984-b397dcc9f9dc\") " Oct 06 10:24:14 crc kubenswrapper[4824]: I1006 10:24:14.886609 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c995dce1-6fa5-456b-b984-b397dcc9f9dc-inventory\") pod \"c995dce1-6fa5-456b-b984-b397dcc9f9dc\" (UID: \"c995dce1-6fa5-456b-b984-b397dcc9f9dc\") " Oct 06 10:24:14 crc kubenswrapper[4824]: I1006 10:24:14.893676 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c995dce1-6fa5-456b-b984-b397dcc9f9dc-kube-api-access-5x8hs" (OuterVolumeSpecName: "kube-api-access-5x8hs") pod "c995dce1-6fa5-456b-b984-b397dcc9f9dc" (UID: "c995dce1-6fa5-456b-b984-b397dcc9f9dc"). InnerVolumeSpecName "kube-api-access-5x8hs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:24:14 crc kubenswrapper[4824]: I1006 10:24:14.894433 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c995dce1-6fa5-456b-b984-b397dcc9f9dc-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "c995dce1-6fa5-456b-b984-b397dcc9f9dc" (UID: "c995dce1-6fa5-456b-b984-b397dcc9f9dc"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:24:14 crc kubenswrapper[4824]: I1006 10:24:14.918060 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c995dce1-6fa5-456b-b984-b397dcc9f9dc-inventory" (OuterVolumeSpecName: "inventory") pod "c995dce1-6fa5-456b-b984-b397dcc9f9dc" (UID: "c995dce1-6fa5-456b-b984-b397dcc9f9dc"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:24:14 crc kubenswrapper[4824]: I1006 10:24:14.939716 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c995dce1-6fa5-456b-b984-b397dcc9f9dc-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "c995dce1-6fa5-456b-b984-b397dcc9f9dc" (UID: "c995dce1-6fa5-456b-b984-b397dcc9f9dc"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:24:14 crc kubenswrapper[4824]: I1006 10:24:14.989287 4824 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c995dce1-6fa5-456b-b984-b397dcc9f9dc-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 10:24:14 crc kubenswrapper[4824]: I1006 10:24:14.989331 4824 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c995dce1-6fa5-456b-b984-b397dcc9f9dc-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 06 10:24:14 crc kubenswrapper[4824]: I1006 10:24:14.989341 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5x8hs\" (UniqueName: \"kubernetes.io/projected/c995dce1-6fa5-456b-b984-b397dcc9f9dc-kube-api-access-5x8hs\") on node \"crc\" DevicePath \"\"" Oct 06 10:24:14 crc kubenswrapper[4824]: I1006 10:24:14.989351 4824 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c995dce1-6fa5-456b-b984-b397dcc9f9dc-inventory\") on node \"crc\" DevicePath \"\"" Oct 06 10:24:15 crc kubenswrapper[4824]: I1006 10:24:15.232822 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-mlckl" event={"ID":"c995dce1-6fa5-456b-b984-b397dcc9f9dc","Type":"ContainerDied","Data":"8b605b390d04aeadf5dda98fad576bb262657d9a1a7c9087513b0de8ca2b18a5"} Oct 06 10:24:15 crc kubenswrapper[4824]: I1006 10:24:15.232878 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8b605b390d04aeadf5dda98fad576bb262657d9a1a7c9087513b0de8ca2b18a5" Oct 06 10:24:15 crc kubenswrapper[4824]: I1006 10:24:15.232883 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-mlckl" Oct 06 10:24:15 crc kubenswrapper[4824]: I1006 10:24:15.362313 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-kcvhh"] Oct 06 10:24:15 crc kubenswrapper[4824]: E1006 10:24:15.362942 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c995dce1-6fa5-456b-b984-b397dcc9f9dc" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Oct 06 10:24:15 crc kubenswrapper[4824]: I1006 10:24:15.362968 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="c995dce1-6fa5-456b-b984-b397dcc9f9dc" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Oct 06 10:24:15 crc kubenswrapper[4824]: E1006 10:24:15.363015 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73ac5c84-1924-49ec-9182-20c2d196e0ec" containerName="registry-server" Oct 06 10:24:15 crc kubenswrapper[4824]: I1006 10:24:15.363024 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="73ac5c84-1924-49ec-9182-20c2d196e0ec" containerName="registry-server" Oct 06 10:24:15 crc kubenswrapper[4824]: E1006 10:24:15.363045 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73ac5c84-1924-49ec-9182-20c2d196e0ec" containerName="extract-content" Oct 06 10:24:15 crc kubenswrapper[4824]: I1006 10:24:15.363053 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="73ac5c84-1924-49ec-9182-20c2d196e0ec" containerName="extract-content" Oct 06 10:24:15 crc kubenswrapper[4824]: E1006 10:24:15.363117 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73ac5c84-1924-49ec-9182-20c2d196e0ec" containerName="extract-utilities" Oct 06 10:24:15 crc kubenswrapper[4824]: I1006 10:24:15.363127 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="73ac5c84-1924-49ec-9182-20c2d196e0ec" containerName="extract-utilities" Oct 06 10:24:15 crc kubenswrapper[4824]: I1006 10:24:15.363374 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="c995dce1-6fa5-456b-b984-b397dcc9f9dc" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Oct 06 10:24:15 crc kubenswrapper[4824]: I1006 10:24:15.363399 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="73ac5c84-1924-49ec-9182-20c2d196e0ec" containerName="registry-server" Oct 06 10:24:15 crc kubenswrapper[4824]: I1006 10:24:15.364349 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-kcvhh" Oct 06 10:24:15 crc kubenswrapper[4824]: I1006 10:24:15.378398 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-kcvhh"] Oct 06 10:24:15 crc kubenswrapper[4824]: I1006 10:24:15.382842 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-f886w" Oct 06 10:24:15 crc kubenswrapper[4824]: I1006 10:24:15.383138 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 06 10:24:15 crc kubenswrapper[4824]: I1006 10:24:15.383640 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 06 10:24:15 crc kubenswrapper[4824]: I1006 10:24:15.383727 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 06 10:24:15 crc kubenswrapper[4824]: I1006 10:24:15.506371 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8ea3d60c-80c7-4163-9f81-c0ec20e758c9-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-kcvhh\" (UID: \"8ea3d60c-80c7-4163-9f81-c0ec20e758c9\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-kcvhh" Oct 06 10:24:15 crc kubenswrapper[4824]: I1006 10:24:15.506547 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8ea3d60c-80c7-4163-9f81-c0ec20e758c9-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-kcvhh\" (UID: \"8ea3d60c-80c7-4163-9f81-c0ec20e758c9\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-kcvhh" Oct 06 10:24:15 crc kubenswrapper[4824]: I1006 10:24:15.507540 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vb6kf\" (UniqueName: \"kubernetes.io/projected/8ea3d60c-80c7-4163-9f81-c0ec20e758c9-kube-api-access-vb6kf\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-kcvhh\" (UID: \"8ea3d60c-80c7-4163-9f81-c0ec20e758c9\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-kcvhh" Oct 06 10:24:15 crc kubenswrapper[4824]: I1006 10:24:15.610118 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8ea3d60c-80c7-4163-9f81-c0ec20e758c9-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-kcvhh\" (UID: \"8ea3d60c-80c7-4163-9f81-c0ec20e758c9\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-kcvhh" Oct 06 10:24:15 crc kubenswrapper[4824]: I1006 10:24:15.610682 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8ea3d60c-80c7-4163-9f81-c0ec20e758c9-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-kcvhh\" (UID: \"8ea3d60c-80c7-4163-9f81-c0ec20e758c9\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-kcvhh" Oct 06 10:24:15 crc kubenswrapper[4824]: I1006 10:24:15.610933 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vb6kf\" (UniqueName: \"kubernetes.io/projected/8ea3d60c-80c7-4163-9f81-c0ec20e758c9-kube-api-access-vb6kf\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-kcvhh\" (UID: \"8ea3d60c-80c7-4163-9f81-c0ec20e758c9\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-kcvhh" Oct 06 10:24:15 crc kubenswrapper[4824]: I1006 10:24:15.614879 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8ea3d60c-80c7-4163-9f81-c0ec20e758c9-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-kcvhh\" (UID: \"8ea3d60c-80c7-4163-9f81-c0ec20e758c9\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-kcvhh" Oct 06 10:24:15 crc kubenswrapper[4824]: I1006 10:24:15.617888 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8ea3d60c-80c7-4163-9f81-c0ec20e758c9-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-kcvhh\" (UID: \"8ea3d60c-80c7-4163-9f81-c0ec20e758c9\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-kcvhh" Oct 06 10:24:15 crc kubenswrapper[4824]: I1006 10:24:15.631138 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vb6kf\" (UniqueName: \"kubernetes.io/projected/8ea3d60c-80c7-4163-9f81-c0ec20e758c9-kube-api-access-vb6kf\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-kcvhh\" (UID: \"8ea3d60c-80c7-4163-9f81-c0ec20e758c9\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-kcvhh" Oct 06 10:24:15 crc kubenswrapper[4824]: I1006 10:24:15.701899 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-kcvhh" Oct 06 10:24:16 crc kubenswrapper[4824]: I1006 10:24:16.266231 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-kcvhh"] Oct 06 10:24:16 crc kubenswrapper[4824]: I1006 10:24:16.271492 4824 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 06 10:24:17 crc kubenswrapper[4824]: I1006 10:24:17.254141 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-kcvhh" event={"ID":"8ea3d60c-80c7-4163-9f81-c0ec20e758c9","Type":"ContainerStarted","Data":"f16c6c16083a3e2b88f2fbcb3faa9ee204a9dd542a8d0dcb2955d3fa4855841b"} Oct 06 10:24:17 crc kubenswrapper[4824]: I1006 10:24:17.254514 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-kcvhh" event={"ID":"8ea3d60c-80c7-4163-9f81-c0ec20e758c9","Type":"ContainerStarted","Data":"4d060b590ac350451bceaac2039d004b906d02efe77aeff328c1fe07659bc5a1"} Oct 06 10:24:17 crc kubenswrapper[4824]: I1006 10:24:17.280630 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-kcvhh" podStartSLOduration=1.690228983 podStartE2EDuration="2.280607367s" podCreationTimestamp="2025-10-06 10:24:15 +0000 UTC" firstStartedPulling="2025-10-06 10:24:16.271046673 +0000 UTC m=+1565.635469544" lastFinishedPulling="2025-10-06 10:24:16.861425067 +0000 UTC m=+1566.225847928" observedRunningTime="2025-10-06 10:24:17.274719678 +0000 UTC m=+1566.639142549" watchObservedRunningTime="2025-10-06 10:24:17.280607367 +0000 UTC m=+1566.645030228" Oct 06 10:24:30 crc kubenswrapper[4824]: I1006 10:24:30.854553 4824 scope.go:117] "RemoveContainer" containerID="5aa36cdff7d909fd98c3a8a65be66bbf3f6a6b660ad12183d0da7f5a94b443b3" Oct 06 10:24:51 crc kubenswrapper[4824]: I1006 10:24:51.044405 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-f4pft"] Oct 06 10:24:51 crc kubenswrapper[4824]: I1006 10:24:51.053121 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-f4pft"] Oct 06 10:24:51 crc kubenswrapper[4824]: I1006 10:24:51.289305 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a55f5293-0c03-479d-a0c7-bb4869445794" path="/var/lib/kubelet/pods/a55f5293-0c03-479d-a0c7-bb4869445794/volumes" Oct 06 10:24:57 crc kubenswrapper[4824]: I1006 10:24:57.040166 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-5rvk5"] Oct 06 10:24:57 crc kubenswrapper[4824]: I1006 10:24:57.049432 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-5rvk5"] Oct 06 10:24:57 crc kubenswrapper[4824]: I1006 10:24:57.290369 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20c54a88-ce86-425a-b12c-cebd09083d14" path="/var/lib/kubelet/pods/20c54a88-ce86-425a-b12c-cebd09083d14/volumes" Oct 06 10:24:58 crc kubenswrapper[4824]: I1006 10:24:58.035465 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-t57w8"] Oct 06 10:24:58 crc kubenswrapper[4824]: I1006 10:24:58.044356 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-t57w8"] Oct 06 10:24:58 crc kubenswrapper[4824]: I1006 10:24:58.056699 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-xzlzd"] Oct 06 10:24:58 crc kubenswrapper[4824]: I1006 10:24:58.067006 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-nmf65"] Oct 06 10:24:58 crc kubenswrapper[4824]: I1006 10:24:58.089015 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-nmf65"] Oct 06 10:24:58 crc kubenswrapper[4824]: I1006 10:24:58.102264 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-xzlzd"] Oct 06 10:24:59 crc kubenswrapper[4824]: I1006 10:24:59.292957 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c9d76f42-5e79-4057-8823-f80071669be7" path="/var/lib/kubelet/pods/c9d76f42-5e79-4057-8823-f80071669be7/volumes" Oct 06 10:24:59 crc kubenswrapper[4824]: I1006 10:24:59.294238 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ceec1f0f-b38f-459c-99fe-2bf0660b8777" path="/var/lib/kubelet/pods/ceec1f0f-b38f-459c-99fe-2bf0660b8777/volumes" Oct 06 10:24:59 crc kubenswrapper[4824]: I1006 10:24:59.295424 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ddcbc15f-464c-44de-9eb3-d1d35d48bacb" path="/var/lib/kubelet/pods/ddcbc15f-464c-44de-9eb3-d1d35d48bacb/volumes" Oct 06 10:25:01 crc kubenswrapper[4824]: I1006 10:25:01.046683 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-bmgl4"] Oct 06 10:25:01 crc kubenswrapper[4824]: I1006 10:25:01.057257 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-bmgl4"] Oct 06 10:25:01 crc kubenswrapper[4824]: I1006 10:25:01.066432 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-4a6e-account-create-cx6mn"] Oct 06 10:25:01 crc kubenswrapper[4824]: I1006 10:25:01.075586 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-4a6e-account-create-cx6mn"] Oct 06 10:25:01 crc kubenswrapper[4824]: I1006 10:25:01.307560 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="256cff80-df3a-481d-a994-05b49d8996c1" path="/var/lib/kubelet/pods/256cff80-df3a-481d-a994-05b49d8996c1/volumes" Oct 06 10:25:01 crc kubenswrapper[4824]: I1006 10:25:01.309162 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7b66b5d9-1c7c-4f7d-aa8d-cba5b9f6c2f2" path="/var/lib/kubelet/pods/7b66b5d9-1c7c-4f7d-aa8d-cba5b9f6c2f2/volumes" Oct 06 10:25:06 crc kubenswrapper[4824]: I1006 10:25:06.042041 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-faf1-account-create-85qr7"] Oct 06 10:25:06 crc kubenswrapper[4824]: I1006 10:25:06.054508 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-faf1-account-create-85qr7"] Oct 06 10:25:07 crc kubenswrapper[4824]: I1006 10:25:07.290641 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c9571f84-f5be-4f46-b939-2ad4d5a42daf" path="/var/lib/kubelet/pods/c9571f84-f5be-4f46-b939-2ad4d5a42daf/volumes" Oct 06 10:25:08 crc kubenswrapper[4824]: I1006 10:25:08.049810 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-8782-account-create-86hgc"] Oct 06 10:25:08 crc kubenswrapper[4824]: I1006 10:25:08.064930 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-d68d-account-create-47bqf"] Oct 06 10:25:08 crc kubenswrapper[4824]: I1006 10:25:08.076801 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-8782-account-create-86hgc"] Oct 06 10:25:08 crc kubenswrapper[4824]: I1006 10:25:08.085732 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-d68d-account-create-47bqf"] Oct 06 10:25:09 crc kubenswrapper[4824]: I1006 10:25:09.042111 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-bbe4-account-create-ghj2p"] Oct 06 10:25:09 crc kubenswrapper[4824]: I1006 10:25:09.051908 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-bbe4-account-create-ghj2p"] Oct 06 10:25:09 crc kubenswrapper[4824]: I1006 10:25:09.289782 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="258a753c-080b-44aa-9636-bfb6f7195b23" path="/var/lib/kubelet/pods/258a753c-080b-44aa-9636-bfb6f7195b23/volumes" Oct 06 10:25:09 crc kubenswrapper[4824]: I1006 10:25:09.290726 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="34475172-27f5-47b0-824e-71eff9ed3570" path="/var/lib/kubelet/pods/34475172-27f5-47b0-824e-71eff9ed3570/volumes" Oct 06 10:25:09 crc kubenswrapper[4824]: I1006 10:25:09.291869 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4c9b54e8-db5a-4837-ab99-30f509b9d2fb" path="/var/lib/kubelet/pods/4c9b54e8-db5a-4837-ab99-30f509b9d2fb/volumes" Oct 06 10:25:12 crc kubenswrapper[4824]: I1006 10:25:12.040486 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-b964-account-create-h2khv"] Oct 06 10:25:12 crc kubenswrapper[4824]: I1006 10:25:12.056964 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-b964-account-create-h2khv"] Oct 06 10:25:13 crc kubenswrapper[4824]: I1006 10:25:13.292236 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c9fd14a4-e0f8-4905-88ad-77371475d9c2" path="/var/lib/kubelet/pods/c9fd14a4-e0f8-4905-88ad-77371475d9c2/volumes" Oct 06 10:25:13 crc kubenswrapper[4824]: I1006 10:25:13.915807 4824 patch_prober.go:28] interesting pod/machine-config-daemon-khgzw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 10:25:13 crc kubenswrapper[4824]: I1006 10:25:13.915910 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 10:25:30 crc kubenswrapper[4824]: I1006 10:25:30.940826 4824 scope.go:117] "RemoveContainer" containerID="10dd696c355b056274df6482c88b5dbbf7520a4852f8252ed2269ed130337ac5" Oct 06 10:25:30 crc kubenswrapper[4824]: I1006 10:25:30.990331 4824 scope.go:117] "RemoveContainer" containerID="87c81576adcaa23fb9324494c703d1c3f84bf4dfc82a548e6d873feb13063e88" Oct 06 10:25:31 crc kubenswrapper[4824]: I1006 10:25:31.055671 4824 scope.go:117] "RemoveContainer" containerID="03a50ad9606a91f866ae2e6342d1b63486e17701e34c1c5ab2d58988ccb2b1a1" Oct 06 10:25:31 crc kubenswrapper[4824]: I1006 10:25:31.110452 4824 scope.go:117] "RemoveContainer" containerID="f2bd5fad013050b478969598e1456a7593913eafa49996689b79b535d76a889d" Oct 06 10:25:31 crc kubenswrapper[4824]: I1006 10:25:31.183432 4824 scope.go:117] "RemoveContainer" containerID="a423d8ebda96fde12328f4eb430b9d6d71aa4c43d91b811bbab42c96f2fb8468" Oct 06 10:25:31 crc kubenswrapper[4824]: I1006 10:25:31.208676 4824 scope.go:117] "RemoveContainer" containerID="78b87860dc45f07561d049b019a8b61447f0972f95e18d7d73e1fb8a6b4d874f" Oct 06 10:25:31 crc kubenswrapper[4824]: I1006 10:25:31.258093 4824 scope.go:117] "RemoveContainer" containerID="aab8d3bb855e53d592920c6636a15e3d7292c18be7bd1a3f5740ad628def69bc" Oct 06 10:25:31 crc kubenswrapper[4824]: I1006 10:25:31.283917 4824 scope.go:117] "RemoveContainer" containerID="c20ec4c680dba96cfce21e0d3a4afec1174331e4d57303bda1d4da5b2de1ef46" Oct 06 10:25:31 crc kubenswrapper[4824]: I1006 10:25:31.307906 4824 scope.go:117] "RemoveContainer" containerID="e9aa345b92d318f37d64b759ec7893557cd190fb224b6cfbba3902cf664d657e" Oct 06 10:25:31 crc kubenswrapper[4824]: I1006 10:25:31.331437 4824 scope.go:117] "RemoveContainer" containerID="e05cd16913d32642ed8363d7ffe7f9912b69414d4757d773e592dfb555d63428" Oct 06 10:25:31 crc kubenswrapper[4824]: I1006 10:25:31.358729 4824 scope.go:117] "RemoveContainer" containerID="1b9e78c64608e0c16a1937508d6125f8c093ac04a3f603d41ea1089ecde42934" Oct 06 10:25:31 crc kubenswrapper[4824]: I1006 10:25:31.382201 4824 scope.go:117] "RemoveContainer" containerID="59c36f0cc6dade51bfb70bc93fe29975a6816585123e2e2b7180f65307c8bd9a" Oct 06 10:25:36 crc kubenswrapper[4824]: I1006 10:25:36.050264 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-59pkf"] Oct 06 10:25:36 crc kubenswrapper[4824]: I1006 10:25:36.062459 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-59pkf"] Oct 06 10:25:37 crc kubenswrapper[4824]: I1006 10:25:37.288262 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bc5d3bf-5bde-4ce8-9b2a-50498e628b86" path="/var/lib/kubelet/pods/1bc5d3bf-5bde-4ce8-9b2a-50498e628b86/volumes" Oct 06 10:25:43 crc kubenswrapper[4824]: I1006 10:25:43.916137 4824 patch_prober.go:28] interesting pod/machine-config-daemon-khgzw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 10:25:43 crc kubenswrapper[4824]: I1006 10:25:43.916744 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 10:25:45 crc kubenswrapper[4824]: I1006 10:25:45.035221 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-s7fhb"] Oct 06 10:25:45 crc kubenswrapper[4824]: I1006 10:25:45.044182 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-s7fhb"] Oct 06 10:25:45 crc kubenswrapper[4824]: I1006 10:25:45.289948 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e843b5e1-97a8-4a1c-8fa0-f3d8cf155ea0" path="/var/lib/kubelet/pods/e843b5e1-97a8-4a1c-8fa0-f3d8cf155ea0/volumes" Oct 06 10:25:47 crc kubenswrapper[4824]: I1006 10:25:47.301454 4824 generic.go:334] "Generic (PLEG): container finished" podID="8ea3d60c-80c7-4163-9f81-c0ec20e758c9" containerID="f16c6c16083a3e2b88f2fbcb3faa9ee204a9dd542a8d0dcb2955d3fa4855841b" exitCode=0 Oct 06 10:25:47 crc kubenswrapper[4824]: I1006 10:25:47.302281 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-kcvhh" event={"ID":"8ea3d60c-80c7-4163-9f81-c0ec20e758c9","Type":"ContainerDied","Data":"f16c6c16083a3e2b88f2fbcb3faa9ee204a9dd542a8d0dcb2955d3fa4855841b"} Oct 06 10:25:48 crc kubenswrapper[4824]: I1006 10:25:48.797216 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-kcvhh" Oct 06 10:25:48 crc kubenswrapper[4824]: I1006 10:25:48.929342 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8ea3d60c-80c7-4163-9f81-c0ec20e758c9-inventory\") pod \"8ea3d60c-80c7-4163-9f81-c0ec20e758c9\" (UID: \"8ea3d60c-80c7-4163-9f81-c0ec20e758c9\") " Oct 06 10:25:48 crc kubenswrapper[4824]: I1006 10:25:48.929421 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8ea3d60c-80c7-4163-9f81-c0ec20e758c9-ssh-key\") pod \"8ea3d60c-80c7-4163-9f81-c0ec20e758c9\" (UID: \"8ea3d60c-80c7-4163-9f81-c0ec20e758c9\") " Oct 06 10:25:48 crc kubenswrapper[4824]: I1006 10:25:48.929551 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vb6kf\" (UniqueName: \"kubernetes.io/projected/8ea3d60c-80c7-4163-9f81-c0ec20e758c9-kube-api-access-vb6kf\") pod \"8ea3d60c-80c7-4163-9f81-c0ec20e758c9\" (UID: \"8ea3d60c-80c7-4163-9f81-c0ec20e758c9\") " Oct 06 10:25:48 crc kubenswrapper[4824]: I1006 10:25:48.936010 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8ea3d60c-80c7-4163-9f81-c0ec20e758c9-kube-api-access-vb6kf" (OuterVolumeSpecName: "kube-api-access-vb6kf") pod "8ea3d60c-80c7-4163-9f81-c0ec20e758c9" (UID: "8ea3d60c-80c7-4163-9f81-c0ec20e758c9"). InnerVolumeSpecName "kube-api-access-vb6kf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:25:48 crc kubenswrapper[4824]: I1006 10:25:48.961313 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ea3d60c-80c7-4163-9f81-c0ec20e758c9-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "8ea3d60c-80c7-4163-9f81-c0ec20e758c9" (UID: "8ea3d60c-80c7-4163-9f81-c0ec20e758c9"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:25:48 crc kubenswrapper[4824]: I1006 10:25:48.964579 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ea3d60c-80c7-4163-9f81-c0ec20e758c9-inventory" (OuterVolumeSpecName: "inventory") pod "8ea3d60c-80c7-4163-9f81-c0ec20e758c9" (UID: "8ea3d60c-80c7-4163-9f81-c0ec20e758c9"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:25:49 crc kubenswrapper[4824]: I1006 10:25:49.032535 4824 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8ea3d60c-80c7-4163-9f81-c0ec20e758c9-inventory\") on node \"crc\" DevicePath \"\"" Oct 06 10:25:49 crc kubenswrapper[4824]: I1006 10:25:49.032586 4824 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8ea3d60c-80c7-4163-9f81-c0ec20e758c9-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 06 10:25:49 crc kubenswrapper[4824]: I1006 10:25:49.032603 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vb6kf\" (UniqueName: \"kubernetes.io/projected/8ea3d60c-80c7-4163-9f81-c0ec20e758c9-kube-api-access-vb6kf\") on node \"crc\" DevicePath \"\"" Oct 06 10:25:49 crc kubenswrapper[4824]: I1006 10:25:49.335764 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-kcvhh" event={"ID":"8ea3d60c-80c7-4163-9f81-c0ec20e758c9","Type":"ContainerDied","Data":"4d060b590ac350451bceaac2039d004b906d02efe77aeff328c1fe07659bc5a1"} Oct 06 10:25:49 crc kubenswrapper[4824]: I1006 10:25:49.335830 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4d060b590ac350451bceaac2039d004b906d02efe77aeff328c1fe07659bc5a1" Oct 06 10:25:49 crc kubenswrapper[4824]: I1006 10:25:49.335922 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-kcvhh" Oct 06 10:25:49 crc kubenswrapper[4824]: I1006 10:25:49.416772 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-xtlm8"] Oct 06 10:25:49 crc kubenswrapper[4824]: E1006 10:25:49.417401 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ea3d60c-80c7-4163-9f81-c0ec20e758c9" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Oct 06 10:25:49 crc kubenswrapper[4824]: I1006 10:25:49.417428 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ea3d60c-80c7-4163-9f81-c0ec20e758c9" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Oct 06 10:25:49 crc kubenswrapper[4824]: I1006 10:25:49.417714 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="8ea3d60c-80c7-4163-9f81-c0ec20e758c9" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Oct 06 10:25:49 crc kubenswrapper[4824]: I1006 10:25:49.418739 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-xtlm8" Oct 06 10:25:49 crc kubenswrapper[4824]: I1006 10:25:49.423753 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 06 10:25:49 crc kubenswrapper[4824]: I1006 10:25:49.424017 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 06 10:25:49 crc kubenswrapper[4824]: I1006 10:25:49.424200 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-f886w" Oct 06 10:25:49 crc kubenswrapper[4824]: I1006 10:25:49.424373 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 06 10:25:49 crc kubenswrapper[4824]: I1006 10:25:49.435025 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-xtlm8"] Oct 06 10:25:49 crc kubenswrapper[4824]: I1006 10:25:49.545970 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1961267a-e5c0-469e-9f0e-9d4edbc8e64c-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-xtlm8\" (UID: \"1961267a-e5c0-469e-9f0e-9d4edbc8e64c\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-xtlm8" Oct 06 10:25:49 crc kubenswrapper[4824]: I1006 10:25:49.546315 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1961267a-e5c0-469e-9f0e-9d4edbc8e64c-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-xtlm8\" (UID: \"1961267a-e5c0-469e-9f0e-9d4edbc8e64c\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-xtlm8" Oct 06 10:25:49 crc kubenswrapper[4824]: I1006 10:25:49.546519 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tw6j9\" (UniqueName: \"kubernetes.io/projected/1961267a-e5c0-469e-9f0e-9d4edbc8e64c-kube-api-access-tw6j9\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-xtlm8\" (UID: \"1961267a-e5c0-469e-9f0e-9d4edbc8e64c\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-xtlm8" Oct 06 10:25:49 crc kubenswrapper[4824]: I1006 10:25:49.649027 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1961267a-e5c0-469e-9f0e-9d4edbc8e64c-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-xtlm8\" (UID: \"1961267a-e5c0-469e-9f0e-9d4edbc8e64c\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-xtlm8" Oct 06 10:25:49 crc kubenswrapper[4824]: I1006 10:25:49.649375 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1961267a-e5c0-469e-9f0e-9d4edbc8e64c-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-xtlm8\" (UID: \"1961267a-e5c0-469e-9f0e-9d4edbc8e64c\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-xtlm8" Oct 06 10:25:49 crc kubenswrapper[4824]: I1006 10:25:49.649571 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tw6j9\" (UniqueName: \"kubernetes.io/projected/1961267a-e5c0-469e-9f0e-9d4edbc8e64c-kube-api-access-tw6j9\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-xtlm8\" (UID: \"1961267a-e5c0-469e-9f0e-9d4edbc8e64c\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-xtlm8" Oct 06 10:25:49 crc kubenswrapper[4824]: I1006 10:25:49.656712 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1961267a-e5c0-469e-9f0e-9d4edbc8e64c-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-xtlm8\" (UID: \"1961267a-e5c0-469e-9f0e-9d4edbc8e64c\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-xtlm8" Oct 06 10:25:49 crc kubenswrapper[4824]: I1006 10:25:49.657665 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1961267a-e5c0-469e-9f0e-9d4edbc8e64c-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-xtlm8\" (UID: \"1961267a-e5c0-469e-9f0e-9d4edbc8e64c\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-xtlm8" Oct 06 10:25:49 crc kubenswrapper[4824]: I1006 10:25:49.685618 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tw6j9\" (UniqueName: \"kubernetes.io/projected/1961267a-e5c0-469e-9f0e-9d4edbc8e64c-kube-api-access-tw6j9\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-xtlm8\" (UID: \"1961267a-e5c0-469e-9f0e-9d4edbc8e64c\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-xtlm8" Oct 06 10:25:49 crc kubenswrapper[4824]: I1006 10:25:49.737088 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-xtlm8" Oct 06 10:25:50 crc kubenswrapper[4824]: I1006 10:25:50.364562 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-xtlm8"] Oct 06 10:25:51 crc kubenswrapper[4824]: I1006 10:25:51.361446 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-xtlm8" event={"ID":"1961267a-e5c0-469e-9f0e-9d4edbc8e64c","Type":"ContainerStarted","Data":"83ff57ec9460bcda2585092553e0380b45ef405854ac0a7cd3997af2b8c76a24"} Oct 06 10:25:51 crc kubenswrapper[4824]: I1006 10:25:51.361643 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-xtlm8" event={"ID":"1961267a-e5c0-469e-9f0e-9d4edbc8e64c","Type":"ContainerStarted","Data":"4ec7ef63f36858b07b10bf60579e7cdb2abad46cbe36c16bc12dd71b807087c7"} Oct 06 10:25:51 crc kubenswrapper[4824]: I1006 10:25:51.385806 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-xtlm8" podStartSLOduration=1.944725933 podStartE2EDuration="2.385780035s" podCreationTimestamp="2025-10-06 10:25:49 +0000 UTC" firstStartedPulling="2025-10-06 10:25:50.364645075 +0000 UTC m=+1659.729067936" lastFinishedPulling="2025-10-06 10:25:50.805699177 +0000 UTC m=+1660.170122038" observedRunningTime="2025-10-06 10:25:51.376080552 +0000 UTC m=+1660.740503443" watchObservedRunningTime="2025-10-06 10:25:51.385780035 +0000 UTC m=+1660.750202906" Oct 06 10:25:56 crc kubenswrapper[4824]: I1006 10:25:56.977453 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-qnxgg"] Oct 06 10:25:56 crc kubenswrapper[4824]: I1006 10:25:56.981223 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qnxgg" Oct 06 10:25:57 crc kubenswrapper[4824]: I1006 10:25:57.015273 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qnxgg"] Oct 06 10:25:57 crc kubenswrapper[4824]: I1006 10:25:57.123310 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4cd0302f-df74-4748-bec8-a8b9198b8dfd-catalog-content\") pod \"redhat-operators-qnxgg\" (UID: \"4cd0302f-df74-4748-bec8-a8b9198b8dfd\") " pod="openshift-marketplace/redhat-operators-qnxgg" Oct 06 10:25:57 crc kubenswrapper[4824]: I1006 10:25:57.123427 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4cd0302f-df74-4748-bec8-a8b9198b8dfd-utilities\") pod \"redhat-operators-qnxgg\" (UID: \"4cd0302f-df74-4748-bec8-a8b9198b8dfd\") " pod="openshift-marketplace/redhat-operators-qnxgg" Oct 06 10:25:57 crc kubenswrapper[4824]: I1006 10:25:57.123462 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vmm5f\" (UniqueName: \"kubernetes.io/projected/4cd0302f-df74-4748-bec8-a8b9198b8dfd-kube-api-access-vmm5f\") pod \"redhat-operators-qnxgg\" (UID: \"4cd0302f-df74-4748-bec8-a8b9198b8dfd\") " pod="openshift-marketplace/redhat-operators-qnxgg" Oct 06 10:25:57 crc kubenswrapper[4824]: I1006 10:25:57.225675 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4cd0302f-df74-4748-bec8-a8b9198b8dfd-catalog-content\") pod \"redhat-operators-qnxgg\" (UID: \"4cd0302f-df74-4748-bec8-a8b9198b8dfd\") " pod="openshift-marketplace/redhat-operators-qnxgg" Oct 06 10:25:57 crc kubenswrapper[4824]: I1006 10:25:57.226108 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4cd0302f-df74-4748-bec8-a8b9198b8dfd-utilities\") pod \"redhat-operators-qnxgg\" (UID: \"4cd0302f-df74-4748-bec8-a8b9198b8dfd\") " pod="openshift-marketplace/redhat-operators-qnxgg" Oct 06 10:25:57 crc kubenswrapper[4824]: I1006 10:25:57.226277 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vmm5f\" (UniqueName: \"kubernetes.io/projected/4cd0302f-df74-4748-bec8-a8b9198b8dfd-kube-api-access-vmm5f\") pod \"redhat-operators-qnxgg\" (UID: \"4cd0302f-df74-4748-bec8-a8b9198b8dfd\") " pod="openshift-marketplace/redhat-operators-qnxgg" Oct 06 10:25:57 crc kubenswrapper[4824]: I1006 10:25:57.226283 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4cd0302f-df74-4748-bec8-a8b9198b8dfd-catalog-content\") pod \"redhat-operators-qnxgg\" (UID: \"4cd0302f-df74-4748-bec8-a8b9198b8dfd\") " pod="openshift-marketplace/redhat-operators-qnxgg" Oct 06 10:25:57 crc kubenswrapper[4824]: I1006 10:25:57.226772 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4cd0302f-df74-4748-bec8-a8b9198b8dfd-utilities\") pod \"redhat-operators-qnxgg\" (UID: \"4cd0302f-df74-4748-bec8-a8b9198b8dfd\") " pod="openshift-marketplace/redhat-operators-qnxgg" Oct 06 10:25:57 crc kubenswrapper[4824]: I1006 10:25:57.260062 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vmm5f\" (UniqueName: \"kubernetes.io/projected/4cd0302f-df74-4748-bec8-a8b9198b8dfd-kube-api-access-vmm5f\") pod \"redhat-operators-qnxgg\" (UID: \"4cd0302f-df74-4748-bec8-a8b9198b8dfd\") " pod="openshift-marketplace/redhat-operators-qnxgg" Oct 06 10:25:57 crc kubenswrapper[4824]: I1006 10:25:57.312683 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qnxgg" Oct 06 10:25:57 crc kubenswrapper[4824]: I1006 10:25:57.839589 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qnxgg"] Oct 06 10:25:58 crc kubenswrapper[4824]: I1006 10:25:58.443762 4824 generic.go:334] "Generic (PLEG): container finished" podID="4cd0302f-df74-4748-bec8-a8b9198b8dfd" containerID="eebb13f2f40afe6b7e2d79bcf2717e3b1849c92244e7e72562cc24b2eec67a01" exitCode=0 Oct 06 10:25:58 crc kubenswrapper[4824]: I1006 10:25:58.443884 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qnxgg" event={"ID":"4cd0302f-df74-4748-bec8-a8b9198b8dfd","Type":"ContainerDied","Data":"eebb13f2f40afe6b7e2d79bcf2717e3b1849c92244e7e72562cc24b2eec67a01"} Oct 06 10:25:58 crc kubenswrapper[4824]: I1006 10:25:58.444258 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qnxgg" event={"ID":"4cd0302f-df74-4748-bec8-a8b9198b8dfd","Type":"ContainerStarted","Data":"b9e645be5ef5718c82d7f998813307063f3aa6acb08a9bd7a81723153c0998ce"} Oct 06 10:25:59 crc kubenswrapper[4824]: I1006 10:25:59.459045 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qnxgg" event={"ID":"4cd0302f-df74-4748-bec8-a8b9198b8dfd","Type":"ContainerStarted","Data":"cc8ffb04c36bccfbde6df3d48652a81cffa2c90621099bd663a19642f4f624e3"} Oct 06 10:26:00 crc kubenswrapper[4824]: I1006 10:26:00.473297 4824 generic.go:334] "Generic (PLEG): container finished" podID="4cd0302f-df74-4748-bec8-a8b9198b8dfd" containerID="cc8ffb04c36bccfbde6df3d48652a81cffa2c90621099bd663a19642f4f624e3" exitCode=0 Oct 06 10:26:00 crc kubenswrapper[4824]: I1006 10:26:00.473458 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qnxgg" event={"ID":"4cd0302f-df74-4748-bec8-a8b9198b8dfd","Type":"ContainerDied","Data":"cc8ffb04c36bccfbde6df3d48652a81cffa2c90621099bd663a19642f4f624e3"} Oct 06 10:26:01 crc kubenswrapper[4824]: I1006 10:26:01.487294 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qnxgg" event={"ID":"4cd0302f-df74-4748-bec8-a8b9198b8dfd","Type":"ContainerStarted","Data":"506fbf40c29559121b22a74be6867129a4f1694c95a293f3e69d108af5fec089"} Oct 06 10:26:01 crc kubenswrapper[4824]: I1006 10:26:01.511697 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-qnxgg" podStartSLOduration=2.767207934 podStartE2EDuration="5.511676016s" podCreationTimestamp="2025-10-06 10:25:56 +0000 UTC" firstStartedPulling="2025-10-06 10:25:58.445914336 +0000 UTC m=+1667.810337197" lastFinishedPulling="2025-10-06 10:26:01.190382408 +0000 UTC m=+1670.554805279" observedRunningTime="2025-10-06 10:26:01.507629734 +0000 UTC m=+1670.872052585" watchObservedRunningTime="2025-10-06 10:26:01.511676016 +0000 UTC m=+1670.876098877" Oct 06 10:26:07 crc kubenswrapper[4824]: I1006 10:26:07.313107 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-qnxgg" Oct 06 10:26:07 crc kubenswrapper[4824]: I1006 10:26:07.313603 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-qnxgg" Oct 06 10:26:07 crc kubenswrapper[4824]: I1006 10:26:07.378793 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-qnxgg" Oct 06 10:26:07 crc kubenswrapper[4824]: I1006 10:26:07.593607 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-qnxgg" Oct 06 10:26:07 crc kubenswrapper[4824]: I1006 10:26:07.650606 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-qnxgg"] Oct 06 10:26:09 crc kubenswrapper[4824]: I1006 10:26:09.565277 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-qnxgg" podUID="4cd0302f-df74-4748-bec8-a8b9198b8dfd" containerName="registry-server" containerID="cri-o://506fbf40c29559121b22a74be6867129a4f1694c95a293f3e69d108af5fec089" gracePeriod=2 Oct 06 10:26:09 crc kubenswrapper[4824]: I1006 10:26:09.996505 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qnxgg" Oct 06 10:26:10 crc kubenswrapper[4824]: I1006 10:26:10.160009 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4cd0302f-df74-4748-bec8-a8b9198b8dfd-utilities\") pod \"4cd0302f-df74-4748-bec8-a8b9198b8dfd\" (UID: \"4cd0302f-df74-4748-bec8-a8b9198b8dfd\") " Oct 06 10:26:10 crc kubenswrapper[4824]: I1006 10:26:10.160189 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4cd0302f-df74-4748-bec8-a8b9198b8dfd-catalog-content\") pod \"4cd0302f-df74-4748-bec8-a8b9198b8dfd\" (UID: \"4cd0302f-df74-4748-bec8-a8b9198b8dfd\") " Oct 06 10:26:10 crc kubenswrapper[4824]: I1006 10:26:10.160250 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vmm5f\" (UniqueName: \"kubernetes.io/projected/4cd0302f-df74-4748-bec8-a8b9198b8dfd-kube-api-access-vmm5f\") pod \"4cd0302f-df74-4748-bec8-a8b9198b8dfd\" (UID: \"4cd0302f-df74-4748-bec8-a8b9198b8dfd\") " Oct 06 10:26:10 crc kubenswrapper[4824]: I1006 10:26:10.161728 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4cd0302f-df74-4748-bec8-a8b9198b8dfd-utilities" (OuterVolumeSpecName: "utilities") pod "4cd0302f-df74-4748-bec8-a8b9198b8dfd" (UID: "4cd0302f-df74-4748-bec8-a8b9198b8dfd"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:26:10 crc kubenswrapper[4824]: I1006 10:26:10.167729 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4cd0302f-df74-4748-bec8-a8b9198b8dfd-kube-api-access-vmm5f" (OuterVolumeSpecName: "kube-api-access-vmm5f") pod "4cd0302f-df74-4748-bec8-a8b9198b8dfd" (UID: "4cd0302f-df74-4748-bec8-a8b9198b8dfd"). InnerVolumeSpecName "kube-api-access-vmm5f". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:26:10 crc kubenswrapper[4824]: I1006 10:26:10.257990 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4cd0302f-df74-4748-bec8-a8b9198b8dfd-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4cd0302f-df74-4748-bec8-a8b9198b8dfd" (UID: "4cd0302f-df74-4748-bec8-a8b9198b8dfd"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:26:10 crc kubenswrapper[4824]: I1006 10:26:10.263115 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4cd0302f-df74-4748-bec8-a8b9198b8dfd-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 06 10:26:10 crc kubenswrapper[4824]: I1006 10:26:10.263170 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vmm5f\" (UniqueName: \"kubernetes.io/projected/4cd0302f-df74-4748-bec8-a8b9198b8dfd-kube-api-access-vmm5f\") on node \"crc\" DevicePath \"\"" Oct 06 10:26:10 crc kubenswrapper[4824]: I1006 10:26:10.263186 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4cd0302f-df74-4748-bec8-a8b9198b8dfd-utilities\") on node \"crc\" DevicePath \"\"" Oct 06 10:26:10 crc kubenswrapper[4824]: I1006 10:26:10.578800 4824 generic.go:334] "Generic (PLEG): container finished" podID="4cd0302f-df74-4748-bec8-a8b9198b8dfd" containerID="506fbf40c29559121b22a74be6867129a4f1694c95a293f3e69d108af5fec089" exitCode=0 Oct 06 10:26:10 crc kubenswrapper[4824]: I1006 10:26:10.578870 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qnxgg" Oct 06 10:26:10 crc kubenswrapper[4824]: I1006 10:26:10.578869 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qnxgg" event={"ID":"4cd0302f-df74-4748-bec8-a8b9198b8dfd","Type":"ContainerDied","Data":"506fbf40c29559121b22a74be6867129a4f1694c95a293f3e69d108af5fec089"} Oct 06 10:26:10 crc kubenswrapper[4824]: I1006 10:26:10.579032 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qnxgg" event={"ID":"4cd0302f-df74-4748-bec8-a8b9198b8dfd","Type":"ContainerDied","Data":"b9e645be5ef5718c82d7f998813307063f3aa6acb08a9bd7a81723153c0998ce"} Oct 06 10:26:10 crc kubenswrapper[4824]: I1006 10:26:10.579066 4824 scope.go:117] "RemoveContainer" containerID="506fbf40c29559121b22a74be6867129a4f1694c95a293f3e69d108af5fec089" Oct 06 10:26:10 crc kubenswrapper[4824]: I1006 10:26:10.619047 4824 scope.go:117] "RemoveContainer" containerID="cc8ffb04c36bccfbde6df3d48652a81cffa2c90621099bd663a19642f4f624e3" Oct 06 10:26:10 crc kubenswrapper[4824]: I1006 10:26:10.633261 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-qnxgg"] Oct 06 10:26:10 crc kubenswrapper[4824]: I1006 10:26:10.653897 4824 scope.go:117] "RemoveContainer" containerID="eebb13f2f40afe6b7e2d79bcf2717e3b1849c92244e7e72562cc24b2eec67a01" Oct 06 10:26:10 crc kubenswrapper[4824]: I1006 10:26:10.657542 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-qnxgg"] Oct 06 10:26:10 crc kubenswrapper[4824]: I1006 10:26:10.693071 4824 scope.go:117] "RemoveContainer" containerID="506fbf40c29559121b22a74be6867129a4f1694c95a293f3e69d108af5fec089" Oct 06 10:26:10 crc kubenswrapper[4824]: E1006 10:26:10.693876 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"506fbf40c29559121b22a74be6867129a4f1694c95a293f3e69d108af5fec089\": container with ID starting with 506fbf40c29559121b22a74be6867129a4f1694c95a293f3e69d108af5fec089 not found: ID does not exist" containerID="506fbf40c29559121b22a74be6867129a4f1694c95a293f3e69d108af5fec089" Oct 06 10:26:10 crc kubenswrapper[4824]: I1006 10:26:10.693937 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"506fbf40c29559121b22a74be6867129a4f1694c95a293f3e69d108af5fec089"} err="failed to get container status \"506fbf40c29559121b22a74be6867129a4f1694c95a293f3e69d108af5fec089\": rpc error: code = NotFound desc = could not find container \"506fbf40c29559121b22a74be6867129a4f1694c95a293f3e69d108af5fec089\": container with ID starting with 506fbf40c29559121b22a74be6867129a4f1694c95a293f3e69d108af5fec089 not found: ID does not exist" Oct 06 10:26:10 crc kubenswrapper[4824]: I1006 10:26:10.694254 4824 scope.go:117] "RemoveContainer" containerID="cc8ffb04c36bccfbde6df3d48652a81cffa2c90621099bd663a19642f4f624e3" Oct 06 10:26:10 crc kubenswrapper[4824]: E1006 10:26:10.695023 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cc8ffb04c36bccfbde6df3d48652a81cffa2c90621099bd663a19642f4f624e3\": container with ID starting with cc8ffb04c36bccfbde6df3d48652a81cffa2c90621099bd663a19642f4f624e3 not found: ID does not exist" containerID="cc8ffb04c36bccfbde6df3d48652a81cffa2c90621099bd663a19642f4f624e3" Oct 06 10:26:10 crc kubenswrapper[4824]: I1006 10:26:10.695082 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cc8ffb04c36bccfbde6df3d48652a81cffa2c90621099bd663a19642f4f624e3"} err="failed to get container status \"cc8ffb04c36bccfbde6df3d48652a81cffa2c90621099bd663a19642f4f624e3\": rpc error: code = NotFound desc = could not find container \"cc8ffb04c36bccfbde6df3d48652a81cffa2c90621099bd663a19642f4f624e3\": container with ID starting with cc8ffb04c36bccfbde6df3d48652a81cffa2c90621099bd663a19642f4f624e3 not found: ID does not exist" Oct 06 10:26:10 crc kubenswrapper[4824]: I1006 10:26:10.695122 4824 scope.go:117] "RemoveContainer" containerID="eebb13f2f40afe6b7e2d79bcf2717e3b1849c92244e7e72562cc24b2eec67a01" Oct 06 10:26:10 crc kubenswrapper[4824]: E1006 10:26:10.695587 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eebb13f2f40afe6b7e2d79bcf2717e3b1849c92244e7e72562cc24b2eec67a01\": container with ID starting with eebb13f2f40afe6b7e2d79bcf2717e3b1849c92244e7e72562cc24b2eec67a01 not found: ID does not exist" containerID="eebb13f2f40afe6b7e2d79bcf2717e3b1849c92244e7e72562cc24b2eec67a01" Oct 06 10:26:10 crc kubenswrapper[4824]: I1006 10:26:10.695619 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eebb13f2f40afe6b7e2d79bcf2717e3b1849c92244e7e72562cc24b2eec67a01"} err="failed to get container status \"eebb13f2f40afe6b7e2d79bcf2717e3b1849c92244e7e72562cc24b2eec67a01\": rpc error: code = NotFound desc = could not find container \"eebb13f2f40afe6b7e2d79bcf2717e3b1849c92244e7e72562cc24b2eec67a01\": container with ID starting with eebb13f2f40afe6b7e2d79bcf2717e3b1849c92244e7e72562cc24b2eec67a01 not found: ID does not exist" Oct 06 10:26:11 crc kubenswrapper[4824]: I1006 10:26:11.309229 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4cd0302f-df74-4748-bec8-a8b9198b8dfd" path="/var/lib/kubelet/pods/4cd0302f-df74-4748-bec8-a8b9198b8dfd/volumes" Oct 06 10:26:13 crc kubenswrapper[4824]: I1006 10:26:13.915719 4824 patch_prober.go:28] interesting pod/machine-config-daemon-khgzw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 10:26:13 crc kubenswrapper[4824]: I1006 10:26:13.916586 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 10:26:13 crc kubenswrapper[4824]: I1006 10:26:13.916671 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" Oct 06 10:26:13 crc kubenswrapper[4824]: I1006 10:26:13.918286 4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"9de43c7f48ed9ad071a24adf937829ab92276bd505bb72d5b89bac4922aac3b0"} pod="openshift-machine-config-operator/machine-config-daemon-khgzw" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 06 10:26:13 crc kubenswrapper[4824]: I1006 10:26:13.918446 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" containerName="machine-config-daemon" containerID="cri-o://9de43c7f48ed9ad071a24adf937829ab92276bd505bb72d5b89bac4922aac3b0" gracePeriod=600 Oct 06 10:26:14 crc kubenswrapper[4824]: E1006 10:26:14.043559 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-khgzw_openshift-machine-config-operator(e1140c8c-93f1-4fce-8c06-a69561ff8a6d)\"" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" Oct 06 10:26:14 crc kubenswrapper[4824]: I1006 10:26:14.620469 4824 generic.go:334] "Generic (PLEG): container finished" podID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" containerID="9de43c7f48ed9ad071a24adf937829ab92276bd505bb72d5b89bac4922aac3b0" exitCode=0 Oct 06 10:26:14 crc kubenswrapper[4824]: I1006 10:26:14.620537 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" event={"ID":"e1140c8c-93f1-4fce-8c06-a69561ff8a6d","Type":"ContainerDied","Data":"9de43c7f48ed9ad071a24adf937829ab92276bd505bb72d5b89bac4922aac3b0"} Oct 06 10:26:14 crc kubenswrapper[4824]: I1006 10:26:14.620587 4824 scope.go:117] "RemoveContainer" containerID="7f680663f7fe4c1e2c1c711b5dc3a94dfad5110e11cc00fd374e4d6a79007f09" Oct 06 10:26:14 crc kubenswrapper[4824]: I1006 10:26:14.621640 4824 scope.go:117] "RemoveContainer" containerID="9de43c7f48ed9ad071a24adf937829ab92276bd505bb72d5b89bac4922aac3b0" Oct 06 10:26:14 crc kubenswrapper[4824]: E1006 10:26:14.623947 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-khgzw_openshift-machine-config-operator(e1140c8c-93f1-4fce-8c06-a69561ff8a6d)\"" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" Oct 06 10:26:17 crc kubenswrapper[4824]: I1006 10:26:17.049224 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-xsgsg"] Oct 06 10:26:17 crc kubenswrapper[4824]: I1006 10:26:17.061038 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-xsgsg"] Oct 06 10:26:17 crc kubenswrapper[4824]: I1006 10:26:17.286414 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3d78bdee-7130-48a9-ad24-6c8c6525784c" path="/var/lib/kubelet/pods/3d78bdee-7130-48a9-ad24-6c8c6525784c/volumes" Oct 06 10:26:29 crc kubenswrapper[4824]: I1006 10:26:29.274622 4824 scope.go:117] "RemoveContainer" containerID="9de43c7f48ed9ad071a24adf937829ab92276bd505bb72d5b89bac4922aac3b0" Oct 06 10:26:29 crc kubenswrapper[4824]: E1006 10:26:29.276159 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-khgzw_openshift-machine-config-operator(e1140c8c-93f1-4fce-8c06-a69561ff8a6d)\"" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" Oct 06 10:26:30 crc kubenswrapper[4824]: I1006 10:26:30.071354 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-wlmn6"] Oct 06 10:26:30 crc kubenswrapper[4824]: I1006 10:26:30.086412 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-fzvv8"] Oct 06 10:26:30 crc kubenswrapper[4824]: I1006 10:26:30.097014 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-wlmn6"] Oct 06 10:26:30 crc kubenswrapper[4824]: I1006 10:26:30.106273 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-fzvv8"] Oct 06 10:26:31 crc kubenswrapper[4824]: I1006 10:26:31.312154 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="554a3f37-c7c8-4d09-9da7-df5319dccecd" path="/var/lib/kubelet/pods/554a3f37-c7c8-4d09-9da7-df5319dccecd/volumes" Oct 06 10:26:31 crc kubenswrapper[4824]: I1006 10:26:31.313045 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="73fb825b-ea3b-43c5-b143-e70310f8c1fd" path="/var/lib/kubelet/pods/73fb825b-ea3b-43c5-b143-e70310f8c1fd/volumes" Oct 06 10:26:31 crc kubenswrapper[4824]: I1006 10:26:31.639875 4824 scope.go:117] "RemoveContainer" containerID="a787dec7530bf4c4cbfc1e0888aebec0261071050df377f25ba2aebb9d060923" Oct 06 10:26:31 crc kubenswrapper[4824]: I1006 10:26:31.679775 4824 scope.go:117] "RemoveContainer" containerID="f23c07b390abda237c8434ff9ade9da85768a81dd5fc716e96733d119b669bce" Oct 06 10:26:31 crc kubenswrapper[4824]: I1006 10:26:31.733652 4824 scope.go:117] "RemoveContainer" containerID="86eeda4af9d4a8081ccdcc7a0cceaeae21a27abc73451b614b28ba15c64ed645" Oct 06 10:26:31 crc kubenswrapper[4824]: I1006 10:26:31.786755 4824 scope.go:117] "RemoveContainer" containerID="90ca09e2651e49075cef289ad0aa7e54e3882d823344ef1a9a6a28355a58ff56" Oct 06 10:26:31 crc kubenswrapper[4824]: I1006 10:26:31.831351 4824 scope.go:117] "RemoveContainer" containerID="0bb94fc42b32d2a0dfc8802382f088d25416230c9e728cf1be47d5229e9d1ebe" Oct 06 10:26:37 crc kubenswrapper[4824]: I1006 10:26:37.029216 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-ljqzg"] Oct 06 10:26:37 crc kubenswrapper[4824]: I1006 10:26:37.037833 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-ljqzg"] Oct 06 10:26:37 crc kubenswrapper[4824]: I1006 10:26:37.290735 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dfc3bcd2-9baa-4a3a-8588-295e692d8e3e" path="/var/lib/kubelet/pods/dfc3bcd2-9baa-4a3a-8588-295e692d8e3e/volumes" Oct 06 10:26:42 crc kubenswrapper[4824]: I1006 10:26:42.033088 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-ldn9v"] Oct 06 10:26:42 crc kubenswrapper[4824]: I1006 10:26:42.042972 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-ldn9v"] Oct 06 10:26:43 crc kubenswrapper[4824]: I1006 10:26:43.293956 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2382b80c-75fc-435f-be5f-0f1eb6a5cd17" path="/var/lib/kubelet/pods/2382b80c-75fc-435f-be5f-0f1eb6a5cd17/volumes" Oct 06 10:26:44 crc kubenswrapper[4824]: I1006 10:26:44.275550 4824 scope.go:117] "RemoveContainer" containerID="9de43c7f48ed9ad071a24adf937829ab92276bd505bb72d5b89bac4922aac3b0" Oct 06 10:26:44 crc kubenswrapper[4824]: E1006 10:26:44.275920 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-khgzw_openshift-machine-config-operator(e1140c8c-93f1-4fce-8c06-a69561ff8a6d)\"" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" Oct 06 10:26:57 crc kubenswrapper[4824]: I1006 10:26:57.276883 4824 scope.go:117] "RemoveContainer" containerID="9de43c7f48ed9ad071a24adf937829ab92276bd505bb72d5b89bac4922aac3b0" Oct 06 10:26:57 crc kubenswrapper[4824]: E1006 10:26:57.277950 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-khgzw_openshift-machine-config-operator(e1140c8c-93f1-4fce-8c06-a69561ff8a6d)\"" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" Oct 06 10:27:02 crc kubenswrapper[4824]: I1006 10:27:02.168536 4824 generic.go:334] "Generic (PLEG): container finished" podID="1961267a-e5c0-469e-9f0e-9d4edbc8e64c" containerID="83ff57ec9460bcda2585092553e0380b45ef405854ac0a7cd3997af2b8c76a24" exitCode=0 Oct 06 10:27:02 crc kubenswrapper[4824]: I1006 10:27:02.168635 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-xtlm8" event={"ID":"1961267a-e5c0-469e-9f0e-9d4edbc8e64c","Type":"ContainerDied","Data":"83ff57ec9460bcda2585092553e0380b45ef405854ac0a7cd3997af2b8c76a24"} Oct 06 10:27:03 crc kubenswrapper[4824]: I1006 10:27:03.607621 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-xtlm8" Oct 06 10:27:03 crc kubenswrapper[4824]: I1006 10:27:03.719052 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tw6j9\" (UniqueName: \"kubernetes.io/projected/1961267a-e5c0-469e-9f0e-9d4edbc8e64c-kube-api-access-tw6j9\") pod \"1961267a-e5c0-469e-9f0e-9d4edbc8e64c\" (UID: \"1961267a-e5c0-469e-9f0e-9d4edbc8e64c\") " Oct 06 10:27:03 crc kubenswrapper[4824]: I1006 10:27:03.719288 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1961267a-e5c0-469e-9f0e-9d4edbc8e64c-ssh-key\") pod \"1961267a-e5c0-469e-9f0e-9d4edbc8e64c\" (UID: \"1961267a-e5c0-469e-9f0e-9d4edbc8e64c\") " Oct 06 10:27:03 crc kubenswrapper[4824]: I1006 10:27:03.719356 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1961267a-e5c0-469e-9f0e-9d4edbc8e64c-inventory\") pod \"1961267a-e5c0-469e-9f0e-9d4edbc8e64c\" (UID: \"1961267a-e5c0-469e-9f0e-9d4edbc8e64c\") " Oct 06 10:27:03 crc kubenswrapper[4824]: I1006 10:27:03.729084 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1961267a-e5c0-469e-9f0e-9d4edbc8e64c-kube-api-access-tw6j9" (OuterVolumeSpecName: "kube-api-access-tw6j9") pod "1961267a-e5c0-469e-9f0e-9d4edbc8e64c" (UID: "1961267a-e5c0-469e-9f0e-9d4edbc8e64c"). InnerVolumeSpecName "kube-api-access-tw6j9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:27:03 crc kubenswrapper[4824]: I1006 10:27:03.766349 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1961267a-e5c0-469e-9f0e-9d4edbc8e64c-inventory" (OuterVolumeSpecName: "inventory") pod "1961267a-e5c0-469e-9f0e-9d4edbc8e64c" (UID: "1961267a-e5c0-469e-9f0e-9d4edbc8e64c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:27:03 crc kubenswrapper[4824]: I1006 10:27:03.775108 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1961267a-e5c0-469e-9f0e-9d4edbc8e64c-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "1961267a-e5c0-469e-9f0e-9d4edbc8e64c" (UID: "1961267a-e5c0-469e-9f0e-9d4edbc8e64c"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:27:03 crc kubenswrapper[4824]: I1006 10:27:03.822047 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tw6j9\" (UniqueName: \"kubernetes.io/projected/1961267a-e5c0-469e-9f0e-9d4edbc8e64c-kube-api-access-tw6j9\") on node \"crc\" DevicePath \"\"" Oct 06 10:27:03 crc kubenswrapper[4824]: I1006 10:27:03.822086 4824 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1961267a-e5c0-469e-9f0e-9d4edbc8e64c-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 06 10:27:03 crc kubenswrapper[4824]: I1006 10:27:03.822097 4824 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1961267a-e5c0-469e-9f0e-9d4edbc8e64c-inventory\") on node \"crc\" DevicePath \"\"" Oct 06 10:27:04 crc kubenswrapper[4824]: I1006 10:27:04.201704 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-xtlm8" Oct 06 10:27:04 crc kubenswrapper[4824]: I1006 10:27:04.201638 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-xtlm8" event={"ID":"1961267a-e5c0-469e-9f0e-9d4edbc8e64c","Type":"ContainerDied","Data":"4ec7ef63f36858b07b10bf60579e7cdb2abad46cbe36c16bc12dd71b807087c7"} Oct 06 10:27:04 crc kubenswrapper[4824]: I1006 10:27:04.209358 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4ec7ef63f36858b07b10bf60579e7cdb2abad46cbe36c16bc12dd71b807087c7" Oct 06 10:27:04 crc kubenswrapper[4824]: I1006 10:27:04.325070 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-jm56f"] Oct 06 10:27:04 crc kubenswrapper[4824]: E1006 10:27:04.325742 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4cd0302f-df74-4748-bec8-a8b9198b8dfd" containerName="extract-utilities" Oct 06 10:27:04 crc kubenswrapper[4824]: I1006 10:27:04.325778 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="4cd0302f-df74-4748-bec8-a8b9198b8dfd" containerName="extract-utilities" Oct 06 10:27:04 crc kubenswrapper[4824]: E1006 10:27:04.325826 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4cd0302f-df74-4748-bec8-a8b9198b8dfd" containerName="extract-content" Oct 06 10:27:04 crc kubenswrapper[4824]: I1006 10:27:04.325841 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="4cd0302f-df74-4748-bec8-a8b9198b8dfd" containerName="extract-content" Oct 06 10:27:04 crc kubenswrapper[4824]: E1006 10:27:04.325874 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1961267a-e5c0-469e-9f0e-9d4edbc8e64c" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Oct 06 10:27:04 crc kubenswrapper[4824]: I1006 10:27:04.325890 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="1961267a-e5c0-469e-9f0e-9d4edbc8e64c" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Oct 06 10:27:04 crc kubenswrapper[4824]: E1006 10:27:04.325926 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4cd0302f-df74-4748-bec8-a8b9198b8dfd" containerName="registry-server" Oct 06 10:27:04 crc kubenswrapper[4824]: I1006 10:27:04.325938 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="4cd0302f-df74-4748-bec8-a8b9198b8dfd" containerName="registry-server" Oct 06 10:27:04 crc kubenswrapper[4824]: I1006 10:27:04.326301 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="1961267a-e5c0-469e-9f0e-9d4edbc8e64c" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Oct 06 10:27:04 crc kubenswrapper[4824]: I1006 10:27:04.326356 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="4cd0302f-df74-4748-bec8-a8b9198b8dfd" containerName="registry-server" Oct 06 10:27:04 crc kubenswrapper[4824]: I1006 10:27:04.327927 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-jm56f" Oct 06 10:27:04 crc kubenswrapper[4824]: I1006 10:27:04.332754 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 06 10:27:04 crc kubenswrapper[4824]: I1006 10:27:04.333058 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-f886w" Oct 06 10:27:04 crc kubenswrapper[4824]: I1006 10:27:04.333285 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 06 10:27:04 crc kubenswrapper[4824]: I1006 10:27:04.333510 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 06 10:27:04 crc kubenswrapper[4824]: I1006 10:27:04.339418 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-jm56f"] Oct 06 10:27:04 crc kubenswrapper[4824]: I1006 10:27:04.434436 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9684787a-cab3-4930-9ada-f29df39d21a6-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-jm56f\" (UID: \"9684787a-cab3-4930-9ada-f29df39d21a6\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-jm56f" Oct 06 10:27:04 crc kubenswrapper[4824]: I1006 10:27:04.434915 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rngf6\" (UniqueName: \"kubernetes.io/projected/9684787a-cab3-4930-9ada-f29df39d21a6-kube-api-access-rngf6\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-jm56f\" (UID: \"9684787a-cab3-4930-9ada-f29df39d21a6\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-jm56f" Oct 06 10:27:04 crc kubenswrapper[4824]: I1006 10:27:04.435229 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9684787a-cab3-4930-9ada-f29df39d21a6-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-jm56f\" (UID: \"9684787a-cab3-4930-9ada-f29df39d21a6\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-jm56f" Oct 06 10:27:04 crc kubenswrapper[4824]: I1006 10:27:04.538294 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9684787a-cab3-4930-9ada-f29df39d21a6-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-jm56f\" (UID: \"9684787a-cab3-4930-9ada-f29df39d21a6\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-jm56f" Oct 06 10:27:04 crc kubenswrapper[4824]: I1006 10:27:04.538457 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rngf6\" (UniqueName: \"kubernetes.io/projected/9684787a-cab3-4930-9ada-f29df39d21a6-kube-api-access-rngf6\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-jm56f\" (UID: \"9684787a-cab3-4930-9ada-f29df39d21a6\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-jm56f" Oct 06 10:27:04 crc kubenswrapper[4824]: I1006 10:27:04.538557 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9684787a-cab3-4930-9ada-f29df39d21a6-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-jm56f\" (UID: \"9684787a-cab3-4930-9ada-f29df39d21a6\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-jm56f" Oct 06 10:27:04 crc kubenswrapper[4824]: I1006 10:27:04.544681 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9684787a-cab3-4930-9ada-f29df39d21a6-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-jm56f\" (UID: \"9684787a-cab3-4930-9ada-f29df39d21a6\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-jm56f" Oct 06 10:27:04 crc kubenswrapper[4824]: I1006 10:27:04.550136 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9684787a-cab3-4930-9ada-f29df39d21a6-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-jm56f\" (UID: \"9684787a-cab3-4930-9ada-f29df39d21a6\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-jm56f" Oct 06 10:27:04 crc kubenswrapper[4824]: I1006 10:27:04.570071 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rngf6\" (UniqueName: \"kubernetes.io/projected/9684787a-cab3-4930-9ada-f29df39d21a6-kube-api-access-rngf6\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-jm56f\" (UID: \"9684787a-cab3-4930-9ada-f29df39d21a6\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-jm56f" Oct 06 10:27:04 crc kubenswrapper[4824]: I1006 10:27:04.653153 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-jm56f" Oct 06 10:27:05 crc kubenswrapper[4824]: I1006 10:27:05.760305 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-jm56f"] Oct 06 10:27:06 crc kubenswrapper[4824]: I1006 10:27:06.225242 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-jm56f" event={"ID":"9684787a-cab3-4930-9ada-f29df39d21a6","Type":"ContainerStarted","Data":"b33cd38bd337838df12e77835e107fda96a594c77b26654bff1dfddcc19a1137"} Oct 06 10:27:07 crc kubenswrapper[4824]: I1006 10:27:07.238072 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-jm56f" event={"ID":"9684787a-cab3-4930-9ada-f29df39d21a6","Type":"ContainerStarted","Data":"20acb1b8649ba50a42839b743e9546a8fa48f39590e0a853e802a2eccfc84a79"} Oct 06 10:27:07 crc kubenswrapper[4824]: I1006 10:27:07.301368 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-jm56f" podStartSLOduration=2.821750737 podStartE2EDuration="3.301340435s" podCreationTimestamp="2025-10-06 10:27:04 +0000 UTC" firstStartedPulling="2025-10-06 10:27:05.769182398 +0000 UTC m=+1735.133605259" lastFinishedPulling="2025-10-06 10:27:06.248772056 +0000 UTC m=+1735.613194957" observedRunningTime="2025-10-06 10:27:07.271560789 +0000 UTC m=+1736.635983680" watchObservedRunningTime="2025-10-06 10:27:07.301340435 +0000 UTC m=+1736.665763306" Oct 06 10:27:10 crc kubenswrapper[4824]: I1006 10:27:10.274305 4824 scope.go:117] "RemoveContainer" containerID="9de43c7f48ed9ad071a24adf937829ab92276bd505bb72d5b89bac4922aac3b0" Oct 06 10:27:10 crc kubenswrapper[4824]: E1006 10:27:10.275243 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-khgzw_openshift-machine-config-operator(e1140c8c-93f1-4fce-8c06-a69561ff8a6d)\"" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" Oct 06 10:27:12 crc kubenswrapper[4824]: I1006 10:27:12.312938 4824 generic.go:334] "Generic (PLEG): container finished" podID="9684787a-cab3-4930-9ada-f29df39d21a6" containerID="20acb1b8649ba50a42839b743e9546a8fa48f39590e0a853e802a2eccfc84a79" exitCode=0 Oct 06 10:27:12 crc kubenswrapper[4824]: I1006 10:27:12.313229 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-jm56f" event={"ID":"9684787a-cab3-4930-9ada-f29df39d21a6","Type":"ContainerDied","Data":"20acb1b8649ba50a42839b743e9546a8fa48f39590e0a853e802a2eccfc84a79"} Oct 06 10:27:13 crc kubenswrapper[4824]: I1006 10:27:13.817519 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-jm56f" Oct 06 10:27:13 crc kubenswrapper[4824]: I1006 10:27:13.889324 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9684787a-cab3-4930-9ada-f29df39d21a6-ssh-key\") pod \"9684787a-cab3-4930-9ada-f29df39d21a6\" (UID: \"9684787a-cab3-4930-9ada-f29df39d21a6\") " Oct 06 10:27:13 crc kubenswrapper[4824]: I1006 10:27:13.889461 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9684787a-cab3-4930-9ada-f29df39d21a6-inventory\") pod \"9684787a-cab3-4930-9ada-f29df39d21a6\" (UID: \"9684787a-cab3-4930-9ada-f29df39d21a6\") " Oct 06 10:27:13 crc kubenswrapper[4824]: I1006 10:27:13.890730 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rngf6\" (UniqueName: \"kubernetes.io/projected/9684787a-cab3-4930-9ada-f29df39d21a6-kube-api-access-rngf6\") pod \"9684787a-cab3-4930-9ada-f29df39d21a6\" (UID: \"9684787a-cab3-4930-9ada-f29df39d21a6\") " Oct 06 10:27:13 crc kubenswrapper[4824]: I1006 10:27:13.898104 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9684787a-cab3-4930-9ada-f29df39d21a6-kube-api-access-rngf6" (OuterVolumeSpecName: "kube-api-access-rngf6") pod "9684787a-cab3-4930-9ada-f29df39d21a6" (UID: "9684787a-cab3-4930-9ada-f29df39d21a6"). InnerVolumeSpecName "kube-api-access-rngf6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:27:13 crc kubenswrapper[4824]: I1006 10:27:13.926522 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9684787a-cab3-4930-9ada-f29df39d21a6-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "9684787a-cab3-4930-9ada-f29df39d21a6" (UID: "9684787a-cab3-4930-9ada-f29df39d21a6"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:27:13 crc kubenswrapper[4824]: I1006 10:27:13.929634 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9684787a-cab3-4930-9ada-f29df39d21a6-inventory" (OuterVolumeSpecName: "inventory") pod "9684787a-cab3-4930-9ada-f29df39d21a6" (UID: "9684787a-cab3-4930-9ada-f29df39d21a6"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:27:13 crc kubenswrapper[4824]: I1006 10:27:13.993376 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rngf6\" (UniqueName: \"kubernetes.io/projected/9684787a-cab3-4930-9ada-f29df39d21a6-kube-api-access-rngf6\") on node \"crc\" DevicePath \"\"" Oct 06 10:27:13 crc kubenswrapper[4824]: I1006 10:27:13.993416 4824 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9684787a-cab3-4930-9ada-f29df39d21a6-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 06 10:27:13 crc kubenswrapper[4824]: I1006 10:27:13.993426 4824 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9684787a-cab3-4930-9ada-f29df39d21a6-inventory\") on node \"crc\" DevicePath \"\"" Oct 06 10:27:14 crc kubenswrapper[4824]: I1006 10:27:14.335071 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-jm56f" event={"ID":"9684787a-cab3-4930-9ada-f29df39d21a6","Type":"ContainerDied","Data":"b33cd38bd337838df12e77835e107fda96a594c77b26654bff1dfddcc19a1137"} Oct 06 10:27:14 crc kubenswrapper[4824]: I1006 10:27:14.335119 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b33cd38bd337838df12e77835e107fda96a594c77b26654bff1dfddcc19a1137" Oct 06 10:27:14 crc kubenswrapper[4824]: I1006 10:27:14.335239 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-jm56f" Oct 06 10:27:14 crc kubenswrapper[4824]: I1006 10:27:14.509298 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-k82nk"] Oct 06 10:27:14 crc kubenswrapper[4824]: E1006 10:27:14.510159 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9684787a-cab3-4930-9ada-f29df39d21a6" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Oct 06 10:27:14 crc kubenswrapper[4824]: I1006 10:27:14.510200 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="9684787a-cab3-4930-9ada-f29df39d21a6" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Oct 06 10:27:14 crc kubenswrapper[4824]: I1006 10:27:14.510591 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="9684787a-cab3-4930-9ada-f29df39d21a6" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Oct 06 10:27:14 crc kubenswrapper[4824]: I1006 10:27:14.511846 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-k82nk" Oct 06 10:27:14 crc kubenswrapper[4824]: I1006 10:27:14.514969 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 06 10:27:14 crc kubenswrapper[4824]: I1006 10:27:14.515531 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-f886w" Oct 06 10:27:14 crc kubenswrapper[4824]: I1006 10:27:14.515951 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 06 10:27:14 crc kubenswrapper[4824]: I1006 10:27:14.516144 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 06 10:27:14 crc kubenswrapper[4824]: I1006 10:27:14.528166 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-k82nk"] Oct 06 10:27:14 crc kubenswrapper[4824]: I1006 10:27:14.609395 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7db64f14-df4b-4519-81c5-f4e03e053925-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-k82nk\" (UID: \"7db64f14-df4b-4519-81c5-f4e03e053925\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-k82nk" Oct 06 10:27:14 crc kubenswrapper[4824]: I1006 10:27:14.609489 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fnnn2\" (UniqueName: \"kubernetes.io/projected/7db64f14-df4b-4519-81c5-f4e03e053925-kube-api-access-fnnn2\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-k82nk\" (UID: \"7db64f14-df4b-4519-81c5-f4e03e053925\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-k82nk" Oct 06 10:27:14 crc kubenswrapper[4824]: I1006 10:27:14.610197 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7db64f14-df4b-4519-81c5-f4e03e053925-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-k82nk\" (UID: \"7db64f14-df4b-4519-81c5-f4e03e053925\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-k82nk" Oct 06 10:27:14 crc kubenswrapper[4824]: I1006 10:27:14.714576 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7db64f14-df4b-4519-81c5-f4e03e053925-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-k82nk\" (UID: \"7db64f14-df4b-4519-81c5-f4e03e053925\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-k82nk" Oct 06 10:27:14 crc kubenswrapper[4824]: I1006 10:27:14.714702 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7db64f14-df4b-4519-81c5-f4e03e053925-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-k82nk\" (UID: \"7db64f14-df4b-4519-81c5-f4e03e053925\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-k82nk" Oct 06 10:27:14 crc kubenswrapper[4824]: I1006 10:27:14.714818 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fnnn2\" (UniqueName: \"kubernetes.io/projected/7db64f14-df4b-4519-81c5-f4e03e053925-kube-api-access-fnnn2\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-k82nk\" (UID: \"7db64f14-df4b-4519-81c5-f4e03e053925\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-k82nk" Oct 06 10:27:14 crc kubenswrapper[4824]: I1006 10:27:14.729899 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7db64f14-df4b-4519-81c5-f4e03e053925-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-k82nk\" (UID: \"7db64f14-df4b-4519-81c5-f4e03e053925\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-k82nk" Oct 06 10:27:14 crc kubenswrapper[4824]: I1006 10:27:14.731747 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7db64f14-df4b-4519-81c5-f4e03e053925-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-k82nk\" (UID: \"7db64f14-df4b-4519-81c5-f4e03e053925\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-k82nk" Oct 06 10:27:14 crc kubenswrapper[4824]: I1006 10:27:14.739853 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fnnn2\" (UniqueName: \"kubernetes.io/projected/7db64f14-df4b-4519-81c5-f4e03e053925-kube-api-access-fnnn2\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-k82nk\" (UID: \"7db64f14-df4b-4519-81c5-f4e03e053925\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-k82nk" Oct 06 10:27:14 crc kubenswrapper[4824]: I1006 10:27:14.833353 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-k82nk" Oct 06 10:27:15 crc kubenswrapper[4824]: I1006 10:27:15.215251 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-k82nk"] Oct 06 10:27:15 crc kubenswrapper[4824]: I1006 10:27:15.348024 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-k82nk" event={"ID":"7db64f14-df4b-4519-81c5-f4e03e053925","Type":"ContainerStarted","Data":"61c2c0d23f8b18f2f8cc7465e6444804de3282fd93dae70dfa8f4a22a37a2e1d"} Oct 06 10:27:16 crc kubenswrapper[4824]: I1006 10:27:16.369430 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-k82nk" event={"ID":"7db64f14-df4b-4519-81c5-f4e03e053925","Type":"ContainerStarted","Data":"38a79cd0455bc00668b3e074b70c34df0664dd44929d85cc8153947f80b4716d"} Oct 06 10:27:16 crc kubenswrapper[4824]: I1006 10:27:16.390664 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-k82nk" podStartSLOduration=1.73719504 podStartE2EDuration="2.390637568s" podCreationTimestamp="2025-10-06 10:27:14 +0000 UTC" firstStartedPulling="2025-10-06 10:27:15.228197624 +0000 UTC m=+1744.592620485" lastFinishedPulling="2025-10-06 10:27:15.881640152 +0000 UTC m=+1745.246063013" observedRunningTime="2025-10-06 10:27:16.38951452 +0000 UTC m=+1745.753937401" watchObservedRunningTime="2025-10-06 10:27:16.390637568 +0000 UTC m=+1745.755060429" Oct 06 10:27:18 crc kubenswrapper[4824]: I1006 10:27:18.047843 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-lgz2q"] Oct 06 10:27:18 crc kubenswrapper[4824]: I1006 10:27:18.057646 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-lgz2q"] Oct 06 10:27:19 crc kubenswrapper[4824]: I1006 10:27:19.043510 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-48qqz"] Oct 06 10:27:19 crc kubenswrapper[4824]: I1006 10:27:19.056028 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-qwq7x"] Oct 06 10:27:19 crc kubenswrapper[4824]: I1006 10:27:19.070564 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-qwq7x"] Oct 06 10:27:19 crc kubenswrapper[4824]: I1006 10:27:19.082701 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-48qqz"] Oct 06 10:27:19 crc kubenswrapper[4824]: I1006 10:27:19.285489 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="547f97c5-ea4d-4fbc-9f6a-d21ebd33d58f" path="/var/lib/kubelet/pods/547f97c5-ea4d-4fbc-9f6a-d21ebd33d58f/volumes" Oct 06 10:27:19 crc kubenswrapper[4824]: I1006 10:27:19.286826 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6d4717f3-9382-4a1b-a6a3-d0f0f37880c7" path="/var/lib/kubelet/pods/6d4717f3-9382-4a1b-a6a3-d0f0f37880c7/volumes" Oct 06 10:27:19 crc kubenswrapper[4824]: I1006 10:27:19.287478 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c492c641-1fe6-44d1-8b6b-0933efbef65d" path="/var/lib/kubelet/pods/c492c641-1fe6-44d1-8b6b-0933efbef65d/volumes" Oct 06 10:27:24 crc kubenswrapper[4824]: I1006 10:27:24.274496 4824 scope.go:117] "RemoveContainer" containerID="9de43c7f48ed9ad071a24adf937829ab92276bd505bb72d5b89bac4922aac3b0" Oct 06 10:27:24 crc kubenswrapper[4824]: E1006 10:27:24.275456 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-khgzw_openshift-machine-config-operator(e1140c8c-93f1-4fce-8c06-a69561ff8a6d)\"" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" Oct 06 10:27:31 crc kubenswrapper[4824]: I1006 10:27:31.984449 4824 scope.go:117] "RemoveContainer" containerID="cd64c0d30963bc4055ec2f373fd1f3c755ad602d382d96c789000d1db60c8a91" Oct 06 10:27:32 crc kubenswrapper[4824]: I1006 10:27:32.040961 4824 scope.go:117] "RemoveContainer" containerID="429cd910467ac58cccbd723e79ad017a0046e53a5d5f63976b2ebc8c40185f77" Oct 06 10:27:32 crc kubenswrapper[4824]: I1006 10:27:32.104963 4824 scope.go:117] "RemoveContainer" containerID="0197e3d5f2022700f4e6658e19fec631b535f5202acfa0bf383ab4c2a7407e42" Oct 06 10:27:32 crc kubenswrapper[4824]: I1006 10:27:32.143822 4824 scope.go:117] "RemoveContainer" containerID="87baaf8f01793ebfb0d0c10a1534bef63ae25ed902ab10b41d81f273be74418b" Oct 06 10:27:32 crc kubenswrapper[4824]: I1006 10:27:32.209051 4824 scope.go:117] "RemoveContainer" containerID="bfe65796f2dadbf11294a2fca859967193ea63bf8d8080fbb171c2311304906e" Oct 06 10:27:34 crc kubenswrapper[4824]: I1006 10:27:34.066533 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-c73c-account-create-q9hfk"] Oct 06 10:27:34 crc kubenswrapper[4824]: I1006 10:27:34.082965 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-e2d6-account-create-rxhwz"] Oct 06 10:27:34 crc kubenswrapper[4824]: I1006 10:27:34.093531 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-e2d6-account-create-rxhwz"] Oct 06 10:27:34 crc kubenswrapper[4824]: I1006 10:27:34.104159 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-c73c-account-create-q9hfk"] Oct 06 10:27:35 crc kubenswrapper[4824]: I1006 10:27:35.054703 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-aa65-account-create-l9c7p"] Oct 06 10:27:35 crc kubenswrapper[4824]: I1006 10:27:35.070852 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-aa65-account-create-l9c7p"] Oct 06 10:27:35 crc kubenswrapper[4824]: I1006 10:27:35.287914 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4007653a-5c39-4285-8ba3-e0b3caed5640" path="/var/lib/kubelet/pods/4007653a-5c39-4285-8ba3-e0b3caed5640/volumes" Oct 06 10:27:35 crc kubenswrapper[4824]: I1006 10:27:35.288452 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7cbd2d79-2d35-4954-9f61-c9bbbbe9d84e" path="/var/lib/kubelet/pods/7cbd2d79-2d35-4954-9f61-c9bbbbe9d84e/volumes" Oct 06 10:27:35 crc kubenswrapper[4824]: I1006 10:27:35.288937 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7dd053a5-58ba-4f1c-9d32-e639fa8b1450" path="/var/lib/kubelet/pods/7dd053a5-58ba-4f1c-9d32-e639fa8b1450/volumes" Oct 06 10:27:38 crc kubenswrapper[4824]: I1006 10:27:38.274104 4824 scope.go:117] "RemoveContainer" containerID="9de43c7f48ed9ad071a24adf937829ab92276bd505bb72d5b89bac4922aac3b0" Oct 06 10:27:38 crc kubenswrapper[4824]: E1006 10:27:38.274694 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-khgzw_openshift-machine-config-operator(e1140c8c-93f1-4fce-8c06-a69561ff8a6d)\"" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" Oct 06 10:27:49 crc kubenswrapper[4824]: I1006 10:27:49.275572 4824 scope.go:117] "RemoveContainer" containerID="9de43c7f48ed9ad071a24adf937829ab92276bd505bb72d5b89bac4922aac3b0" Oct 06 10:27:49 crc kubenswrapper[4824]: E1006 10:27:49.285487 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-khgzw_openshift-machine-config-operator(e1140c8c-93f1-4fce-8c06-a69561ff8a6d)\"" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" Oct 06 10:27:56 crc kubenswrapper[4824]: I1006 10:27:56.815793 4824 generic.go:334] "Generic (PLEG): container finished" podID="7db64f14-df4b-4519-81c5-f4e03e053925" containerID="38a79cd0455bc00668b3e074b70c34df0664dd44929d85cc8153947f80b4716d" exitCode=0 Oct 06 10:27:56 crc kubenswrapper[4824]: I1006 10:27:56.816039 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-k82nk" event={"ID":"7db64f14-df4b-4519-81c5-f4e03e053925","Type":"ContainerDied","Data":"38a79cd0455bc00668b3e074b70c34df0664dd44929d85cc8153947f80b4716d"} Oct 06 10:27:58 crc kubenswrapper[4824]: I1006 10:27:58.059067 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-7txxj"] Oct 06 10:27:58 crc kubenswrapper[4824]: I1006 10:27:58.071904 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-7txxj"] Oct 06 10:27:58 crc kubenswrapper[4824]: I1006 10:27:58.461059 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-k82nk" Oct 06 10:27:58 crc kubenswrapper[4824]: I1006 10:27:58.594824 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fnnn2\" (UniqueName: \"kubernetes.io/projected/7db64f14-df4b-4519-81c5-f4e03e053925-kube-api-access-fnnn2\") pod \"7db64f14-df4b-4519-81c5-f4e03e053925\" (UID: \"7db64f14-df4b-4519-81c5-f4e03e053925\") " Oct 06 10:27:58 crc kubenswrapper[4824]: I1006 10:27:58.594929 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7db64f14-df4b-4519-81c5-f4e03e053925-inventory\") pod \"7db64f14-df4b-4519-81c5-f4e03e053925\" (UID: \"7db64f14-df4b-4519-81c5-f4e03e053925\") " Oct 06 10:27:58 crc kubenswrapper[4824]: I1006 10:27:58.595018 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7db64f14-df4b-4519-81c5-f4e03e053925-ssh-key\") pod \"7db64f14-df4b-4519-81c5-f4e03e053925\" (UID: \"7db64f14-df4b-4519-81c5-f4e03e053925\") " Oct 06 10:27:58 crc kubenswrapper[4824]: I1006 10:27:58.603282 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7db64f14-df4b-4519-81c5-f4e03e053925-kube-api-access-fnnn2" (OuterVolumeSpecName: "kube-api-access-fnnn2") pod "7db64f14-df4b-4519-81c5-f4e03e053925" (UID: "7db64f14-df4b-4519-81c5-f4e03e053925"). InnerVolumeSpecName "kube-api-access-fnnn2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:27:58 crc kubenswrapper[4824]: I1006 10:27:58.642202 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7db64f14-df4b-4519-81c5-f4e03e053925-inventory" (OuterVolumeSpecName: "inventory") pod "7db64f14-df4b-4519-81c5-f4e03e053925" (UID: "7db64f14-df4b-4519-81c5-f4e03e053925"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:27:58 crc kubenswrapper[4824]: I1006 10:27:58.660918 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7db64f14-df4b-4519-81c5-f4e03e053925-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "7db64f14-df4b-4519-81c5-f4e03e053925" (UID: "7db64f14-df4b-4519-81c5-f4e03e053925"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:27:58 crc kubenswrapper[4824]: I1006 10:27:58.697800 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fnnn2\" (UniqueName: \"kubernetes.io/projected/7db64f14-df4b-4519-81c5-f4e03e053925-kube-api-access-fnnn2\") on node \"crc\" DevicePath \"\"" Oct 06 10:27:58 crc kubenswrapper[4824]: I1006 10:27:58.697948 4824 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7db64f14-df4b-4519-81c5-f4e03e053925-inventory\") on node \"crc\" DevicePath \"\"" Oct 06 10:27:58 crc kubenswrapper[4824]: I1006 10:27:58.697959 4824 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7db64f14-df4b-4519-81c5-f4e03e053925-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 06 10:27:58 crc kubenswrapper[4824]: I1006 10:27:58.837096 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-k82nk" event={"ID":"7db64f14-df4b-4519-81c5-f4e03e053925","Type":"ContainerDied","Data":"61c2c0d23f8b18f2f8cc7465e6444804de3282fd93dae70dfa8f4a22a37a2e1d"} Oct 06 10:27:58 crc kubenswrapper[4824]: I1006 10:27:58.837153 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="61c2c0d23f8b18f2f8cc7465e6444804de3282fd93dae70dfa8f4a22a37a2e1d" Oct 06 10:27:58 crc kubenswrapper[4824]: I1006 10:27:58.837209 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-k82nk" Oct 06 10:27:58 crc kubenswrapper[4824]: I1006 10:27:58.930147 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-shvtw"] Oct 06 10:27:58 crc kubenswrapper[4824]: E1006 10:27:58.930681 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7db64f14-df4b-4519-81c5-f4e03e053925" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Oct 06 10:27:58 crc kubenswrapper[4824]: I1006 10:27:58.930707 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="7db64f14-df4b-4519-81c5-f4e03e053925" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Oct 06 10:27:58 crc kubenswrapper[4824]: I1006 10:27:58.931097 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="7db64f14-df4b-4519-81c5-f4e03e053925" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Oct 06 10:27:58 crc kubenswrapper[4824]: I1006 10:27:58.932079 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-shvtw" Oct 06 10:27:58 crc kubenswrapper[4824]: I1006 10:27:58.934348 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 06 10:27:58 crc kubenswrapper[4824]: I1006 10:27:58.934831 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 06 10:27:58 crc kubenswrapper[4824]: I1006 10:27:58.934880 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-f886w" Oct 06 10:27:58 crc kubenswrapper[4824]: I1006 10:27:58.936404 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 06 10:27:58 crc kubenswrapper[4824]: I1006 10:27:58.944356 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-shvtw"] Oct 06 10:27:59 crc kubenswrapper[4824]: I1006 10:27:59.005011 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3a56b28d-58ae-4cd4-a06a-b942a9365de7-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-shvtw\" (UID: \"3a56b28d-58ae-4cd4-a06a-b942a9365de7\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-shvtw" Oct 06 10:27:59 crc kubenswrapper[4824]: I1006 10:27:59.005173 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pwgh6\" (UniqueName: \"kubernetes.io/projected/3a56b28d-58ae-4cd4-a06a-b942a9365de7-kube-api-access-pwgh6\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-shvtw\" (UID: \"3a56b28d-58ae-4cd4-a06a-b942a9365de7\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-shvtw" Oct 06 10:27:59 crc kubenswrapper[4824]: I1006 10:27:59.005204 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3a56b28d-58ae-4cd4-a06a-b942a9365de7-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-shvtw\" (UID: \"3a56b28d-58ae-4cd4-a06a-b942a9365de7\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-shvtw" Oct 06 10:27:59 crc kubenswrapper[4824]: I1006 10:27:59.107943 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3a56b28d-58ae-4cd4-a06a-b942a9365de7-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-shvtw\" (UID: \"3a56b28d-58ae-4cd4-a06a-b942a9365de7\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-shvtw" Oct 06 10:27:59 crc kubenswrapper[4824]: I1006 10:27:59.108364 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pwgh6\" (UniqueName: \"kubernetes.io/projected/3a56b28d-58ae-4cd4-a06a-b942a9365de7-kube-api-access-pwgh6\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-shvtw\" (UID: \"3a56b28d-58ae-4cd4-a06a-b942a9365de7\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-shvtw" Oct 06 10:27:59 crc kubenswrapper[4824]: I1006 10:27:59.108398 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3a56b28d-58ae-4cd4-a06a-b942a9365de7-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-shvtw\" (UID: \"3a56b28d-58ae-4cd4-a06a-b942a9365de7\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-shvtw" Oct 06 10:27:59 crc kubenswrapper[4824]: I1006 10:27:59.112293 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3a56b28d-58ae-4cd4-a06a-b942a9365de7-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-shvtw\" (UID: \"3a56b28d-58ae-4cd4-a06a-b942a9365de7\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-shvtw" Oct 06 10:27:59 crc kubenswrapper[4824]: I1006 10:27:59.113813 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3a56b28d-58ae-4cd4-a06a-b942a9365de7-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-shvtw\" (UID: \"3a56b28d-58ae-4cd4-a06a-b942a9365de7\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-shvtw" Oct 06 10:27:59 crc kubenswrapper[4824]: I1006 10:27:59.131037 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pwgh6\" (UniqueName: \"kubernetes.io/projected/3a56b28d-58ae-4cd4-a06a-b942a9365de7-kube-api-access-pwgh6\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-shvtw\" (UID: \"3a56b28d-58ae-4cd4-a06a-b942a9365de7\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-shvtw" Oct 06 10:27:59 crc kubenswrapper[4824]: I1006 10:27:59.250160 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-shvtw" Oct 06 10:27:59 crc kubenswrapper[4824]: I1006 10:27:59.287613 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8a08722d-580d-4dbc-80d5-f33caa5cae5b" path="/var/lib/kubelet/pods/8a08722d-580d-4dbc-80d5-f33caa5cae5b/volumes" Oct 06 10:27:59 crc kubenswrapper[4824]: I1006 10:27:59.871420 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-shvtw"] Oct 06 10:28:00 crc kubenswrapper[4824]: I1006 10:28:00.859594 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-shvtw" event={"ID":"3a56b28d-58ae-4cd4-a06a-b942a9365de7","Type":"ContainerStarted","Data":"c75b5649eb048b623e3f769ae033c1dc28c9e6acdd175d89024a31d4dc3ff844"} Oct 06 10:28:00 crc kubenswrapper[4824]: I1006 10:28:00.860177 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-shvtw" event={"ID":"3a56b28d-58ae-4cd4-a06a-b942a9365de7","Type":"ContainerStarted","Data":"d4f8f24c246c822be68b900336dcfc30e8e7b3a3ecfbdf100d0d8456a7d1c1ef"} Oct 06 10:28:00 crc kubenswrapper[4824]: I1006 10:28:00.882002 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-shvtw" podStartSLOduration=2.436649939 podStartE2EDuration="2.881956587s" podCreationTimestamp="2025-10-06 10:27:58 +0000 UTC" firstStartedPulling="2025-10-06 10:27:59.871140946 +0000 UTC m=+1789.235563807" lastFinishedPulling="2025-10-06 10:28:00.316447574 +0000 UTC m=+1789.680870455" observedRunningTime="2025-10-06 10:28:00.881423015 +0000 UTC m=+1790.245845896" watchObservedRunningTime="2025-10-06 10:28:00.881956587 +0000 UTC m=+1790.246379458" Oct 06 10:28:03 crc kubenswrapper[4824]: I1006 10:28:03.274792 4824 scope.go:117] "RemoveContainer" containerID="9de43c7f48ed9ad071a24adf937829ab92276bd505bb72d5b89bac4922aac3b0" Oct 06 10:28:03 crc kubenswrapper[4824]: E1006 10:28:03.275637 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-khgzw_openshift-machine-config-operator(e1140c8c-93f1-4fce-8c06-a69561ff8a6d)\"" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" Oct 06 10:28:18 crc kubenswrapper[4824]: I1006 10:28:18.274994 4824 scope.go:117] "RemoveContainer" containerID="9de43c7f48ed9ad071a24adf937829ab92276bd505bb72d5b89bac4922aac3b0" Oct 06 10:28:18 crc kubenswrapper[4824]: E1006 10:28:18.276318 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-khgzw_openshift-machine-config-operator(e1140c8c-93f1-4fce-8c06-a69561ff8a6d)\"" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" Oct 06 10:28:23 crc kubenswrapper[4824]: I1006 10:28:23.043603 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-cbbc7"] Oct 06 10:28:23 crc kubenswrapper[4824]: I1006 10:28:23.052165 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-cbbc7"] Oct 06 10:28:23 crc kubenswrapper[4824]: I1006 10:28:23.287033 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="795af81e-370f-4845-8632-acfe7988e3cf" path="/var/lib/kubelet/pods/795af81e-370f-4845-8632-acfe7988e3cf/volumes" Oct 06 10:28:24 crc kubenswrapper[4824]: I1006 10:28:24.034964 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-zq5td"] Oct 06 10:28:24 crc kubenswrapper[4824]: I1006 10:28:24.044360 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-zq5td"] Oct 06 10:28:25 crc kubenswrapper[4824]: I1006 10:28:25.293704 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd95e609-bd54-4234-a449-66cd65e2558a" path="/var/lib/kubelet/pods/cd95e609-bd54-4234-a449-66cd65e2558a/volumes" Oct 06 10:28:32 crc kubenswrapper[4824]: I1006 10:28:32.435302 4824 scope.go:117] "RemoveContainer" containerID="d7a21bc36e001b5d47e59fe8dd2712c72c5714c4632964f952a86eead6ad34bc" Oct 06 10:28:32 crc kubenswrapper[4824]: I1006 10:28:32.497252 4824 scope.go:117] "RemoveContainer" containerID="ad61d4bd99222d6b606f4e3eb3c1f5d55d69b5ac369233c8514fa1400dca4650" Oct 06 10:28:32 crc kubenswrapper[4824]: I1006 10:28:32.545585 4824 scope.go:117] "RemoveContainer" containerID="380cdc21f47ff79047047cde2a74671a6cd209a886c1a0dc68647e0acebb1e32" Oct 06 10:28:32 crc kubenswrapper[4824]: I1006 10:28:32.615280 4824 scope.go:117] "RemoveContainer" containerID="7ebd18e643bfc8279abcc47f6aeb5952741001d0ec2376eecfaf1520463b43e1" Oct 06 10:28:32 crc kubenswrapper[4824]: I1006 10:28:32.676055 4824 scope.go:117] "RemoveContainer" containerID="aed46891458e2473b16dd98b74441f192fc6315b49f05c1ec0728035643b5019" Oct 06 10:28:32 crc kubenswrapper[4824]: I1006 10:28:32.704610 4824 scope.go:117] "RemoveContainer" containerID="e93f66db5683a4761c779328c5d70b916fcb98226c72e9ab3dc9a4e98af10ebd" Oct 06 10:28:33 crc kubenswrapper[4824]: I1006 10:28:33.275246 4824 scope.go:117] "RemoveContainer" containerID="9de43c7f48ed9ad071a24adf937829ab92276bd505bb72d5b89bac4922aac3b0" Oct 06 10:28:33 crc kubenswrapper[4824]: E1006 10:28:33.275716 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-khgzw_openshift-machine-config-operator(e1140c8c-93f1-4fce-8c06-a69561ff8a6d)\"" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" Oct 06 10:28:47 crc kubenswrapper[4824]: I1006 10:28:47.275464 4824 scope.go:117] "RemoveContainer" containerID="9de43c7f48ed9ad071a24adf937829ab92276bd505bb72d5b89bac4922aac3b0" Oct 06 10:28:47 crc kubenswrapper[4824]: E1006 10:28:47.276768 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-khgzw_openshift-machine-config-operator(e1140c8c-93f1-4fce-8c06-a69561ff8a6d)\"" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" Oct 06 10:28:55 crc kubenswrapper[4824]: I1006 10:28:55.467227 4824 generic.go:334] "Generic (PLEG): container finished" podID="3a56b28d-58ae-4cd4-a06a-b942a9365de7" containerID="c75b5649eb048b623e3f769ae033c1dc28c9e6acdd175d89024a31d4dc3ff844" exitCode=2 Oct 06 10:28:55 crc kubenswrapper[4824]: I1006 10:28:55.467319 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-shvtw" event={"ID":"3a56b28d-58ae-4cd4-a06a-b942a9365de7","Type":"ContainerDied","Data":"c75b5649eb048b623e3f769ae033c1dc28c9e6acdd175d89024a31d4dc3ff844"} Oct 06 10:28:56 crc kubenswrapper[4824]: I1006 10:28:56.954562 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-shvtw" Oct 06 10:28:57 crc kubenswrapper[4824]: I1006 10:28:57.056727 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3a56b28d-58ae-4cd4-a06a-b942a9365de7-inventory\") pod \"3a56b28d-58ae-4cd4-a06a-b942a9365de7\" (UID: \"3a56b28d-58ae-4cd4-a06a-b942a9365de7\") " Oct 06 10:28:57 crc kubenswrapper[4824]: I1006 10:28:57.056933 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pwgh6\" (UniqueName: \"kubernetes.io/projected/3a56b28d-58ae-4cd4-a06a-b942a9365de7-kube-api-access-pwgh6\") pod \"3a56b28d-58ae-4cd4-a06a-b942a9365de7\" (UID: \"3a56b28d-58ae-4cd4-a06a-b942a9365de7\") " Oct 06 10:28:57 crc kubenswrapper[4824]: I1006 10:28:57.057129 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3a56b28d-58ae-4cd4-a06a-b942a9365de7-ssh-key\") pod \"3a56b28d-58ae-4cd4-a06a-b942a9365de7\" (UID: \"3a56b28d-58ae-4cd4-a06a-b942a9365de7\") " Oct 06 10:28:57 crc kubenswrapper[4824]: I1006 10:28:57.065479 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3a56b28d-58ae-4cd4-a06a-b942a9365de7-kube-api-access-pwgh6" (OuterVolumeSpecName: "kube-api-access-pwgh6") pod "3a56b28d-58ae-4cd4-a06a-b942a9365de7" (UID: "3a56b28d-58ae-4cd4-a06a-b942a9365de7"). InnerVolumeSpecName "kube-api-access-pwgh6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:28:57 crc kubenswrapper[4824]: I1006 10:28:57.091802 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a56b28d-58ae-4cd4-a06a-b942a9365de7-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "3a56b28d-58ae-4cd4-a06a-b942a9365de7" (UID: "3a56b28d-58ae-4cd4-a06a-b942a9365de7"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:28:57 crc kubenswrapper[4824]: I1006 10:28:57.093613 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a56b28d-58ae-4cd4-a06a-b942a9365de7-inventory" (OuterVolumeSpecName: "inventory") pod "3a56b28d-58ae-4cd4-a06a-b942a9365de7" (UID: "3a56b28d-58ae-4cd4-a06a-b942a9365de7"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:28:57 crc kubenswrapper[4824]: I1006 10:28:57.161587 4824 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3a56b28d-58ae-4cd4-a06a-b942a9365de7-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 06 10:28:57 crc kubenswrapper[4824]: I1006 10:28:57.161618 4824 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3a56b28d-58ae-4cd4-a06a-b942a9365de7-inventory\") on node \"crc\" DevicePath \"\"" Oct 06 10:28:57 crc kubenswrapper[4824]: I1006 10:28:57.161628 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pwgh6\" (UniqueName: \"kubernetes.io/projected/3a56b28d-58ae-4cd4-a06a-b942a9365de7-kube-api-access-pwgh6\") on node \"crc\" DevicePath \"\"" Oct 06 10:28:57 crc kubenswrapper[4824]: I1006 10:28:57.488124 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-shvtw" event={"ID":"3a56b28d-58ae-4cd4-a06a-b942a9365de7","Type":"ContainerDied","Data":"d4f8f24c246c822be68b900336dcfc30e8e7b3a3ecfbdf100d0d8456a7d1c1ef"} Oct 06 10:28:57 crc kubenswrapper[4824]: I1006 10:28:57.488190 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d4f8f24c246c822be68b900336dcfc30e8e7b3a3ecfbdf100d0d8456a7d1c1ef" Oct 06 10:28:57 crc kubenswrapper[4824]: I1006 10:28:57.488236 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-shvtw" Oct 06 10:28:58 crc kubenswrapper[4824]: I1006 10:28:58.274593 4824 scope.go:117] "RemoveContainer" containerID="9de43c7f48ed9ad071a24adf937829ab92276bd505bb72d5b89bac4922aac3b0" Oct 06 10:28:58 crc kubenswrapper[4824]: E1006 10:28:58.275308 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-khgzw_openshift-machine-config-operator(e1140c8c-93f1-4fce-8c06-a69561ff8a6d)\"" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" Oct 06 10:29:04 crc kubenswrapper[4824]: I1006 10:29:04.034020 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-nm76g"] Oct 06 10:29:04 crc kubenswrapper[4824]: E1006 10:29:04.036295 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a56b28d-58ae-4cd4-a06a-b942a9365de7" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 06 10:29:04 crc kubenswrapper[4824]: I1006 10:29:04.036374 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a56b28d-58ae-4cd4-a06a-b942a9365de7" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 06 10:29:04 crc kubenswrapper[4824]: I1006 10:29:04.036648 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a56b28d-58ae-4cd4-a06a-b942a9365de7" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 06 10:29:04 crc kubenswrapper[4824]: I1006 10:29:04.037648 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-nm76g" Oct 06 10:29:04 crc kubenswrapper[4824]: I1006 10:29:04.040712 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 06 10:29:04 crc kubenswrapper[4824]: I1006 10:29:04.040888 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-f886w" Oct 06 10:29:04 crc kubenswrapper[4824]: I1006 10:29:04.041144 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 06 10:29:04 crc kubenswrapper[4824]: I1006 10:29:04.041655 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 06 10:29:04 crc kubenswrapper[4824]: I1006 10:29:04.050732 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-nm76g"] Oct 06 10:29:04 crc kubenswrapper[4824]: I1006 10:29:04.131963 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/87320828-94da-481c-b903-e7d478e3df65-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-nm76g\" (UID: \"87320828-94da-481c-b903-e7d478e3df65\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-nm76g" Oct 06 10:29:04 crc kubenswrapper[4824]: I1006 10:29:04.132530 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-thshr\" (UniqueName: \"kubernetes.io/projected/87320828-94da-481c-b903-e7d478e3df65-kube-api-access-thshr\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-nm76g\" (UID: \"87320828-94da-481c-b903-e7d478e3df65\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-nm76g" Oct 06 10:29:04 crc kubenswrapper[4824]: I1006 10:29:04.132556 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/87320828-94da-481c-b903-e7d478e3df65-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-nm76g\" (UID: \"87320828-94da-481c-b903-e7d478e3df65\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-nm76g" Oct 06 10:29:04 crc kubenswrapper[4824]: I1006 10:29:04.234656 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-thshr\" (UniqueName: \"kubernetes.io/projected/87320828-94da-481c-b903-e7d478e3df65-kube-api-access-thshr\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-nm76g\" (UID: \"87320828-94da-481c-b903-e7d478e3df65\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-nm76g" Oct 06 10:29:04 crc kubenswrapper[4824]: I1006 10:29:04.234714 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/87320828-94da-481c-b903-e7d478e3df65-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-nm76g\" (UID: \"87320828-94da-481c-b903-e7d478e3df65\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-nm76g" Oct 06 10:29:04 crc kubenswrapper[4824]: I1006 10:29:04.234876 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/87320828-94da-481c-b903-e7d478e3df65-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-nm76g\" (UID: \"87320828-94da-481c-b903-e7d478e3df65\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-nm76g" Oct 06 10:29:04 crc kubenswrapper[4824]: I1006 10:29:04.242898 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/87320828-94da-481c-b903-e7d478e3df65-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-nm76g\" (UID: \"87320828-94da-481c-b903-e7d478e3df65\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-nm76g" Oct 06 10:29:04 crc kubenswrapper[4824]: I1006 10:29:04.244187 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/87320828-94da-481c-b903-e7d478e3df65-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-nm76g\" (UID: \"87320828-94da-481c-b903-e7d478e3df65\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-nm76g" Oct 06 10:29:04 crc kubenswrapper[4824]: I1006 10:29:04.262231 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-thshr\" (UniqueName: \"kubernetes.io/projected/87320828-94da-481c-b903-e7d478e3df65-kube-api-access-thshr\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-nm76g\" (UID: \"87320828-94da-481c-b903-e7d478e3df65\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-nm76g" Oct 06 10:29:04 crc kubenswrapper[4824]: I1006 10:29:04.416251 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-nm76g" Oct 06 10:29:05 crc kubenswrapper[4824]: I1006 10:29:05.054724 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-nm76g"] Oct 06 10:29:05 crc kubenswrapper[4824]: I1006 10:29:05.591321 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-nm76g" event={"ID":"87320828-94da-481c-b903-e7d478e3df65","Type":"ContainerStarted","Data":"03b85f8eaf22bdd22f521e494874c668cad5a105517a80ef3a4ab75e1db2322a"} Oct 06 10:29:06 crc kubenswrapper[4824]: I1006 10:29:06.606590 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-nm76g" event={"ID":"87320828-94da-481c-b903-e7d478e3df65","Type":"ContainerStarted","Data":"e75e6f4ad2f06d545749a03035a0f5dc5d072736aad46941dea33e092e92d144"} Oct 06 10:29:06 crc kubenswrapper[4824]: I1006 10:29:06.636328 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-nm76g" podStartSLOduration=2.173511558 podStartE2EDuration="2.636295475s" podCreationTimestamp="2025-10-06 10:29:04 +0000 UTC" firstStartedPulling="2025-10-06 10:29:05.095146672 +0000 UTC m=+1854.459569533" lastFinishedPulling="2025-10-06 10:29:05.557930549 +0000 UTC m=+1854.922353450" observedRunningTime="2025-10-06 10:29:06.635040223 +0000 UTC m=+1855.999463094" watchObservedRunningTime="2025-10-06 10:29:06.636295475 +0000 UTC m=+1856.000718346" Oct 06 10:29:08 crc kubenswrapper[4824]: I1006 10:29:08.048818 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-mrzrt"] Oct 06 10:29:08 crc kubenswrapper[4824]: I1006 10:29:08.057716 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-mrzrt"] Oct 06 10:29:09 crc kubenswrapper[4824]: I1006 10:29:09.292752 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="06af5e40-91c0-45c9-9890-ffe7673be037" path="/var/lib/kubelet/pods/06af5e40-91c0-45c9-9890-ffe7673be037/volumes" Oct 06 10:29:10 crc kubenswrapper[4824]: I1006 10:29:10.274298 4824 scope.go:117] "RemoveContainer" containerID="9de43c7f48ed9ad071a24adf937829ab92276bd505bb72d5b89bac4922aac3b0" Oct 06 10:29:10 crc kubenswrapper[4824]: E1006 10:29:10.274829 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-khgzw_openshift-machine-config-operator(e1140c8c-93f1-4fce-8c06-a69561ff8a6d)\"" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" Oct 06 10:29:22 crc kubenswrapper[4824]: I1006 10:29:22.274467 4824 scope.go:117] "RemoveContainer" containerID="9de43c7f48ed9ad071a24adf937829ab92276bd505bb72d5b89bac4922aac3b0" Oct 06 10:29:22 crc kubenswrapper[4824]: E1006 10:29:22.275429 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-khgzw_openshift-machine-config-operator(e1140c8c-93f1-4fce-8c06-a69561ff8a6d)\"" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" Oct 06 10:29:32 crc kubenswrapper[4824]: I1006 10:29:32.904670 4824 scope.go:117] "RemoveContainer" containerID="100570af227d742275ce56ad92c0165761a6c359e8215e08efe07f27d976decc" Oct 06 10:29:34 crc kubenswrapper[4824]: I1006 10:29:34.275051 4824 scope.go:117] "RemoveContainer" containerID="9de43c7f48ed9ad071a24adf937829ab92276bd505bb72d5b89bac4922aac3b0" Oct 06 10:29:34 crc kubenswrapper[4824]: E1006 10:29:34.275916 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-khgzw_openshift-machine-config-operator(e1140c8c-93f1-4fce-8c06-a69561ff8a6d)\"" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" Oct 06 10:29:48 crc kubenswrapper[4824]: I1006 10:29:48.274302 4824 scope.go:117] "RemoveContainer" containerID="9de43c7f48ed9ad071a24adf937829ab92276bd505bb72d5b89bac4922aac3b0" Oct 06 10:29:48 crc kubenswrapper[4824]: E1006 10:29:48.276174 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-khgzw_openshift-machine-config-operator(e1140c8c-93f1-4fce-8c06-a69561ff8a6d)\"" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" Oct 06 10:29:52 crc kubenswrapper[4824]: I1006 10:29:52.093175 4824 generic.go:334] "Generic (PLEG): container finished" podID="87320828-94da-481c-b903-e7d478e3df65" containerID="e75e6f4ad2f06d545749a03035a0f5dc5d072736aad46941dea33e092e92d144" exitCode=0 Oct 06 10:29:52 crc kubenswrapper[4824]: I1006 10:29:52.093303 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-nm76g" event={"ID":"87320828-94da-481c-b903-e7d478e3df65","Type":"ContainerDied","Data":"e75e6f4ad2f06d545749a03035a0f5dc5d072736aad46941dea33e092e92d144"} Oct 06 10:29:53 crc kubenswrapper[4824]: I1006 10:29:53.591114 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-nm76g" Oct 06 10:29:53 crc kubenswrapper[4824]: I1006 10:29:53.704938 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-thshr\" (UniqueName: \"kubernetes.io/projected/87320828-94da-481c-b903-e7d478e3df65-kube-api-access-thshr\") pod \"87320828-94da-481c-b903-e7d478e3df65\" (UID: \"87320828-94da-481c-b903-e7d478e3df65\") " Oct 06 10:29:53 crc kubenswrapper[4824]: I1006 10:29:53.705108 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/87320828-94da-481c-b903-e7d478e3df65-ssh-key\") pod \"87320828-94da-481c-b903-e7d478e3df65\" (UID: \"87320828-94da-481c-b903-e7d478e3df65\") " Oct 06 10:29:53 crc kubenswrapper[4824]: I1006 10:29:53.705317 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/87320828-94da-481c-b903-e7d478e3df65-inventory\") pod \"87320828-94da-481c-b903-e7d478e3df65\" (UID: \"87320828-94da-481c-b903-e7d478e3df65\") " Oct 06 10:29:53 crc kubenswrapper[4824]: I1006 10:29:53.712261 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87320828-94da-481c-b903-e7d478e3df65-kube-api-access-thshr" (OuterVolumeSpecName: "kube-api-access-thshr") pod "87320828-94da-481c-b903-e7d478e3df65" (UID: "87320828-94da-481c-b903-e7d478e3df65"). InnerVolumeSpecName "kube-api-access-thshr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:29:53 crc kubenswrapper[4824]: I1006 10:29:53.736687 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87320828-94da-481c-b903-e7d478e3df65-inventory" (OuterVolumeSpecName: "inventory") pod "87320828-94da-481c-b903-e7d478e3df65" (UID: "87320828-94da-481c-b903-e7d478e3df65"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:29:53 crc kubenswrapper[4824]: I1006 10:29:53.743126 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87320828-94da-481c-b903-e7d478e3df65-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "87320828-94da-481c-b903-e7d478e3df65" (UID: "87320828-94da-481c-b903-e7d478e3df65"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:29:53 crc kubenswrapper[4824]: I1006 10:29:53.807531 4824 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/87320828-94da-481c-b903-e7d478e3df65-inventory\") on node \"crc\" DevicePath \"\"" Oct 06 10:29:53 crc kubenswrapper[4824]: I1006 10:29:53.807570 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-thshr\" (UniqueName: \"kubernetes.io/projected/87320828-94da-481c-b903-e7d478e3df65-kube-api-access-thshr\") on node \"crc\" DevicePath \"\"" Oct 06 10:29:53 crc kubenswrapper[4824]: I1006 10:29:53.807584 4824 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/87320828-94da-481c-b903-e7d478e3df65-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 06 10:29:54 crc kubenswrapper[4824]: I1006 10:29:54.122302 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-nm76g" event={"ID":"87320828-94da-481c-b903-e7d478e3df65","Type":"ContainerDied","Data":"03b85f8eaf22bdd22f521e494874c668cad5a105517a80ef3a4ab75e1db2322a"} Oct 06 10:29:54 crc kubenswrapper[4824]: I1006 10:29:54.122823 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="03b85f8eaf22bdd22f521e494874c668cad5a105517a80ef3a4ab75e1db2322a" Oct 06 10:29:54 crc kubenswrapper[4824]: I1006 10:29:54.122387 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-nm76g" Oct 06 10:29:54 crc kubenswrapper[4824]: I1006 10:29:54.234572 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-d24cq"] Oct 06 10:29:54 crc kubenswrapper[4824]: E1006 10:29:54.235192 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87320828-94da-481c-b903-e7d478e3df65" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 06 10:29:54 crc kubenswrapper[4824]: I1006 10:29:54.235222 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="87320828-94da-481c-b903-e7d478e3df65" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 06 10:29:54 crc kubenswrapper[4824]: I1006 10:29:54.235582 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="87320828-94da-481c-b903-e7d478e3df65" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 06 10:29:54 crc kubenswrapper[4824]: I1006 10:29:54.236651 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-d24cq" Oct 06 10:29:54 crc kubenswrapper[4824]: I1006 10:29:54.240562 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 06 10:29:54 crc kubenswrapper[4824]: I1006 10:29:54.240717 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 06 10:29:54 crc kubenswrapper[4824]: I1006 10:29:54.240739 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-f886w" Oct 06 10:29:54 crc kubenswrapper[4824]: I1006 10:29:54.241105 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 06 10:29:54 crc kubenswrapper[4824]: I1006 10:29:54.264089 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-d24cq"] Oct 06 10:29:54 crc kubenswrapper[4824]: I1006 10:29:54.319087 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/7dcfc0e7-18d6-4a54-9033-acf435eda511-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-d24cq\" (UID: \"7dcfc0e7-18d6-4a54-9033-acf435eda511\") " pod="openstack/ssh-known-hosts-edpm-deployment-d24cq" Oct 06 10:29:54 crc kubenswrapper[4824]: I1006 10:29:54.319207 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/7dcfc0e7-18d6-4a54-9033-acf435eda511-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-d24cq\" (UID: \"7dcfc0e7-18d6-4a54-9033-acf435eda511\") " pod="openstack/ssh-known-hosts-edpm-deployment-d24cq" Oct 06 10:29:54 crc kubenswrapper[4824]: I1006 10:29:54.319297 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-npjn5\" (UniqueName: \"kubernetes.io/projected/7dcfc0e7-18d6-4a54-9033-acf435eda511-kube-api-access-npjn5\") pod \"ssh-known-hosts-edpm-deployment-d24cq\" (UID: \"7dcfc0e7-18d6-4a54-9033-acf435eda511\") " pod="openstack/ssh-known-hosts-edpm-deployment-d24cq" Oct 06 10:29:54 crc kubenswrapper[4824]: I1006 10:29:54.421673 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-npjn5\" (UniqueName: \"kubernetes.io/projected/7dcfc0e7-18d6-4a54-9033-acf435eda511-kube-api-access-npjn5\") pod \"ssh-known-hosts-edpm-deployment-d24cq\" (UID: \"7dcfc0e7-18d6-4a54-9033-acf435eda511\") " pod="openstack/ssh-known-hosts-edpm-deployment-d24cq" Oct 06 10:29:54 crc kubenswrapper[4824]: I1006 10:29:54.422146 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/7dcfc0e7-18d6-4a54-9033-acf435eda511-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-d24cq\" (UID: \"7dcfc0e7-18d6-4a54-9033-acf435eda511\") " pod="openstack/ssh-known-hosts-edpm-deployment-d24cq" Oct 06 10:29:54 crc kubenswrapper[4824]: I1006 10:29:54.422464 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/7dcfc0e7-18d6-4a54-9033-acf435eda511-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-d24cq\" (UID: \"7dcfc0e7-18d6-4a54-9033-acf435eda511\") " pod="openstack/ssh-known-hosts-edpm-deployment-d24cq" Oct 06 10:29:54 crc kubenswrapper[4824]: I1006 10:29:54.427795 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/7dcfc0e7-18d6-4a54-9033-acf435eda511-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-d24cq\" (UID: \"7dcfc0e7-18d6-4a54-9033-acf435eda511\") " pod="openstack/ssh-known-hosts-edpm-deployment-d24cq" Oct 06 10:29:54 crc kubenswrapper[4824]: I1006 10:29:54.435371 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/7dcfc0e7-18d6-4a54-9033-acf435eda511-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-d24cq\" (UID: \"7dcfc0e7-18d6-4a54-9033-acf435eda511\") " pod="openstack/ssh-known-hosts-edpm-deployment-d24cq" Oct 06 10:29:54 crc kubenswrapper[4824]: I1006 10:29:54.446296 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-npjn5\" (UniqueName: \"kubernetes.io/projected/7dcfc0e7-18d6-4a54-9033-acf435eda511-kube-api-access-npjn5\") pod \"ssh-known-hosts-edpm-deployment-d24cq\" (UID: \"7dcfc0e7-18d6-4a54-9033-acf435eda511\") " pod="openstack/ssh-known-hosts-edpm-deployment-d24cq" Oct 06 10:29:54 crc kubenswrapper[4824]: I1006 10:29:54.566859 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-d24cq" Oct 06 10:29:55 crc kubenswrapper[4824]: I1006 10:29:55.118813 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-d24cq"] Oct 06 10:29:55 crc kubenswrapper[4824]: I1006 10:29:55.126945 4824 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 06 10:29:55 crc kubenswrapper[4824]: I1006 10:29:55.138282 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-d24cq" event={"ID":"7dcfc0e7-18d6-4a54-9033-acf435eda511","Type":"ContainerStarted","Data":"7232f1a18888efac80a18560e07c18718f0976b90b1d87de0abc6ff98cd14d9b"} Oct 06 10:29:57 crc kubenswrapper[4824]: I1006 10:29:57.172078 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-d24cq" event={"ID":"7dcfc0e7-18d6-4a54-9033-acf435eda511","Type":"ContainerStarted","Data":"04567b6b64348f7930057ed8df5ad38d5c002e917e376376fa0d1dd52e471c65"} Oct 06 10:29:57 crc kubenswrapper[4824]: I1006 10:29:57.202837 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-d24cq" podStartSLOduration=1.692329282 podStartE2EDuration="3.20281575s" podCreationTimestamp="2025-10-06 10:29:54 +0000 UTC" firstStartedPulling="2025-10-06 10:29:55.126651917 +0000 UTC m=+1904.491074798" lastFinishedPulling="2025-10-06 10:29:56.637138375 +0000 UTC m=+1906.001561266" observedRunningTime="2025-10-06 10:29:57.192575442 +0000 UTC m=+1906.556998303" watchObservedRunningTime="2025-10-06 10:29:57.20281575 +0000 UTC m=+1906.567238611" Oct 06 10:30:00 crc kubenswrapper[4824]: I1006 10:30:00.145837 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29329110-4hb5w"] Oct 06 10:30:00 crc kubenswrapper[4824]: I1006 10:30:00.147386 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29329110-4hb5w" Oct 06 10:30:00 crc kubenswrapper[4824]: I1006 10:30:00.149849 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 06 10:30:00 crc kubenswrapper[4824]: I1006 10:30:00.150766 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 06 10:30:00 crc kubenswrapper[4824]: I1006 10:30:00.160693 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29329110-4hb5w"] Oct 06 10:30:00 crc kubenswrapper[4824]: I1006 10:30:00.243137 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ea07beb1-3770-4bd3-85a1-898132a74838-config-volume\") pod \"collect-profiles-29329110-4hb5w\" (UID: \"ea07beb1-3770-4bd3-85a1-898132a74838\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329110-4hb5w" Oct 06 10:30:00 crc kubenswrapper[4824]: I1006 10:30:00.243233 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ea07beb1-3770-4bd3-85a1-898132a74838-secret-volume\") pod \"collect-profiles-29329110-4hb5w\" (UID: \"ea07beb1-3770-4bd3-85a1-898132a74838\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329110-4hb5w" Oct 06 10:30:00 crc kubenswrapper[4824]: I1006 10:30:00.243353 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2vlxg\" (UniqueName: \"kubernetes.io/projected/ea07beb1-3770-4bd3-85a1-898132a74838-kube-api-access-2vlxg\") pod \"collect-profiles-29329110-4hb5w\" (UID: \"ea07beb1-3770-4bd3-85a1-898132a74838\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329110-4hb5w" Oct 06 10:30:00 crc kubenswrapper[4824]: I1006 10:30:00.346138 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ea07beb1-3770-4bd3-85a1-898132a74838-config-volume\") pod \"collect-profiles-29329110-4hb5w\" (UID: \"ea07beb1-3770-4bd3-85a1-898132a74838\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329110-4hb5w" Oct 06 10:30:00 crc kubenswrapper[4824]: I1006 10:30:00.346269 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ea07beb1-3770-4bd3-85a1-898132a74838-secret-volume\") pod \"collect-profiles-29329110-4hb5w\" (UID: \"ea07beb1-3770-4bd3-85a1-898132a74838\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329110-4hb5w" Oct 06 10:30:00 crc kubenswrapper[4824]: I1006 10:30:00.346505 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2vlxg\" (UniqueName: \"kubernetes.io/projected/ea07beb1-3770-4bd3-85a1-898132a74838-kube-api-access-2vlxg\") pod \"collect-profiles-29329110-4hb5w\" (UID: \"ea07beb1-3770-4bd3-85a1-898132a74838\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329110-4hb5w" Oct 06 10:30:00 crc kubenswrapper[4824]: I1006 10:30:00.348689 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ea07beb1-3770-4bd3-85a1-898132a74838-config-volume\") pod \"collect-profiles-29329110-4hb5w\" (UID: \"ea07beb1-3770-4bd3-85a1-898132a74838\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329110-4hb5w" Oct 06 10:30:00 crc kubenswrapper[4824]: I1006 10:30:00.353305 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ea07beb1-3770-4bd3-85a1-898132a74838-secret-volume\") pod \"collect-profiles-29329110-4hb5w\" (UID: \"ea07beb1-3770-4bd3-85a1-898132a74838\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329110-4hb5w" Oct 06 10:30:00 crc kubenswrapper[4824]: I1006 10:30:00.384302 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2vlxg\" (UniqueName: \"kubernetes.io/projected/ea07beb1-3770-4bd3-85a1-898132a74838-kube-api-access-2vlxg\") pod \"collect-profiles-29329110-4hb5w\" (UID: \"ea07beb1-3770-4bd3-85a1-898132a74838\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329110-4hb5w" Oct 06 10:30:00 crc kubenswrapper[4824]: I1006 10:30:00.465388 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29329110-4hb5w" Oct 06 10:30:00 crc kubenswrapper[4824]: I1006 10:30:00.938504 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29329110-4hb5w"] Oct 06 10:30:00 crc kubenswrapper[4824]: W1006 10:30:00.943846 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podea07beb1_3770_4bd3_85a1_898132a74838.slice/crio-5bd3b72db00a4193f6f3ba1c5984a41ab01d8dbaceed2b43a066150774268fea WatchSource:0}: Error finding container 5bd3b72db00a4193f6f3ba1c5984a41ab01d8dbaceed2b43a066150774268fea: Status 404 returned error can't find the container with id 5bd3b72db00a4193f6f3ba1c5984a41ab01d8dbaceed2b43a066150774268fea Oct 06 10:30:01 crc kubenswrapper[4824]: I1006 10:30:01.216628 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29329110-4hb5w" event={"ID":"ea07beb1-3770-4bd3-85a1-898132a74838","Type":"ContainerStarted","Data":"badcee0d59c683b35600cf58130942cc62f79e43db20fae72297b3d329477d47"} Oct 06 10:30:01 crc kubenswrapper[4824]: I1006 10:30:01.217627 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29329110-4hb5w" event={"ID":"ea07beb1-3770-4bd3-85a1-898132a74838","Type":"ContainerStarted","Data":"5bd3b72db00a4193f6f3ba1c5984a41ab01d8dbaceed2b43a066150774268fea"} Oct 06 10:30:01 crc kubenswrapper[4824]: I1006 10:30:01.235021 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29329110-4hb5w" podStartSLOduration=1.235007328 podStartE2EDuration="1.235007328s" podCreationTimestamp="2025-10-06 10:30:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:30:01.234781712 +0000 UTC m=+1910.599204573" watchObservedRunningTime="2025-10-06 10:30:01.235007328 +0000 UTC m=+1910.599430189" Oct 06 10:30:02 crc kubenswrapper[4824]: I1006 10:30:02.236024 4824 generic.go:334] "Generic (PLEG): container finished" podID="ea07beb1-3770-4bd3-85a1-898132a74838" containerID="badcee0d59c683b35600cf58130942cc62f79e43db20fae72297b3d329477d47" exitCode=0 Oct 06 10:30:02 crc kubenswrapper[4824]: I1006 10:30:02.236076 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29329110-4hb5w" event={"ID":"ea07beb1-3770-4bd3-85a1-898132a74838","Type":"ContainerDied","Data":"badcee0d59c683b35600cf58130942cc62f79e43db20fae72297b3d329477d47"} Oct 06 10:30:03 crc kubenswrapper[4824]: I1006 10:30:03.274739 4824 scope.go:117] "RemoveContainer" containerID="9de43c7f48ed9ad071a24adf937829ab92276bd505bb72d5b89bac4922aac3b0" Oct 06 10:30:03 crc kubenswrapper[4824]: E1006 10:30:03.275262 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-khgzw_openshift-machine-config-operator(e1140c8c-93f1-4fce-8c06-a69561ff8a6d)\"" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" Oct 06 10:30:03 crc kubenswrapper[4824]: I1006 10:30:03.619664 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29329110-4hb5w" Oct 06 10:30:03 crc kubenswrapper[4824]: I1006 10:30:03.710321 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2vlxg\" (UniqueName: \"kubernetes.io/projected/ea07beb1-3770-4bd3-85a1-898132a74838-kube-api-access-2vlxg\") pod \"ea07beb1-3770-4bd3-85a1-898132a74838\" (UID: \"ea07beb1-3770-4bd3-85a1-898132a74838\") " Oct 06 10:30:03 crc kubenswrapper[4824]: I1006 10:30:03.710408 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ea07beb1-3770-4bd3-85a1-898132a74838-config-volume\") pod \"ea07beb1-3770-4bd3-85a1-898132a74838\" (UID: \"ea07beb1-3770-4bd3-85a1-898132a74838\") " Oct 06 10:30:03 crc kubenswrapper[4824]: I1006 10:30:03.710501 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ea07beb1-3770-4bd3-85a1-898132a74838-secret-volume\") pod \"ea07beb1-3770-4bd3-85a1-898132a74838\" (UID: \"ea07beb1-3770-4bd3-85a1-898132a74838\") " Oct 06 10:30:03 crc kubenswrapper[4824]: I1006 10:30:03.711757 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ea07beb1-3770-4bd3-85a1-898132a74838-config-volume" (OuterVolumeSpecName: "config-volume") pod "ea07beb1-3770-4bd3-85a1-898132a74838" (UID: "ea07beb1-3770-4bd3-85a1-898132a74838"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:30:03 crc kubenswrapper[4824]: I1006 10:30:03.719345 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ea07beb1-3770-4bd3-85a1-898132a74838-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "ea07beb1-3770-4bd3-85a1-898132a74838" (UID: "ea07beb1-3770-4bd3-85a1-898132a74838"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:30:03 crc kubenswrapper[4824]: I1006 10:30:03.722599 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ea07beb1-3770-4bd3-85a1-898132a74838-kube-api-access-2vlxg" (OuterVolumeSpecName: "kube-api-access-2vlxg") pod "ea07beb1-3770-4bd3-85a1-898132a74838" (UID: "ea07beb1-3770-4bd3-85a1-898132a74838"). InnerVolumeSpecName "kube-api-access-2vlxg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:30:03 crc kubenswrapper[4824]: I1006 10:30:03.813863 4824 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ea07beb1-3770-4bd3-85a1-898132a74838-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 06 10:30:03 crc kubenswrapper[4824]: I1006 10:30:03.813920 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2vlxg\" (UniqueName: \"kubernetes.io/projected/ea07beb1-3770-4bd3-85a1-898132a74838-kube-api-access-2vlxg\") on node \"crc\" DevicePath \"\"" Oct 06 10:30:03 crc kubenswrapper[4824]: I1006 10:30:03.813937 4824 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ea07beb1-3770-4bd3-85a1-898132a74838-config-volume\") on node \"crc\" DevicePath \"\"" Oct 06 10:30:04 crc kubenswrapper[4824]: I1006 10:30:04.265839 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29329110-4hb5w" Oct 06 10:30:04 crc kubenswrapper[4824]: I1006 10:30:04.265829 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29329110-4hb5w" event={"ID":"ea07beb1-3770-4bd3-85a1-898132a74838","Type":"ContainerDied","Data":"5bd3b72db00a4193f6f3ba1c5984a41ab01d8dbaceed2b43a066150774268fea"} Oct 06 10:30:04 crc kubenswrapper[4824]: I1006 10:30:04.266639 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5bd3b72db00a4193f6f3ba1c5984a41ab01d8dbaceed2b43a066150774268fea" Oct 06 10:30:04 crc kubenswrapper[4824]: I1006 10:30:04.268110 4824 generic.go:334] "Generic (PLEG): container finished" podID="7dcfc0e7-18d6-4a54-9033-acf435eda511" containerID="04567b6b64348f7930057ed8df5ad38d5c002e917e376376fa0d1dd52e471c65" exitCode=0 Oct 06 10:30:04 crc kubenswrapper[4824]: I1006 10:30:04.268183 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-d24cq" event={"ID":"7dcfc0e7-18d6-4a54-9033-acf435eda511","Type":"ContainerDied","Data":"04567b6b64348f7930057ed8df5ad38d5c002e917e376376fa0d1dd52e471c65"} Oct 06 10:30:05 crc kubenswrapper[4824]: I1006 10:30:05.785828 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-d24cq" Oct 06 10:30:05 crc kubenswrapper[4824]: I1006 10:30:05.860332 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/7dcfc0e7-18d6-4a54-9033-acf435eda511-inventory-0\") pod \"7dcfc0e7-18d6-4a54-9033-acf435eda511\" (UID: \"7dcfc0e7-18d6-4a54-9033-acf435eda511\") " Oct 06 10:30:05 crc kubenswrapper[4824]: I1006 10:30:05.860436 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-npjn5\" (UniqueName: \"kubernetes.io/projected/7dcfc0e7-18d6-4a54-9033-acf435eda511-kube-api-access-npjn5\") pod \"7dcfc0e7-18d6-4a54-9033-acf435eda511\" (UID: \"7dcfc0e7-18d6-4a54-9033-acf435eda511\") " Oct 06 10:30:05 crc kubenswrapper[4824]: I1006 10:30:05.860493 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/7dcfc0e7-18d6-4a54-9033-acf435eda511-ssh-key-openstack-edpm-ipam\") pod \"7dcfc0e7-18d6-4a54-9033-acf435eda511\" (UID: \"7dcfc0e7-18d6-4a54-9033-acf435eda511\") " Oct 06 10:30:05 crc kubenswrapper[4824]: I1006 10:30:05.868748 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7dcfc0e7-18d6-4a54-9033-acf435eda511-kube-api-access-npjn5" (OuterVolumeSpecName: "kube-api-access-npjn5") pod "7dcfc0e7-18d6-4a54-9033-acf435eda511" (UID: "7dcfc0e7-18d6-4a54-9033-acf435eda511"). InnerVolumeSpecName "kube-api-access-npjn5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:30:05 crc kubenswrapper[4824]: I1006 10:30:05.904281 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7dcfc0e7-18d6-4a54-9033-acf435eda511-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "7dcfc0e7-18d6-4a54-9033-acf435eda511" (UID: "7dcfc0e7-18d6-4a54-9033-acf435eda511"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:30:05 crc kubenswrapper[4824]: I1006 10:30:05.905702 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7dcfc0e7-18d6-4a54-9033-acf435eda511-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "7dcfc0e7-18d6-4a54-9033-acf435eda511" (UID: "7dcfc0e7-18d6-4a54-9033-acf435eda511"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:30:05 crc kubenswrapper[4824]: I1006 10:30:05.963308 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-npjn5\" (UniqueName: \"kubernetes.io/projected/7dcfc0e7-18d6-4a54-9033-acf435eda511-kube-api-access-npjn5\") on node \"crc\" DevicePath \"\"" Oct 06 10:30:05 crc kubenswrapper[4824]: I1006 10:30:05.963365 4824 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/7dcfc0e7-18d6-4a54-9033-acf435eda511-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Oct 06 10:30:05 crc kubenswrapper[4824]: I1006 10:30:05.963376 4824 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/7dcfc0e7-18d6-4a54-9033-acf435eda511-inventory-0\") on node \"crc\" DevicePath \"\"" Oct 06 10:30:06 crc kubenswrapper[4824]: I1006 10:30:06.292471 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-d24cq" event={"ID":"7dcfc0e7-18d6-4a54-9033-acf435eda511","Type":"ContainerDied","Data":"7232f1a18888efac80a18560e07c18718f0976b90b1d87de0abc6ff98cd14d9b"} Oct 06 10:30:06 crc kubenswrapper[4824]: I1006 10:30:06.292517 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7232f1a18888efac80a18560e07c18718f0976b90b1d87de0abc6ff98cd14d9b" Oct 06 10:30:06 crc kubenswrapper[4824]: I1006 10:30:06.292523 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-d24cq" Oct 06 10:30:06 crc kubenswrapper[4824]: I1006 10:30:06.371390 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-vwg8r"] Oct 06 10:30:06 crc kubenswrapper[4824]: E1006 10:30:06.371926 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7dcfc0e7-18d6-4a54-9033-acf435eda511" containerName="ssh-known-hosts-edpm-deployment" Oct 06 10:30:06 crc kubenswrapper[4824]: I1006 10:30:06.371951 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="7dcfc0e7-18d6-4a54-9033-acf435eda511" containerName="ssh-known-hosts-edpm-deployment" Oct 06 10:30:06 crc kubenswrapper[4824]: E1006 10:30:06.371995 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea07beb1-3770-4bd3-85a1-898132a74838" containerName="collect-profiles" Oct 06 10:30:06 crc kubenswrapper[4824]: I1006 10:30:06.372006 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea07beb1-3770-4bd3-85a1-898132a74838" containerName="collect-profiles" Oct 06 10:30:06 crc kubenswrapper[4824]: I1006 10:30:06.372229 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="ea07beb1-3770-4bd3-85a1-898132a74838" containerName="collect-profiles" Oct 06 10:30:06 crc kubenswrapper[4824]: I1006 10:30:06.372268 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="7dcfc0e7-18d6-4a54-9033-acf435eda511" containerName="ssh-known-hosts-edpm-deployment" Oct 06 10:30:06 crc kubenswrapper[4824]: I1006 10:30:06.373602 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-vwg8r" Oct 06 10:30:06 crc kubenswrapper[4824]: I1006 10:30:06.375764 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 06 10:30:06 crc kubenswrapper[4824]: I1006 10:30:06.376136 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 06 10:30:06 crc kubenswrapper[4824]: I1006 10:30:06.376204 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 06 10:30:06 crc kubenswrapper[4824]: I1006 10:30:06.377038 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-f886w" Oct 06 10:30:06 crc kubenswrapper[4824]: I1006 10:30:06.386019 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-vwg8r"] Oct 06 10:30:06 crc kubenswrapper[4824]: I1006 10:30:06.473478 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/626b5d8f-93f2-4bbc-add6-a530ddf9a6fa-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-vwg8r\" (UID: \"626b5d8f-93f2-4bbc-add6-a530ddf9a6fa\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-vwg8r" Oct 06 10:30:06 crc kubenswrapper[4824]: I1006 10:30:06.474063 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-576tl\" (UniqueName: \"kubernetes.io/projected/626b5d8f-93f2-4bbc-add6-a530ddf9a6fa-kube-api-access-576tl\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-vwg8r\" (UID: \"626b5d8f-93f2-4bbc-add6-a530ddf9a6fa\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-vwg8r" Oct 06 10:30:06 crc kubenswrapper[4824]: I1006 10:30:06.474114 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/626b5d8f-93f2-4bbc-add6-a530ddf9a6fa-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-vwg8r\" (UID: \"626b5d8f-93f2-4bbc-add6-a530ddf9a6fa\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-vwg8r" Oct 06 10:30:06 crc kubenswrapper[4824]: I1006 10:30:06.576183 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-576tl\" (UniqueName: \"kubernetes.io/projected/626b5d8f-93f2-4bbc-add6-a530ddf9a6fa-kube-api-access-576tl\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-vwg8r\" (UID: \"626b5d8f-93f2-4bbc-add6-a530ddf9a6fa\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-vwg8r" Oct 06 10:30:06 crc kubenswrapper[4824]: I1006 10:30:06.576226 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/626b5d8f-93f2-4bbc-add6-a530ddf9a6fa-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-vwg8r\" (UID: \"626b5d8f-93f2-4bbc-add6-a530ddf9a6fa\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-vwg8r" Oct 06 10:30:06 crc kubenswrapper[4824]: I1006 10:30:06.576274 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/626b5d8f-93f2-4bbc-add6-a530ddf9a6fa-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-vwg8r\" (UID: \"626b5d8f-93f2-4bbc-add6-a530ddf9a6fa\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-vwg8r" Oct 06 10:30:06 crc kubenswrapper[4824]: I1006 10:30:06.580318 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/626b5d8f-93f2-4bbc-add6-a530ddf9a6fa-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-vwg8r\" (UID: \"626b5d8f-93f2-4bbc-add6-a530ddf9a6fa\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-vwg8r" Oct 06 10:30:06 crc kubenswrapper[4824]: I1006 10:30:06.580318 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/626b5d8f-93f2-4bbc-add6-a530ddf9a6fa-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-vwg8r\" (UID: \"626b5d8f-93f2-4bbc-add6-a530ddf9a6fa\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-vwg8r" Oct 06 10:30:06 crc kubenswrapper[4824]: I1006 10:30:06.597427 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-576tl\" (UniqueName: \"kubernetes.io/projected/626b5d8f-93f2-4bbc-add6-a530ddf9a6fa-kube-api-access-576tl\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-vwg8r\" (UID: \"626b5d8f-93f2-4bbc-add6-a530ddf9a6fa\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-vwg8r" Oct 06 10:30:06 crc kubenswrapper[4824]: I1006 10:30:06.691787 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-vwg8r" Oct 06 10:30:07 crc kubenswrapper[4824]: I1006 10:30:07.192603 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-vwg8r"] Oct 06 10:30:07 crc kubenswrapper[4824]: I1006 10:30:07.302657 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-vwg8r" event={"ID":"626b5d8f-93f2-4bbc-add6-a530ddf9a6fa","Type":"ContainerStarted","Data":"3c859b9129ca48e09e14fd03c232504361780c575b202f243d6c4d213a200522"} Oct 06 10:30:08 crc kubenswrapper[4824]: I1006 10:30:08.315813 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-vwg8r" event={"ID":"626b5d8f-93f2-4bbc-add6-a530ddf9a6fa","Type":"ContainerStarted","Data":"e1e60e8722f7333a73167bb0de64363bbf5fc71ba9bb299ee77bfc0c87251444"} Oct 06 10:30:08 crc kubenswrapper[4824]: I1006 10:30:08.340381 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-vwg8r" podStartSLOduration=1.696282498 podStartE2EDuration="2.340359878s" podCreationTimestamp="2025-10-06 10:30:06 +0000 UTC" firstStartedPulling="2025-10-06 10:30:07.204233667 +0000 UTC m=+1916.568656568" lastFinishedPulling="2025-10-06 10:30:07.848311047 +0000 UTC m=+1917.212733948" observedRunningTime="2025-10-06 10:30:08.334920141 +0000 UTC m=+1917.699343012" watchObservedRunningTime="2025-10-06 10:30:08.340359878 +0000 UTC m=+1917.704782739" Oct 06 10:30:14 crc kubenswrapper[4824]: I1006 10:30:14.274452 4824 scope.go:117] "RemoveContainer" containerID="9de43c7f48ed9ad071a24adf937829ab92276bd505bb72d5b89bac4922aac3b0" Oct 06 10:30:14 crc kubenswrapper[4824]: E1006 10:30:14.275235 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-khgzw_openshift-machine-config-operator(e1140c8c-93f1-4fce-8c06-a69561ff8a6d)\"" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" Oct 06 10:30:17 crc kubenswrapper[4824]: I1006 10:30:17.405772 4824 generic.go:334] "Generic (PLEG): container finished" podID="626b5d8f-93f2-4bbc-add6-a530ddf9a6fa" containerID="e1e60e8722f7333a73167bb0de64363bbf5fc71ba9bb299ee77bfc0c87251444" exitCode=0 Oct 06 10:30:17 crc kubenswrapper[4824]: I1006 10:30:17.405909 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-vwg8r" event={"ID":"626b5d8f-93f2-4bbc-add6-a530ddf9a6fa","Type":"ContainerDied","Data":"e1e60e8722f7333a73167bb0de64363bbf5fc71ba9bb299ee77bfc0c87251444"} Oct 06 10:30:18 crc kubenswrapper[4824]: I1006 10:30:18.836632 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-vwg8r" Oct 06 10:30:18 crc kubenswrapper[4824]: I1006 10:30:18.943426 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/626b5d8f-93f2-4bbc-add6-a530ddf9a6fa-ssh-key\") pod \"626b5d8f-93f2-4bbc-add6-a530ddf9a6fa\" (UID: \"626b5d8f-93f2-4bbc-add6-a530ddf9a6fa\") " Oct 06 10:30:18 crc kubenswrapper[4824]: I1006 10:30:18.943661 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-576tl\" (UniqueName: \"kubernetes.io/projected/626b5d8f-93f2-4bbc-add6-a530ddf9a6fa-kube-api-access-576tl\") pod \"626b5d8f-93f2-4bbc-add6-a530ddf9a6fa\" (UID: \"626b5d8f-93f2-4bbc-add6-a530ddf9a6fa\") " Oct 06 10:30:18 crc kubenswrapper[4824]: I1006 10:30:18.943698 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/626b5d8f-93f2-4bbc-add6-a530ddf9a6fa-inventory\") pod \"626b5d8f-93f2-4bbc-add6-a530ddf9a6fa\" (UID: \"626b5d8f-93f2-4bbc-add6-a530ddf9a6fa\") " Oct 06 10:30:18 crc kubenswrapper[4824]: I1006 10:30:18.950857 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/626b5d8f-93f2-4bbc-add6-a530ddf9a6fa-kube-api-access-576tl" (OuterVolumeSpecName: "kube-api-access-576tl") pod "626b5d8f-93f2-4bbc-add6-a530ddf9a6fa" (UID: "626b5d8f-93f2-4bbc-add6-a530ddf9a6fa"). InnerVolumeSpecName "kube-api-access-576tl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:30:18 crc kubenswrapper[4824]: I1006 10:30:18.977196 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/626b5d8f-93f2-4bbc-add6-a530ddf9a6fa-inventory" (OuterVolumeSpecName: "inventory") pod "626b5d8f-93f2-4bbc-add6-a530ddf9a6fa" (UID: "626b5d8f-93f2-4bbc-add6-a530ddf9a6fa"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:30:18 crc kubenswrapper[4824]: I1006 10:30:18.979480 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/626b5d8f-93f2-4bbc-add6-a530ddf9a6fa-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "626b5d8f-93f2-4bbc-add6-a530ddf9a6fa" (UID: "626b5d8f-93f2-4bbc-add6-a530ddf9a6fa"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:30:19 crc kubenswrapper[4824]: I1006 10:30:19.045727 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-576tl\" (UniqueName: \"kubernetes.io/projected/626b5d8f-93f2-4bbc-add6-a530ddf9a6fa-kube-api-access-576tl\") on node \"crc\" DevicePath \"\"" Oct 06 10:30:19 crc kubenswrapper[4824]: I1006 10:30:19.045852 4824 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/626b5d8f-93f2-4bbc-add6-a530ddf9a6fa-inventory\") on node \"crc\" DevicePath \"\"" Oct 06 10:30:19 crc kubenswrapper[4824]: I1006 10:30:19.045906 4824 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/626b5d8f-93f2-4bbc-add6-a530ddf9a6fa-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 06 10:30:19 crc kubenswrapper[4824]: I1006 10:30:19.426293 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-vwg8r" event={"ID":"626b5d8f-93f2-4bbc-add6-a530ddf9a6fa","Type":"ContainerDied","Data":"3c859b9129ca48e09e14fd03c232504361780c575b202f243d6c4d213a200522"} Oct 06 10:30:19 crc kubenswrapper[4824]: I1006 10:30:19.426641 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3c859b9129ca48e09e14fd03c232504361780c575b202f243d6c4d213a200522" Oct 06 10:30:19 crc kubenswrapper[4824]: I1006 10:30:19.426381 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-vwg8r" Oct 06 10:30:19 crc kubenswrapper[4824]: I1006 10:30:19.537221 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-97h9g"] Oct 06 10:30:19 crc kubenswrapper[4824]: E1006 10:30:19.538050 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="626b5d8f-93f2-4bbc-add6-a530ddf9a6fa" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Oct 06 10:30:19 crc kubenswrapper[4824]: I1006 10:30:19.538174 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="626b5d8f-93f2-4bbc-add6-a530ddf9a6fa" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Oct 06 10:30:19 crc kubenswrapper[4824]: I1006 10:30:19.539176 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="626b5d8f-93f2-4bbc-add6-a530ddf9a6fa" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Oct 06 10:30:19 crc kubenswrapper[4824]: I1006 10:30:19.540228 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-97h9g" Oct 06 10:30:19 crc kubenswrapper[4824]: I1006 10:30:19.542514 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 06 10:30:19 crc kubenswrapper[4824]: I1006 10:30:19.542579 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 06 10:30:19 crc kubenswrapper[4824]: I1006 10:30:19.542648 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 06 10:30:19 crc kubenswrapper[4824]: I1006 10:30:19.547105 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-f886w" Oct 06 10:30:19 crc kubenswrapper[4824]: I1006 10:30:19.547432 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-97h9g"] Oct 06 10:30:19 crc kubenswrapper[4824]: I1006 10:30:19.659658 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/073c5011-b49a-4900-b00b-488164295d4d-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-97h9g\" (UID: \"073c5011-b49a-4900-b00b-488164295d4d\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-97h9g" Oct 06 10:30:19 crc kubenswrapper[4824]: I1006 10:30:19.659702 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/073c5011-b49a-4900-b00b-488164295d4d-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-97h9g\" (UID: \"073c5011-b49a-4900-b00b-488164295d4d\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-97h9g" Oct 06 10:30:19 crc kubenswrapper[4824]: I1006 10:30:19.659800 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wz252\" (UniqueName: \"kubernetes.io/projected/073c5011-b49a-4900-b00b-488164295d4d-kube-api-access-wz252\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-97h9g\" (UID: \"073c5011-b49a-4900-b00b-488164295d4d\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-97h9g" Oct 06 10:30:19 crc kubenswrapper[4824]: I1006 10:30:19.761849 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/073c5011-b49a-4900-b00b-488164295d4d-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-97h9g\" (UID: \"073c5011-b49a-4900-b00b-488164295d4d\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-97h9g" Oct 06 10:30:19 crc kubenswrapper[4824]: I1006 10:30:19.761910 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/073c5011-b49a-4900-b00b-488164295d4d-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-97h9g\" (UID: \"073c5011-b49a-4900-b00b-488164295d4d\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-97h9g" Oct 06 10:30:19 crc kubenswrapper[4824]: I1006 10:30:19.762069 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wz252\" (UniqueName: \"kubernetes.io/projected/073c5011-b49a-4900-b00b-488164295d4d-kube-api-access-wz252\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-97h9g\" (UID: \"073c5011-b49a-4900-b00b-488164295d4d\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-97h9g" Oct 06 10:30:19 crc kubenswrapper[4824]: I1006 10:30:19.771067 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/073c5011-b49a-4900-b00b-488164295d4d-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-97h9g\" (UID: \"073c5011-b49a-4900-b00b-488164295d4d\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-97h9g" Oct 06 10:30:19 crc kubenswrapper[4824]: I1006 10:30:19.777218 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/073c5011-b49a-4900-b00b-488164295d4d-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-97h9g\" (UID: \"073c5011-b49a-4900-b00b-488164295d4d\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-97h9g" Oct 06 10:30:19 crc kubenswrapper[4824]: I1006 10:30:19.782807 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wz252\" (UniqueName: \"kubernetes.io/projected/073c5011-b49a-4900-b00b-488164295d4d-kube-api-access-wz252\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-97h9g\" (UID: \"073c5011-b49a-4900-b00b-488164295d4d\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-97h9g" Oct 06 10:30:19 crc kubenswrapper[4824]: I1006 10:30:19.860501 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-97h9g" Oct 06 10:30:20 crc kubenswrapper[4824]: I1006 10:30:20.440184 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-97h9g"] Oct 06 10:30:21 crc kubenswrapper[4824]: I1006 10:30:21.454786 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-97h9g" event={"ID":"073c5011-b49a-4900-b00b-488164295d4d","Type":"ContainerStarted","Data":"c00a9ff14ec1010a848efdeb976144fa7526dcf98f6035e109a68b2066804c34"} Oct 06 10:30:21 crc kubenswrapper[4824]: I1006 10:30:21.455229 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-97h9g" event={"ID":"073c5011-b49a-4900-b00b-488164295d4d","Type":"ContainerStarted","Data":"14dde3109c01386de8ac02e6cc8464a176b223302d777de39736f2a0c5106bd9"} Oct 06 10:30:21 crc kubenswrapper[4824]: I1006 10:30:21.471606 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-97h9g" podStartSLOduration=2.042598978 podStartE2EDuration="2.471583349s" podCreationTimestamp="2025-10-06 10:30:19 +0000 UTC" firstStartedPulling="2025-10-06 10:30:20.439484322 +0000 UTC m=+1929.803907193" lastFinishedPulling="2025-10-06 10:30:20.868468693 +0000 UTC m=+1930.232891564" observedRunningTime="2025-10-06 10:30:21.469940541 +0000 UTC m=+1930.834363402" watchObservedRunningTime="2025-10-06 10:30:21.471583349 +0000 UTC m=+1930.836006220" Oct 06 10:30:27 crc kubenswrapper[4824]: I1006 10:30:27.287037 4824 scope.go:117] "RemoveContainer" containerID="9de43c7f48ed9ad071a24adf937829ab92276bd505bb72d5b89bac4922aac3b0" Oct 06 10:30:27 crc kubenswrapper[4824]: E1006 10:30:27.288585 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-khgzw_openshift-machine-config-operator(e1140c8c-93f1-4fce-8c06-a69561ff8a6d)\"" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" Oct 06 10:30:31 crc kubenswrapper[4824]: I1006 10:30:31.571086 4824 generic.go:334] "Generic (PLEG): container finished" podID="073c5011-b49a-4900-b00b-488164295d4d" containerID="c00a9ff14ec1010a848efdeb976144fa7526dcf98f6035e109a68b2066804c34" exitCode=0 Oct 06 10:30:31 crc kubenswrapper[4824]: I1006 10:30:31.571189 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-97h9g" event={"ID":"073c5011-b49a-4900-b00b-488164295d4d","Type":"ContainerDied","Data":"c00a9ff14ec1010a848efdeb976144fa7526dcf98f6035e109a68b2066804c34"} Oct 06 10:30:33 crc kubenswrapper[4824]: I1006 10:30:33.008972 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-97h9g" Oct 06 10:30:33 crc kubenswrapper[4824]: I1006 10:30:33.057738 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/073c5011-b49a-4900-b00b-488164295d4d-ssh-key\") pod \"073c5011-b49a-4900-b00b-488164295d4d\" (UID: \"073c5011-b49a-4900-b00b-488164295d4d\") " Oct 06 10:30:33 crc kubenswrapper[4824]: I1006 10:30:33.057878 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/073c5011-b49a-4900-b00b-488164295d4d-inventory\") pod \"073c5011-b49a-4900-b00b-488164295d4d\" (UID: \"073c5011-b49a-4900-b00b-488164295d4d\") " Oct 06 10:30:33 crc kubenswrapper[4824]: I1006 10:30:33.058001 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wz252\" (UniqueName: \"kubernetes.io/projected/073c5011-b49a-4900-b00b-488164295d4d-kube-api-access-wz252\") pod \"073c5011-b49a-4900-b00b-488164295d4d\" (UID: \"073c5011-b49a-4900-b00b-488164295d4d\") " Oct 06 10:30:33 crc kubenswrapper[4824]: I1006 10:30:33.066683 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/073c5011-b49a-4900-b00b-488164295d4d-kube-api-access-wz252" (OuterVolumeSpecName: "kube-api-access-wz252") pod "073c5011-b49a-4900-b00b-488164295d4d" (UID: "073c5011-b49a-4900-b00b-488164295d4d"). InnerVolumeSpecName "kube-api-access-wz252". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:30:33 crc kubenswrapper[4824]: I1006 10:30:33.104771 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/073c5011-b49a-4900-b00b-488164295d4d-inventory" (OuterVolumeSpecName: "inventory") pod "073c5011-b49a-4900-b00b-488164295d4d" (UID: "073c5011-b49a-4900-b00b-488164295d4d"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:30:33 crc kubenswrapper[4824]: I1006 10:30:33.105197 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/073c5011-b49a-4900-b00b-488164295d4d-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "073c5011-b49a-4900-b00b-488164295d4d" (UID: "073c5011-b49a-4900-b00b-488164295d4d"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:30:33 crc kubenswrapper[4824]: I1006 10:30:33.161116 4824 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/073c5011-b49a-4900-b00b-488164295d4d-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 06 10:30:33 crc kubenswrapper[4824]: I1006 10:30:33.161193 4824 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/073c5011-b49a-4900-b00b-488164295d4d-inventory\") on node \"crc\" DevicePath \"\"" Oct 06 10:30:33 crc kubenswrapper[4824]: I1006 10:30:33.161208 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wz252\" (UniqueName: \"kubernetes.io/projected/073c5011-b49a-4900-b00b-488164295d4d-kube-api-access-wz252\") on node \"crc\" DevicePath \"\"" Oct 06 10:30:33 crc kubenswrapper[4824]: I1006 10:30:33.594485 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-97h9g" event={"ID":"073c5011-b49a-4900-b00b-488164295d4d","Type":"ContainerDied","Data":"14dde3109c01386de8ac02e6cc8464a176b223302d777de39736f2a0c5106bd9"} Oct 06 10:30:33 crc kubenswrapper[4824]: I1006 10:30:33.594547 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="14dde3109c01386de8ac02e6cc8464a176b223302d777de39736f2a0c5106bd9" Oct 06 10:30:33 crc kubenswrapper[4824]: I1006 10:30:33.594600 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-97h9g" Oct 06 10:30:33 crc kubenswrapper[4824]: I1006 10:30:33.750967 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jqgjd"] Oct 06 10:30:33 crc kubenswrapper[4824]: E1006 10:30:33.751413 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="073c5011-b49a-4900-b00b-488164295d4d" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Oct 06 10:30:33 crc kubenswrapper[4824]: I1006 10:30:33.751428 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="073c5011-b49a-4900-b00b-488164295d4d" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Oct 06 10:30:33 crc kubenswrapper[4824]: I1006 10:30:33.751650 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="073c5011-b49a-4900-b00b-488164295d4d" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Oct 06 10:30:33 crc kubenswrapper[4824]: I1006 10:30:33.752299 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jqgjd" Oct 06 10:30:33 crc kubenswrapper[4824]: I1006 10:30:33.755308 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Oct 06 10:30:33 crc kubenswrapper[4824]: I1006 10:30:33.760372 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 06 10:30:33 crc kubenswrapper[4824]: I1006 10:30:33.760517 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Oct 06 10:30:33 crc kubenswrapper[4824]: I1006 10:30:33.760628 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-f886w" Oct 06 10:30:33 crc kubenswrapper[4824]: I1006 10:30:33.760735 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-default-certs-0" Oct 06 10:30:33 crc kubenswrapper[4824]: I1006 10:30:33.760894 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 06 10:30:33 crc kubenswrapper[4824]: I1006 10:30:33.772547 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 06 10:30:33 crc kubenswrapper[4824]: I1006 10:30:33.772784 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Oct 06 10:30:33 crc kubenswrapper[4824]: I1006 10:30:33.776485 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jqgjd"] Oct 06 10:30:33 crc kubenswrapper[4824]: I1006 10:30:33.879872 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54e35b51-aeb1-4e2f-9be0-11230cb89fbe-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jqgjd\" (UID: \"54e35b51-aeb1-4e2f-9be0-11230cb89fbe\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jqgjd" Oct 06 10:30:33 crc kubenswrapper[4824]: I1006 10:30:33.879930 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54e35b51-aeb1-4e2f-9be0-11230cb89fbe-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jqgjd\" (UID: \"54e35b51-aeb1-4e2f-9be0-11230cb89fbe\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jqgjd" Oct 06 10:30:33 crc kubenswrapper[4824]: I1006 10:30:33.879948 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54e35b51-aeb1-4e2f-9be0-11230cb89fbe-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jqgjd\" (UID: \"54e35b51-aeb1-4e2f-9be0-11230cb89fbe\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jqgjd" Oct 06 10:30:33 crc kubenswrapper[4824]: I1006 10:30:33.879990 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54e35b51-aeb1-4e2f-9be0-11230cb89fbe-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jqgjd\" (UID: \"54e35b51-aeb1-4e2f-9be0-11230cb89fbe\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jqgjd" Oct 06 10:30:33 crc kubenswrapper[4824]: I1006 10:30:33.880032 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dwrrn\" (UniqueName: \"kubernetes.io/projected/54e35b51-aeb1-4e2f-9be0-11230cb89fbe-kube-api-access-dwrrn\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jqgjd\" (UID: \"54e35b51-aeb1-4e2f-9be0-11230cb89fbe\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jqgjd" Oct 06 10:30:33 crc kubenswrapper[4824]: I1006 10:30:33.880051 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54e35b51-aeb1-4e2f-9be0-11230cb89fbe-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jqgjd\" (UID: \"54e35b51-aeb1-4e2f-9be0-11230cb89fbe\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jqgjd" Oct 06 10:30:33 crc kubenswrapper[4824]: I1006 10:30:33.880082 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/54e35b51-aeb1-4e2f-9be0-11230cb89fbe-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jqgjd\" (UID: \"54e35b51-aeb1-4e2f-9be0-11230cb89fbe\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jqgjd" Oct 06 10:30:33 crc kubenswrapper[4824]: I1006 10:30:33.880139 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/54e35b51-aeb1-4e2f-9be0-11230cb89fbe-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jqgjd\" (UID: \"54e35b51-aeb1-4e2f-9be0-11230cb89fbe\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jqgjd" Oct 06 10:30:33 crc kubenswrapper[4824]: I1006 10:30:33.880159 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54e35b51-aeb1-4e2f-9be0-11230cb89fbe-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jqgjd\" (UID: \"54e35b51-aeb1-4e2f-9be0-11230cb89fbe\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jqgjd" Oct 06 10:30:33 crc kubenswrapper[4824]: I1006 10:30:33.880189 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/54e35b51-aeb1-4e2f-9be0-11230cb89fbe-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jqgjd\" (UID: \"54e35b51-aeb1-4e2f-9be0-11230cb89fbe\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jqgjd" Oct 06 10:30:33 crc kubenswrapper[4824]: I1006 10:30:33.880209 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/54e35b51-aeb1-4e2f-9be0-11230cb89fbe-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jqgjd\" (UID: \"54e35b51-aeb1-4e2f-9be0-11230cb89fbe\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jqgjd" Oct 06 10:30:33 crc kubenswrapper[4824]: I1006 10:30:33.880236 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/54e35b51-aeb1-4e2f-9be0-11230cb89fbe-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jqgjd\" (UID: \"54e35b51-aeb1-4e2f-9be0-11230cb89fbe\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jqgjd" Oct 06 10:30:33 crc kubenswrapper[4824]: I1006 10:30:33.880262 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54e35b51-aeb1-4e2f-9be0-11230cb89fbe-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jqgjd\" (UID: \"54e35b51-aeb1-4e2f-9be0-11230cb89fbe\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jqgjd" Oct 06 10:30:33 crc kubenswrapper[4824]: I1006 10:30:33.880281 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/54e35b51-aeb1-4e2f-9be0-11230cb89fbe-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jqgjd\" (UID: \"54e35b51-aeb1-4e2f-9be0-11230cb89fbe\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jqgjd" Oct 06 10:30:33 crc kubenswrapper[4824]: I1006 10:30:33.982626 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/54e35b51-aeb1-4e2f-9be0-11230cb89fbe-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jqgjd\" (UID: \"54e35b51-aeb1-4e2f-9be0-11230cb89fbe\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jqgjd" Oct 06 10:30:33 crc kubenswrapper[4824]: I1006 10:30:33.982713 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54e35b51-aeb1-4e2f-9be0-11230cb89fbe-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jqgjd\" (UID: \"54e35b51-aeb1-4e2f-9be0-11230cb89fbe\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jqgjd" Oct 06 10:30:33 crc kubenswrapper[4824]: I1006 10:30:33.982747 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/54e35b51-aeb1-4e2f-9be0-11230cb89fbe-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jqgjd\" (UID: \"54e35b51-aeb1-4e2f-9be0-11230cb89fbe\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jqgjd" Oct 06 10:30:33 crc kubenswrapper[4824]: I1006 10:30:33.982783 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54e35b51-aeb1-4e2f-9be0-11230cb89fbe-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jqgjd\" (UID: \"54e35b51-aeb1-4e2f-9be0-11230cb89fbe\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jqgjd" Oct 06 10:30:33 crc kubenswrapper[4824]: I1006 10:30:33.982812 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54e35b51-aeb1-4e2f-9be0-11230cb89fbe-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jqgjd\" (UID: \"54e35b51-aeb1-4e2f-9be0-11230cb89fbe\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jqgjd" Oct 06 10:30:33 crc kubenswrapper[4824]: I1006 10:30:33.982832 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54e35b51-aeb1-4e2f-9be0-11230cb89fbe-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jqgjd\" (UID: \"54e35b51-aeb1-4e2f-9be0-11230cb89fbe\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jqgjd" Oct 06 10:30:33 crc kubenswrapper[4824]: I1006 10:30:33.982861 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54e35b51-aeb1-4e2f-9be0-11230cb89fbe-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jqgjd\" (UID: \"54e35b51-aeb1-4e2f-9be0-11230cb89fbe\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jqgjd" Oct 06 10:30:33 crc kubenswrapper[4824]: I1006 10:30:33.982906 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dwrrn\" (UniqueName: \"kubernetes.io/projected/54e35b51-aeb1-4e2f-9be0-11230cb89fbe-kube-api-access-dwrrn\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jqgjd\" (UID: \"54e35b51-aeb1-4e2f-9be0-11230cb89fbe\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jqgjd" Oct 06 10:30:33 crc kubenswrapper[4824]: I1006 10:30:33.982926 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54e35b51-aeb1-4e2f-9be0-11230cb89fbe-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jqgjd\" (UID: \"54e35b51-aeb1-4e2f-9be0-11230cb89fbe\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jqgjd" Oct 06 10:30:33 crc kubenswrapper[4824]: I1006 10:30:33.982965 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/54e35b51-aeb1-4e2f-9be0-11230cb89fbe-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jqgjd\" (UID: \"54e35b51-aeb1-4e2f-9be0-11230cb89fbe\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jqgjd" Oct 06 10:30:33 crc kubenswrapper[4824]: I1006 10:30:33.983053 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/54e35b51-aeb1-4e2f-9be0-11230cb89fbe-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jqgjd\" (UID: \"54e35b51-aeb1-4e2f-9be0-11230cb89fbe\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jqgjd" Oct 06 10:30:33 crc kubenswrapper[4824]: I1006 10:30:33.983077 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54e35b51-aeb1-4e2f-9be0-11230cb89fbe-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jqgjd\" (UID: \"54e35b51-aeb1-4e2f-9be0-11230cb89fbe\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jqgjd" Oct 06 10:30:33 crc kubenswrapper[4824]: I1006 10:30:33.983109 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/54e35b51-aeb1-4e2f-9be0-11230cb89fbe-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jqgjd\" (UID: \"54e35b51-aeb1-4e2f-9be0-11230cb89fbe\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jqgjd" Oct 06 10:30:33 crc kubenswrapper[4824]: I1006 10:30:33.983133 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/54e35b51-aeb1-4e2f-9be0-11230cb89fbe-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jqgjd\" (UID: \"54e35b51-aeb1-4e2f-9be0-11230cb89fbe\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jqgjd" Oct 06 10:30:33 crc kubenswrapper[4824]: I1006 10:30:33.990102 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54e35b51-aeb1-4e2f-9be0-11230cb89fbe-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jqgjd\" (UID: \"54e35b51-aeb1-4e2f-9be0-11230cb89fbe\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jqgjd" Oct 06 10:30:33 crc kubenswrapper[4824]: I1006 10:30:33.990329 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54e35b51-aeb1-4e2f-9be0-11230cb89fbe-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jqgjd\" (UID: \"54e35b51-aeb1-4e2f-9be0-11230cb89fbe\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jqgjd" Oct 06 10:30:33 crc kubenswrapper[4824]: I1006 10:30:33.990461 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/54e35b51-aeb1-4e2f-9be0-11230cb89fbe-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jqgjd\" (UID: \"54e35b51-aeb1-4e2f-9be0-11230cb89fbe\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jqgjd" Oct 06 10:30:33 crc kubenswrapper[4824]: I1006 10:30:33.991216 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/54e35b51-aeb1-4e2f-9be0-11230cb89fbe-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jqgjd\" (UID: \"54e35b51-aeb1-4e2f-9be0-11230cb89fbe\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jqgjd" Oct 06 10:30:33 crc kubenswrapper[4824]: I1006 10:30:33.991221 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54e35b51-aeb1-4e2f-9be0-11230cb89fbe-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jqgjd\" (UID: \"54e35b51-aeb1-4e2f-9be0-11230cb89fbe\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jqgjd" Oct 06 10:30:33 crc kubenswrapper[4824]: I1006 10:30:33.992156 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54e35b51-aeb1-4e2f-9be0-11230cb89fbe-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jqgjd\" (UID: \"54e35b51-aeb1-4e2f-9be0-11230cb89fbe\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jqgjd" Oct 06 10:30:33 crc kubenswrapper[4824]: I1006 10:30:33.992430 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/54e35b51-aeb1-4e2f-9be0-11230cb89fbe-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jqgjd\" (UID: \"54e35b51-aeb1-4e2f-9be0-11230cb89fbe\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jqgjd" Oct 06 10:30:33 crc kubenswrapper[4824]: I1006 10:30:33.992507 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/54e35b51-aeb1-4e2f-9be0-11230cb89fbe-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jqgjd\" (UID: \"54e35b51-aeb1-4e2f-9be0-11230cb89fbe\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jqgjd" Oct 06 10:30:33 crc kubenswrapper[4824]: I1006 10:30:33.993749 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/54e35b51-aeb1-4e2f-9be0-11230cb89fbe-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jqgjd\" (UID: \"54e35b51-aeb1-4e2f-9be0-11230cb89fbe\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jqgjd" Oct 06 10:30:33 crc kubenswrapper[4824]: I1006 10:30:33.998020 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54e35b51-aeb1-4e2f-9be0-11230cb89fbe-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jqgjd\" (UID: \"54e35b51-aeb1-4e2f-9be0-11230cb89fbe\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jqgjd" Oct 06 10:30:33 crc kubenswrapper[4824]: I1006 10:30:33.998591 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54e35b51-aeb1-4e2f-9be0-11230cb89fbe-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jqgjd\" (UID: \"54e35b51-aeb1-4e2f-9be0-11230cb89fbe\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jqgjd" Oct 06 10:30:33 crc kubenswrapper[4824]: I1006 10:30:33.999046 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/54e35b51-aeb1-4e2f-9be0-11230cb89fbe-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jqgjd\" (UID: \"54e35b51-aeb1-4e2f-9be0-11230cb89fbe\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jqgjd" Oct 06 10:30:34 crc kubenswrapper[4824]: I1006 10:30:34.001010 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dwrrn\" (UniqueName: \"kubernetes.io/projected/54e35b51-aeb1-4e2f-9be0-11230cb89fbe-kube-api-access-dwrrn\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jqgjd\" (UID: \"54e35b51-aeb1-4e2f-9be0-11230cb89fbe\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jqgjd" Oct 06 10:30:34 crc kubenswrapper[4824]: I1006 10:30:34.009245 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54e35b51-aeb1-4e2f-9be0-11230cb89fbe-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jqgjd\" (UID: \"54e35b51-aeb1-4e2f-9be0-11230cb89fbe\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jqgjd" Oct 06 10:30:34 crc kubenswrapper[4824]: I1006 10:30:34.102631 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jqgjd" Oct 06 10:30:34 crc kubenswrapper[4824]: I1006 10:30:34.706224 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jqgjd"] Oct 06 10:30:34 crc kubenswrapper[4824]: W1006 10:30:34.716562 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod54e35b51_aeb1_4e2f_9be0_11230cb89fbe.slice/crio-43ebdee308c2f00bc86fee3040375a6be3d2b23da03e3c6e3d075b75f0db4c21 WatchSource:0}: Error finding container 43ebdee308c2f00bc86fee3040375a6be3d2b23da03e3c6e3d075b75f0db4c21: Status 404 returned error can't find the container with id 43ebdee308c2f00bc86fee3040375a6be3d2b23da03e3c6e3d075b75f0db4c21 Oct 06 10:30:35 crc kubenswrapper[4824]: I1006 10:30:35.613731 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jqgjd" event={"ID":"54e35b51-aeb1-4e2f-9be0-11230cb89fbe","Type":"ContainerStarted","Data":"3944d219cfa36795655c6e78c98093f238207f103dd21c0456b404d89368c56b"} Oct 06 10:30:35 crc kubenswrapper[4824]: I1006 10:30:35.613788 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jqgjd" event={"ID":"54e35b51-aeb1-4e2f-9be0-11230cb89fbe","Type":"ContainerStarted","Data":"43ebdee308c2f00bc86fee3040375a6be3d2b23da03e3c6e3d075b75f0db4c21"} Oct 06 10:30:35 crc kubenswrapper[4824]: I1006 10:30:35.657234 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jqgjd" podStartSLOduration=2.245529637 podStartE2EDuration="2.657212325s" podCreationTimestamp="2025-10-06 10:30:33 +0000 UTC" firstStartedPulling="2025-10-06 10:30:34.7205161 +0000 UTC m=+1944.084939001" lastFinishedPulling="2025-10-06 10:30:35.132198818 +0000 UTC m=+1944.496621689" observedRunningTime="2025-10-06 10:30:35.642596864 +0000 UTC m=+1945.007019725" watchObservedRunningTime="2025-10-06 10:30:35.657212325 +0000 UTC m=+1945.021635196" Oct 06 10:30:41 crc kubenswrapper[4824]: I1006 10:30:41.280485 4824 scope.go:117] "RemoveContainer" containerID="9de43c7f48ed9ad071a24adf937829ab92276bd505bb72d5b89bac4922aac3b0" Oct 06 10:30:41 crc kubenswrapper[4824]: E1006 10:30:41.281439 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-khgzw_openshift-machine-config-operator(e1140c8c-93f1-4fce-8c06-a69561ff8a6d)\"" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" Oct 06 10:30:56 crc kubenswrapper[4824]: I1006 10:30:56.275830 4824 scope.go:117] "RemoveContainer" containerID="9de43c7f48ed9ad071a24adf937829ab92276bd505bb72d5b89bac4922aac3b0" Oct 06 10:30:56 crc kubenswrapper[4824]: E1006 10:30:56.277292 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-khgzw_openshift-machine-config-operator(e1140c8c-93f1-4fce-8c06-a69561ff8a6d)\"" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" Oct 06 10:31:07 crc kubenswrapper[4824]: I1006 10:31:07.274855 4824 scope.go:117] "RemoveContainer" containerID="9de43c7f48ed9ad071a24adf937829ab92276bd505bb72d5b89bac4922aac3b0" Oct 06 10:31:07 crc kubenswrapper[4824]: E1006 10:31:07.275853 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-khgzw_openshift-machine-config-operator(e1140c8c-93f1-4fce-8c06-a69561ff8a6d)\"" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" Oct 06 10:31:17 crc kubenswrapper[4824]: I1006 10:31:17.047492 4824 generic.go:334] "Generic (PLEG): container finished" podID="54e35b51-aeb1-4e2f-9be0-11230cb89fbe" containerID="3944d219cfa36795655c6e78c98093f238207f103dd21c0456b404d89368c56b" exitCode=0 Oct 06 10:31:17 crc kubenswrapper[4824]: I1006 10:31:17.047607 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jqgjd" event={"ID":"54e35b51-aeb1-4e2f-9be0-11230cb89fbe","Type":"ContainerDied","Data":"3944d219cfa36795655c6e78c98093f238207f103dd21c0456b404d89368c56b"} Oct 06 10:31:18 crc kubenswrapper[4824]: I1006 10:31:18.494675 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jqgjd" Oct 06 10:31:18 crc kubenswrapper[4824]: I1006 10:31:18.647845 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54e35b51-aeb1-4e2f-9be0-11230cb89fbe-bootstrap-combined-ca-bundle\") pod \"54e35b51-aeb1-4e2f-9be0-11230cb89fbe\" (UID: \"54e35b51-aeb1-4e2f-9be0-11230cb89fbe\") " Oct 06 10:31:18 crc kubenswrapper[4824]: I1006 10:31:18.647921 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/54e35b51-aeb1-4e2f-9be0-11230cb89fbe-ssh-key\") pod \"54e35b51-aeb1-4e2f-9be0-11230cb89fbe\" (UID: \"54e35b51-aeb1-4e2f-9be0-11230cb89fbe\") " Oct 06 10:31:18 crc kubenswrapper[4824]: I1006 10:31:18.647947 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54e35b51-aeb1-4e2f-9be0-11230cb89fbe-repo-setup-combined-ca-bundle\") pod \"54e35b51-aeb1-4e2f-9be0-11230cb89fbe\" (UID: \"54e35b51-aeb1-4e2f-9be0-11230cb89fbe\") " Oct 06 10:31:18 crc kubenswrapper[4824]: I1006 10:31:18.648071 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/54e35b51-aeb1-4e2f-9be0-11230cb89fbe-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"54e35b51-aeb1-4e2f-9be0-11230cb89fbe\" (UID: \"54e35b51-aeb1-4e2f-9be0-11230cb89fbe\") " Oct 06 10:31:18 crc kubenswrapper[4824]: I1006 10:31:18.648128 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54e35b51-aeb1-4e2f-9be0-11230cb89fbe-telemetry-combined-ca-bundle\") pod \"54e35b51-aeb1-4e2f-9be0-11230cb89fbe\" (UID: \"54e35b51-aeb1-4e2f-9be0-11230cb89fbe\") " Oct 06 10:31:18 crc kubenswrapper[4824]: I1006 10:31:18.648170 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54e35b51-aeb1-4e2f-9be0-11230cb89fbe-libvirt-combined-ca-bundle\") pod \"54e35b51-aeb1-4e2f-9be0-11230cb89fbe\" (UID: \"54e35b51-aeb1-4e2f-9be0-11230cb89fbe\") " Oct 06 10:31:18 crc kubenswrapper[4824]: I1006 10:31:18.648197 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/54e35b51-aeb1-4e2f-9be0-11230cb89fbe-openstack-edpm-ipam-ovn-default-certs-0\") pod \"54e35b51-aeb1-4e2f-9be0-11230cb89fbe\" (UID: \"54e35b51-aeb1-4e2f-9be0-11230cb89fbe\") " Oct 06 10:31:18 crc kubenswrapper[4824]: I1006 10:31:18.648225 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54e35b51-aeb1-4e2f-9be0-11230cb89fbe-neutron-metadata-combined-ca-bundle\") pod \"54e35b51-aeb1-4e2f-9be0-11230cb89fbe\" (UID: \"54e35b51-aeb1-4e2f-9be0-11230cb89fbe\") " Oct 06 10:31:18 crc kubenswrapper[4824]: I1006 10:31:18.648250 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dwrrn\" (UniqueName: \"kubernetes.io/projected/54e35b51-aeb1-4e2f-9be0-11230cb89fbe-kube-api-access-dwrrn\") pod \"54e35b51-aeb1-4e2f-9be0-11230cb89fbe\" (UID: \"54e35b51-aeb1-4e2f-9be0-11230cb89fbe\") " Oct 06 10:31:18 crc kubenswrapper[4824]: I1006 10:31:18.648365 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/54e35b51-aeb1-4e2f-9be0-11230cb89fbe-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"54e35b51-aeb1-4e2f-9be0-11230cb89fbe\" (UID: \"54e35b51-aeb1-4e2f-9be0-11230cb89fbe\") " Oct 06 10:31:18 crc kubenswrapper[4824]: I1006 10:31:18.648389 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/54e35b51-aeb1-4e2f-9be0-11230cb89fbe-inventory\") pod \"54e35b51-aeb1-4e2f-9be0-11230cb89fbe\" (UID: \"54e35b51-aeb1-4e2f-9be0-11230cb89fbe\") " Oct 06 10:31:18 crc kubenswrapper[4824]: I1006 10:31:18.648414 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54e35b51-aeb1-4e2f-9be0-11230cb89fbe-nova-combined-ca-bundle\") pod \"54e35b51-aeb1-4e2f-9be0-11230cb89fbe\" (UID: \"54e35b51-aeb1-4e2f-9be0-11230cb89fbe\") " Oct 06 10:31:18 crc kubenswrapper[4824]: I1006 10:31:18.648440 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/54e35b51-aeb1-4e2f-9be0-11230cb89fbe-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"54e35b51-aeb1-4e2f-9be0-11230cb89fbe\" (UID: \"54e35b51-aeb1-4e2f-9be0-11230cb89fbe\") " Oct 06 10:31:18 crc kubenswrapper[4824]: I1006 10:31:18.648469 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54e35b51-aeb1-4e2f-9be0-11230cb89fbe-ovn-combined-ca-bundle\") pod \"54e35b51-aeb1-4e2f-9be0-11230cb89fbe\" (UID: \"54e35b51-aeb1-4e2f-9be0-11230cb89fbe\") " Oct 06 10:31:18 crc kubenswrapper[4824]: I1006 10:31:18.655224 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/54e35b51-aeb1-4e2f-9be0-11230cb89fbe-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "54e35b51-aeb1-4e2f-9be0-11230cb89fbe" (UID: "54e35b51-aeb1-4e2f-9be0-11230cb89fbe"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:31:18 crc kubenswrapper[4824]: I1006 10:31:18.655320 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/54e35b51-aeb1-4e2f-9be0-11230cb89fbe-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "54e35b51-aeb1-4e2f-9be0-11230cb89fbe" (UID: "54e35b51-aeb1-4e2f-9be0-11230cb89fbe"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:31:18 crc kubenswrapper[4824]: I1006 10:31:18.655356 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/54e35b51-aeb1-4e2f-9be0-11230cb89fbe-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "54e35b51-aeb1-4e2f-9be0-11230cb89fbe" (UID: "54e35b51-aeb1-4e2f-9be0-11230cb89fbe"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:31:18 crc kubenswrapper[4824]: I1006 10:31:18.655905 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/54e35b51-aeb1-4e2f-9be0-11230cb89fbe-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "54e35b51-aeb1-4e2f-9be0-11230cb89fbe" (UID: "54e35b51-aeb1-4e2f-9be0-11230cb89fbe"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:31:18 crc kubenswrapper[4824]: I1006 10:31:18.656003 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/54e35b51-aeb1-4e2f-9be0-11230cb89fbe-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "54e35b51-aeb1-4e2f-9be0-11230cb89fbe" (UID: "54e35b51-aeb1-4e2f-9be0-11230cb89fbe"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:31:18 crc kubenswrapper[4824]: I1006 10:31:18.656405 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/54e35b51-aeb1-4e2f-9be0-11230cb89fbe-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "54e35b51-aeb1-4e2f-9be0-11230cb89fbe" (UID: "54e35b51-aeb1-4e2f-9be0-11230cb89fbe"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:31:18 crc kubenswrapper[4824]: I1006 10:31:18.656484 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/54e35b51-aeb1-4e2f-9be0-11230cb89fbe-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "54e35b51-aeb1-4e2f-9be0-11230cb89fbe" (UID: "54e35b51-aeb1-4e2f-9be0-11230cb89fbe"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:31:18 crc kubenswrapper[4824]: I1006 10:31:18.658330 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/54e35b51-aeb1-4e2f-9be0-11230cb89fbe-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "54e35b51-aeb1-4e2f-9be0-11230cb89fbe" (UID: "54e35b51-aeb1-4e2f-9be0-11230cb89fbe"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:31:18 crc kubenswrapper[4824]: I1006 10:31:18.659255 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/54e35b51-aeb1-4e2f-9be0-11230cb89fbe-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "54e35b51-aeb1-4e2f-9be0-11230cb89fbe" (UID: "54e35b51-aeb1-4e2f-9be0-11230cb89fbe"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:31:18 crc kubenswrapper[4824]: I1006 10:31:18.659281 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/54e35b51-aeb1-4e2f-9be0-11230cb89fbe-openstack-edpm-ipam-telemetry-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-default-certs-0") pod "54e35b51-aeb1-4e2f-9be0-11230cb89fbe" (UID: "54e35b51-aeb1-4e2f-9be0-11230cb89fbe"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:31:18 crc kubenswrapper[4824]: I1006 10:31:18.662181 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/54e35b51-aeb1-4e2f-9be0-11230cb89fbe-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "54e35b51-aeb1-4e2f-9be0-11230cb89fbe" (UID: "54e35b51-aeb1-4e2f-9be0-11230cb89fbe"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:31:18 crc kubenswrapper[4824]: I1006 10:31:18.662282 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/54e35b51-aeb1-4e2f-9be0-11230cb89fbe-kube-api-access-dwrrn" (OuterVolumeSpecName: "kube-api-access-dwrrn") pod "54e35b51-aeb1-4e2f-9be0-11230cb89fbe" (UID: "54e35b51-aeb1-4e2f-9be0-11230cb89fbe"). InnerVolumeSpecName "kube-api-access-dwrrn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:31:18 crc kubenswrapper[4824]: I1006 10:31:18.683303 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/54e35b51-aeb1-4e2f-9be0-11230cb89fbe-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "54e35b51-aeb1-4e2f-9be0-11230cb89fbe" (UID: "54e35b51-aeb1-4e2f-9be0-11230cb89fbe"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:31:18 crc kubenswrapper[4824]: I1006 10:31:18.685761 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/54e35b51-aeb1-4e2f-9be0-11230cb89fbe-inventory" (OuterVolumeSpecName: "inventory") pod "54e35b51-aeb1-4e2f-9be0-11230cb89fbe" (UID: "54e35b51-aeb1-4e2f-9be0-11230cb89fbe"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:31:18 crc kubenswrapper[4824]: I1006 10:31:18.759647 4824 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54e35b51-aeb1-4e2f-9be0-11230cb89fbe-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 10:31:18 crc kubenswrapper[4824]: I1006 10:31:18.759726 4824 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54e35b51-aeb1-4e2f-9be0-11230cb89fbe-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 10:31:18 crc kubenswrapper[4824]: I1006 10:31:18.759745 4824 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/54e35b51-aeb1-4e2f-9be0-11230cb89fbe-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 06 10:31:18 crc kubenswrapper[4824]: I1006 10:31:18.759761 4824 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54e35b51-aeb1-4e2f-9be0-11230cb89fbe-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 10:31:18 crc kubenswrapper[4824]: I1006 10:31:18.759779 4824 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/54e35b51-aeb1-4e2f-9be0-11230cb89fbe-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 06 10:31:18 crc kubenswrapper[4824]: I1006 10:31:18.759797 4824 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54e35b51-aeb1-4e2f-9be0-11230cb89fbe-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 10:31:18 crc kubenswrapper[4824]: I1006 10:31:18.759810 4824 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54e35b51-aeb1-4e2f-9be0-11230cb89fbe-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 10:31:18 crc kubenswrapper[4824]: I1006 10:31:18.759824 4824 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/54e35b51-aeb1-4e2f-9be0-11230cb89fbe-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 06 10:31:18 crc kubenswrapper[4824]: I1006 10:31:18.759842 4824 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54e35b51-aeb1-4e2f-9be0-11230cb89fbe-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 10:31:18 crc kubenswrapper[4824]: I1006 10:31:18.759858 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dwrrn\" (UniqueName: \"kubernetes.io/projected/54e35b51-aeb1-4e2f-9be0-11230cb89fbe-kube-api-access-dwrrn\") on node \"crc\" DevicePath \"\"" Oct 06 10:31:18 crc kubenswrapper[4824]: I1006 10:31:18.759871 4824 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/54e35b51-aeb1-4e2f-9be0-11230cb89fbe-openstack-edpm-ipam-telemetry-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 06 10:31:18 crc kubenswrapper[4824]: I1006 10:31:18.759888 4824 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/54e35b51-aeb1-4e2f-9be0-11230cb89fbe-inventory\") on node \"crc\" DevicePath \"\"" Oct 06 10:31:18 crc kubenswrapper[4824]: I1006 10:31:18.759905 4824 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54e35b51-aeb1-4e2f-9be0-11230cb89fbe-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 10:31:18 crc kubenswrapper[4824]: I1006 10:31:18.759918 4824 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/54e35b51-aeb1-4e2f-9be0-11230cb89fbe-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 06 10:31:19 crc kubenswrapper[4824]: I1006 10:31:19.067877 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jqgjd" event={"ID":"54e35b51-aeb1-4e2f-9be0-11230cb89fbe","Type":"ContainerDied","Data":"43ebdee308c2f00bc86fee3040375a6be3d2b23da03e3c6e3d075b75f0db4c21"} Oct 06 10:31:19 crc kubenswrapper[4824]: I1006 10:31:19.068164 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="43ebdee308c2f00bc86fee3040375a6be3d2b23da03e3c6e3d075b75f0db4c21" Oct 06 10:31:19 crc kubenswrapper[4824]: I1006 10:31:19.067934 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jqgjd" Oct 06 10:31:19 crc kubenswrapper[4824]: I1006 10:31:19.179998 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-fn42c"] Oct 06 10:31:19 crc kubenswrapper[4824]: E1006 10:31:19.180523 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54e35b51-aeb1-4e2f-9be0-11230cb89fbe" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Oct 06 10:31:19 crc kubenswrapper[4824]: I1006 10:31:19.180547 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="54e35b51-aeb1-4e2f-9be0-11230cb89fbe" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Oct 06 10:31:19 crc kubenswrapper[4824]: I1006 10:31:19.180828 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="54e35b51-aeb1-4e2f-9be0-11230cb89fbe" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Oct 06 10:31:19 crc kubenswrapper[4824]: I1006 10:31:19.183178 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-fn42c" Oct 06 10:31:19 crc kubenswrapper[4824]: I1006 10:31:19.194936 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 06 10:31:19 crc kubenswrapper[4824]: I1006 10:31:19.195334 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 06 10:31:19 crc kubenswrapper[4824]: I1006 10:31:19.195458 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 06 10:31:19 crc kubenswrapper[4824]: I1006 10:31:19.195348 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Oct 06 10:31:19 crc kubenswrapper[4824]: I1006 10:31:19.195692 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-f886w" Oct 06 10:31:19 crc kubenswrapper[4824]: I1006 10:31:19.218162 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-fn42c"] Oct 06 10:31:19 crc kubenswrapper[4824]: I1006 10:31:19.377124 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4rcwt\" (UniqueName: \"kubernetes.io/projected/97deb005-9b94-45a5-8d7f-84147ae5807c-kube-api-access-4rcwt\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-fn42c\" (UID: \"97deb005-9b94-45a5-8d7f-84147ae5807c\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-fn42c" Oct 06 10:31:19 crc kubenswrapper[4824]: I1006 10:31:19.377588 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/97deb005-9b94-45a5-8d7f-84147ae5807c-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-fn42c\" (UID: \"97deb005-9b94-45a5-8d7f-84147ae5807c\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-fn42c" Oct 06 10:31:19 crc kubenswrapper[4824]: I1006 10:31:19.377686 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/97deb005-9b94-45a5-8d7f-84147ae5807c-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-fn42c\" (UID: \"97deb005-9b94-45a5-8d7f-84147ae5807c\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-fn42c" Oct 06 10:31:19 crc kubenswrapper[4824]: I1006 10:31:19.377859 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97deb005-9b94-45a5-8d7f-84147ae5807c-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-fn42c\" (UID: \"97deb005-9b94-45a5-8d7f-84147ae5807c\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-fn42c" Oct 06 10:31:19 crc kubenswrapper[4824]: I1006 10:31:19.377934 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/97deb005-9b94-45a5-8d7f-84147ae5807c-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-fn42c\" (UID: \"97deb005-9b94-45a5-8d7f-84147ae5807c\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-fn42c" Oct 06 10:31:19 crc kubenswrapper[4824]: I1006 10:31:19.499054 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/97deb005-9b94-45a5-8d7f-84147ae5807c-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-fn42c\" (UID: \"97deb005-9b94-45a5-8d7f-84147ae5807c\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-fn42c" Oct 06 10:31:19 crc kubenswrapper[4824]: I1006 10:31:19.499251 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/97deb005-9b94-45a5-8d7f-84147ae5807c-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-fn42c\" (UID: \"97deb005-9b94-45a5-8d7f-84147ae5807c\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-fn42c" Oct 06 10:31:19 crc kubenswrapper[4824]: I1006 10:31:19.499687 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97deb005-9b94-45a5-8d7f-84147ae5807c-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-fn42c\" (UID: \"97deb005-9b94-45a5-8d7f-84147ae5807c\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-fn42c" Oct 06 10:31:19 crc kubenswrapper[4824]: I1006 10:31:19.499773 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/97deb005-9b94-45a5-8d7f-84147ae5807c-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-fn42c\" (UID: \"97deb005-9b94-45a5-8d7f-84147ae5807c\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-fn42c" Oct 06 10:31:19 crc kubenswrapper[4824]: I1006 10:31:19.499936 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4rcwt\" (UniqueName: \"kubernetes.io/projected/97deb005-9b94-45a5-8d7f-84147ae5807c-kube-api-access-4rcwt\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-fn42c\" (UID: \"97deb005-9b94-45a5-8d7f-84147ae5807c\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-fn42c" Oct 06 10:31:19 crc kubenswrapper[4824]: I1006 10:31:19.503739 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/97deb005-9b94-45a5-8d7f-84147ae5807c-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-fn42c\" (UID: \"97deb005-9b94-45a5-8d7f-84147ae5807c\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-fn42c" Oct 06 10:31:19 crc kubenswrapper[4824]: I1006 10:31:19.510639 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/97deb005-9b94-45a5-8d7f-84147ae5807c-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-fn42c\" (UID: \"97deb005-9b94-45a5-8d7f-84147ae5807c\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-fn42c" Oct 06 10:31:19 crc kubenswrapper[4824]: I1006 10:31:19.520473 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97deb005-9b94-45a5-8d7f-84147ae5807c-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-fn42c\" (UID: \"97deb005-9b94-45a5-8d7f-84147ae5807c\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-fn42c" Oct 06 10:31:19 crc kubenswrapper[4824]: I1006 10:31:19.524997 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4rcwt\" (UniqueName: \"kubernetes.io/projected/97deb005-9b94-45a5-8d7f-84147ae5807c-kube-api-access-4rcwt\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-fn42c\" (UID: \"97deb005-9b94-45a5-8d7f-84147ae5807c\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-fn42c" Oct 06 10:31:19 crc kubenswrapper[4824]: I1006 10:31:19.527142 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/97deb005-9b94-45a5-8d7f-84147ae5807c-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-fn42c\" (UID: \"97deb005-9b94-45a5-8d7f-84147ae5807c\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-fn42c" Oct 06 10:31:19 crc kubenswrapper[4824]: I1006 10:31:19.821905 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-fn42c" Oct 06 10:31:20 crc kubenswrapper[4824]: I1006 10:31:20.275240 4824 scope.go:117] "RemoveContainer" containerID="9de43c7f48ed9ad071a24adf937829ab92276bd505bb72d5b89bac4922aac3b0" Oct 06 10:31:20 crc kubenswrapper[4824]: I1006 10:31:20.401666 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-fn42c"] Oct 06 10:31:20 crc kubenswrapper[4824]: W1006 10:31:20.418240 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod97deb005_9b94_45a5_8d7f_84147ae5807c.slice/crio-44fdf78201a1b3f5fd961a30659b65407e3cbdd4e4fc0ebed8634e2f4d74380d WatchSource:0}: Error finding container 44fdf78201a1b3f5fd961a30659b65407e3cbdd4e4fc0ebed8634e2f4d74380d: Status 404 returned error can't find the container with id 44fdf78201a1b3f5fd961a30659b65407e3cbdd4e4fc0ebed8634e2f4d74380d Oct 06 10:31:21 crc kubenswrapper[4824]: I1006 10:31:21.088413 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-fn42c" event={"ID":"97deb005-9b94-45a5-8d7f-84147ae5807c","Type":"ContainerStarted","Data":"44fdf78201a1b3f5fd961a30659b65407e3cbdd4e4fc0ebed8634e2f4d74380d"} Oct 06 10:31:21 crc kubenswrapper[4824]: I1006 10:31:21.093207 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" event={"ID":"e1140c8c-93f1-4fce-8c06-a69561ff8a6d","Type":"ContainerStarted","Data":"2e9a092782b8371eeb30f1c342e8665ca42002786b61ad0319bff781e0406a48"} Oct 06 10:31:22 crc kubenswrapper[4824]: I1006 10:31:22.104228 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-fn42c" event={"ID":"97deb005-9b94-45a5-8d7f-84147ae5807c","Type":"ContainerStarted","Data":"f65097274bd3872a6704d968a2bc05ff5315c2d701eeac6fc0d27eee50d68702"} Oct 06 10:31:22 crc kubenswrapper[4824]: I1006 10:31:22.144184 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-fn42c" podStartSLOduration=2.711869964 podStartE2EDuration="3.144160932s" podCreationTimestamp="2025-10-06 10:31:19 +0000 UTC" firstStartedPulling="2025-10-06 10:31:20.423326964 +0000 UTC m=+1989.787749825" lastFinishedPulling="2025-10-06 10:31:20.855617902 +0000 UTC m=+1990.220040793" observedRunningTime="2025-10-06 10:31:22.143605629 +0000 UTC m=+1991.508028490" watchObservedRunningTime="2025-10-06 10:31:22.144160932 +0000 UTC m=+1991.508583823" Oct 06 10:32:04 crc kubenswrapper[4824]: I1006 10:32:04.081738 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-jj7z9"] Oct 06 10:32:04 crc kubenswrapper[4824]: I1006 10:32:04.086685 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jj7z9" Oct 06 10:32:04 crc kubenswrapper[4824]: I1006 10:32:04.098532 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jj7z9"] Oct 06 10:32:04 crc kubenswrapper[4824]: I1006 10:32:04.230298 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8xg7k\" (UniqueName: \"kubernetes.io/projected/16176c02-b5b6-4c01-926a-bf54896090c1-kube-api-access-8xg7k\") pod \"community-operators-jj7z9\" (UID: \"16176c02-b5b6-4c01-926a-bf54896090c1\") " pod="openshift-marketplace/community-operators-jj7z9" Oct 06 10:32:04 crc kubenswrapper[4824]: I1006 10:32:04.230429 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/16176c02-b5b6-4c01-926a-bf54896090c1-catalog-content\") pod \"community-operators-jj7z9\" (UID: \"16176c02-b5b6-4c01-926a-bf54896090c1\") " pod="openshift-marketplace/community-operators-jj7z9" Oct 06 10:32:04 crc kubenswrapper[4824]: I1006 10:32:04.230470 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/16176c02-b5b6-4c01-926a-bf54896090c1-utilities\") pod \"community-operators-jj7z9\" (UID: \"16176c02-b5b6-4c01-926a-bf54896090c1\") " pod="openshift-marketplace/community-operators-jj7z9" Oct 06 10:32:04 crc kubenswrapper[4824]: I1006 10:32:04.332363 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8xg7k\" (UniqueName: \"kubernetes.io/projected/16176c02-b5b6-4c01-926a-bf54896090c1-kube-api-access-8xg7k\") pod \"community-operators-jj7z9\" (UID: \"16176c02-b5b6-4c01-926a-bf54896090c1\") " pod="openshift-marketplace/community-operators-jj7z9" Oct 06 10:32:04 crc kubenswrapper[4824]: I1006 10:32:04.332495 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/16176c02-b5b6-4c01-926a-bf54896090c1-catalog-content\") pod \"community-operators-jj7z9\" (UID: \"16176c02-b5b6-4c01-926a-bf54896090c1\") " pod="openshift-marketplace/community-operators-jj7z9" Oct 06 10:32:04 crc kubenswrapper[4824]: I1006 10:32:04.332538 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/16176c02-b5b6-4c01-926a-bf54896090c1-utilities\") pod \"community-operators-jj7z9\" (UID: \"16176c02-b5b6-4c01-926a-bf54896090c1\") " pod="openshift-marketplace/community-operators-jj7z9" Oct 06 10:32:04 crc kubenswrapper[4824]: I1006 10:32:04.333136 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/16176c02-b5b6-4c01-926a-bf54896090c1-utilities\") pod \"community-operators-jj7z9\" (UID: \"16176c02-b5b6-4c01-926a-bf54896090c1\") " pod="openshift-marketplace/community-operators-jj7z9" Oct 06 10:32:04 crc kubenswrapper[4824]: I1006 10:32:04.333136 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/16176c02-b5b6-4c01-926a-bf54896090c1-catalog-content\") pod \"community-operators-jj7z9\" (UID: \"16176c02-b5b6-4c01-926a-bf54896090c1\") " pod="openshift-marketplace/community-operators-jj7z9" Oct 06 10:32:04 crc kubenswrapper[4824]: I1006 10:32:04.352935 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8xg7k\" (UniqueName: \"kubernetes.io/projected/16176c02-b5b6-4c01-926a-bf54896090c1-kube-api-access-8xg7k\") pod \"community-operators-jj7z9\" (UID: \"16176c02-b5b6-4c01-926a-bf54896090c1\") " pod="openshift-marketplace/community-operators-jj7z9" Oct 06 10:32:04 crc kubenswrapper[4824]: I1006 10:32:04.424484 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jj7z9" Oct 06 10:32:05 crc kubenswrapper[4824]: I1006 10:32:05.004678 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jj7z9"] Oct 06 10:32:05 crc kubenswrapper[4824]: W1006 10:32:05.013700 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod16176c02_b5b6_4c01_926a_bf54896090c1.slice/crio-2be94c12ad146870f85bf0d324db4148de6174330fbbce8d99d7f3f92e07c92a WatchSource:0}: Error finding container 2be94c12ad146870f85bf0d324db4148de6174330fbbce8d99d7f3f92e07c92a: Status 404 returned error can't find the container with id 2be94c12ad146870f85bf0d324db4148de6174330fbbce8d99d7f3f92e07c92a Oct 06 10:32:05 crc kubenswrapper[4824]: I1006 10:32:05.558375 4824 generic.go:334] "Generic (PLEG): container finished" podID="16176c02-b5b6-4c01-926a-bf54896090c1" containerID="1a075a645e57650f2e96218fda40a74390cf304ddaa0a4c643b45b738a0d5db5" exitCode=0 Oct 06 10:32:05 crc kubenswrapper[4824]: I1006 10:32:05.558626 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jj7z9" event={"ID":"16176c02-b5b6-4c01-926a-bf54896090c1","Type":"ContainerDied","Data":"1a075a645e57650f2e96218fda40a74390cf304ddaa0a4c643b45b738a0d5db5"} Oct 06 10:32:05 crc kubenswrapper[4824]: I1006 10:32:05.558654 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jj7z9" event={"ID":"16176c02-b5b6-4c01-926a-bf54896090c1","Type":"ContainerStarted","Data":"2be94c12ad146870f85bf0d324db4148de6174330fbbce8d99d7f3f92e07c92a"} Oct 06 10:32:07 crc kubenswrapper[4824]: I1006 10:32:07.578367 4824 generic.go:334] "Generic (PLEG): container finished" podID="16176c02-b5b6-4c01-926a-bf54896090c1" containerID="84e36255d17f1d661e72d78dad8d8d7d580f6df006dde59a3a1c082c6ddf8838" exitCode=0 Oct 06 10:32:07 crc kubenswrapper[4824]: I1006 10:32:07.578437 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jj7z9" event={"ID":"16176c02-b5b6-4c01-926a-bf54896090c1","Type":"ContainerDied","Data":"84e36255d17f1d661e72d78dad8d8d7d580f6df006dde59a3a1c082c6ddf8838"} Oct 06 10:32:08 crc kubenswrapper[4824]: I1006 10:32:08.591289 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jj7z9" event={"ID":"16176c02-b5b6-4c01-926a-bf54896090c1","Type":"ContainerStarted","Data":"1b64f620afa612b719a1748b5967547d7c4c638652c9844a66d00f3ad19c7fc0"} Oct 06 10:32:08 crc kubenswrapper[4824]: I1006 10:32:08.619424 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-jj7z9" podStartSLOduration=2.086030245 podStartE2EDuration="4.619402916s" podCreationTimestamp="2025-10-06 10:32:04 +0000 UTC" firstStartedPulling="2025-10-06 10:32:05.56066535 +0000 UTC m=+2034.925088211" lastFinishedPulling="2025-10-06 10:32:08.094038021 +0000 UTC m=+2037.458460882" observedRunningTime="2025-10-06 10:32:08.614094162 +0000 UTC m=+2037.978517043" watchObservedRunningTime="2025-10-06 10:32:08.619402916 +0000 UTC m=+2037.983825777" Oct 06 10:32:14 crc kubenswrapper[4824]: I1006 10:32:14.424820 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-jj7z9" Oct 06 10:32:14 crc kubenswrapper[4824]: I1006 10:32:14.426121 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-jj7z9" Oct 06 10:32:14 crc kubenswrapper[4824]: I1006 10:32:14.498603 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-jj7z9" Oct 06 10:32:14 crc kubenswrapper[4824]: I1006 10:32:14.710314 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-jj7z9" Oct 06 10:32:14 crc kubenswrapper[4824]: I1006 10:32:14.759759 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-jj7z9"] Oct 06 10:32:16 crc kubenswrapper[4824]: I1006 10:32:16.686708 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-jj7z9" podUID="16176c02-b5b6-4c01-926a-bf54896090c1" containerName="registry-server" containerID="cri-o://1b64f620afa612b719a1748b5967547d7c4c638652c9844a66d00f3ad19c7fc0" gracePeriod=2 Oct 06 10:32:16 crc kubenswrapper[4824]: E1006 10:32:16.905138 4824 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod16176c02_b5b6_4c01_926a_bf54896090c1.slice/crio-conmon-1b64f620afa612b719a1748b5967547d7c4c638652c9844a66d00f3ad19c7fc0.scope\": RecentStats: unable to find data in memory cache]" Oct 06 10:32:17 crc kubenswrapper[4824]: I1006 10:32:17.196470 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jj7z9" Oct 06 10:32:17 crc kubenswrapper[4824]: I1006 10:32:17.246087 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/16176c02-b5b6-4c01-926a-bf54896090c1-utilities\") pod \"16176c02-b5b6-4c01-926a-bf54896090c1\" (UID: \"16176c02-b5b6-4c01-926a-bf54896090c1\") " Oct 06 10:32:17 crc kubenswrapper[4824]: I1006 10:32:17.246164 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8xg7k\" (UniqueName: \"kubernetes.io/projected/16176c02-b5b6-4c01-926a-bf54896090c1-kube-api-access-8xg7k\") pod \"16176c02-b5b6-4c01-926a-bf54896090c1\" (UID: \"16176c02-b5b6-4c01-926a-bf54896090c1\") " Oct 06 10:32:17 crc kubenswrapper[4824]: I1006 10:32:17.246963 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/16176c02-b5b6-4c01-926a-bf54896090c1-utilities" (OuterVolumeSpecName: "utilities") pod "16176c02-b5b6-4c01-926a-bf54896090c1" (UID: "16176c02-b5b6-4c01-926a-bf54896090c1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:32:17 crc kubenswrapper[4824]: I1006 10:32:17.254038 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/16176c02-b5b6-4c01-926a-bf54896090c1-kube-api-access-8xg7k" (OuterVolumeSpecName: "kube-api-access-8xg7k") pod "16176c02-b5b6-4c01-926a-bf54896090c1" (UID: "16176c02-b5b6-4c01-926a-bf54896090c1"). InnerVolumeSpecName "kube-api-access-8xg7k". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:32:17 crc kubenswrapper[4824]: I1006 10:32:17.352036 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/16176c02-b5b6-4c01-926a-bf54896090c1-catalog-content\") pod \"16176c02-b5b6-4c01-926a-bf54896090c1\" (UID: \"16176c02-b5b6-4c01-926a-bf54896090c1\") " Oct 06 10:32:17 crc kubenswrapper[4824]: I1006 10:32:17.352864 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8xg7k\" (UniqueName: \"kubernetes.io/projected/16176c02-b5b6-4c01-926a-bf54896090c1-kube-api-access-8xg7k\") on node \"crc\" DevicePath \"\"" Oct 06 10:32:17 crc kubenswrapper[4824]: I1006 10:32:17.352884 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/16176c02-b5b6-4c01-926a-bf54896090c1-utilities\") on node \"crc\" DevicePath \"\"" Oct 06 10:32:17 crc kubenswrapper[4824]: I1006 10:32:17.400103 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/16176c02-b5b6-4c01-926a-bf54896090c1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "16176c02-b5b6-4c01-926a-bf54896090c1" (UID: "16176c02-b5b6-4c01-926a-bf54896090c1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:32:17 crc kubenswrapper[4824]: I1006 10:32:17.453887 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/16176c02-b5b6-4c01-926a-bf54896090c1-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 06 10:32:17 crc kubenswrapper[4824]: I1006 10:32:17.711538 4824 generic.go:334] "Generic (PLEG): container finished" podID="16176c02-b5b6-4c01-926a-bf54896090c1" containerID="1b64f620afa612b719a1748b5967547d7c4c638652c9844a66d00f3ad19c7fc0" exitCode=0 Oct 06 10:32:17 crc kubenswrapper[4824]: I1006 10:32:17.711600 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jj7z9" event={"ID":"16176c02-b5b6-4c01-926a-bf54896090c1","Type":"ContainerDied","Data":"1b64f620afa612b719a1748b5967547d7c4c638652c9844a66d00f3ad19c7fc0"} Oct 06 10:32:17 crc kubenswrapper[4824]: I1006 10:32:17.711653 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jj7z9" event={"ID":"16176c02-b5b6-4c01-926a-bf54896090c1","Type":"ContainerDied","Data":"2be94c12ad146870f85bf0d324db4148de6174330fbbce8d99d7f3f92e07c92a"} Oct 06 10:32:17 crc kubenswrapper[4824]: I1006 10:32:17.711678 4824 scope.go:117] "RemoveContainer" containerID="1b64f620afa612b719a1748b5967547d7c4c638652c9844a66d00f3ad19c7fc0" Oct 06 10:32:17 crc kubenswrapper[4824]: I1006 10:32:17.711926 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jj7z9" Oct 06 10:32:17 crc kubenswrapper[4824]: I1006 10:32:17.743019 4824 scope.go:117] "RemoveContainer" containerID="84e36255d17f1d661e72d78dad8d8d7d580f6df006dde59a3a1c082c6ddf8838" Oct 06 10:32:17 crc kubenswrapper[4824]: I1006 10:32:17.756157 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-jj7z9"] Oct 06 10:32:17 crc kubenswrapper[4824]: I1006 10:32:17.764262 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-jj7z9"] Oct 06 10:32:17 crc kubenswrapper[4824]: I1006 10:32:17.775408 4824 scope.go:117] "RemoveContainer" containerID="1a075a645e57650f2e96218fda40a74390cf304ddaa0a4c643b45b738a0d5db5" Oct 06 10:32:17 crc kubenswrapper[4824]: I1006 10:32:17.808159 4824 scope.go:117] "RemoveContainer" containerID="1b64f620afa612b719a1748b5967547d7c4c638652c9844a66d00f3ad19c7fc0" Oct 06 10:32:17 crc kubenswrapper[4824]: E1006 10:32:17.808722 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1b64f620afa612b719a1748b5967547d7c4c638652c9844a66d00f3ad19c7fc0\": container with ID starting with 1b64f620afa612b719a1748b5967547d7c4c638652c9844a66d00f3ad19c7fc0 not found: ID does not exist" containerID="1b64f620afa612b719a1748b5967547d7c4c638652c9844a66d00f3ad19c7fc0" Oct 06 10:32:17 crc kubenswrapper[4824]: I1006 10:32:17.808754 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1b64f620afa612b719a1748b5967547d7c4c638652c9844a66d00f3ad19c7fc0"} err="failed to get container status \"1b64f620afa612b719a1748b5967547d7c4c638652c9844a66d00f3ad19c7fc0\": rpc error: code = NotFound desc = could not find container \"1b64f620afa612b719a1748b5967547d7c4c638652c9844a66d00f3ad19c7fc0\": container with ID starting with 1b64f620afa612b719a1748b5967547d7c4c638652c9844a66d00f3ad19c7fc0 not found: ID does not exist" Oct 06 10:32:17 crc kubenswrapper[4824]: I1006 10:32:17.808774 4824 scope.go:117] "RemoveContainer" containerID="84e36255d17f1d661e72d78dad8d8d7d580f6df006dde59a3a1c082c6ddf8838" Oct 06 10:32:17 crc kubenswrapper[4824]: E1006 10:32:17.809294 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"84e36255d17f1d661e72d78dad8d8d7d580f6df006dde59a3a1c082c6ddf8838\": container with ID starting with 84e36255d17f1d661e72d78dad8d8d7d580f6df006dde59a3a1c082c6ddf8838 not found: ID does not exist" containerID="84e36255d17f1d661e72d78dad8d8d7d580f6df006dde59a3a1c082c6ddf8838" Oct 06 10:32:17 crc kubenswrapper[4824]: I1006 10:32:17.809346 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"84e36255d17f1d661e72d78dad8d8d7d580f6df006dde59a3a1c082c6ddf8838"} err="failed to get container status \"84e36255d17f1d661e72d78dad8d8d7d580f6df006dde59a3a1c082c6ddf8838\": rpc error: code = NotFound desc = could not find container \"84e36255d17f1d661e72d78dad8d8d7d580f6df006dde59a3a1c082c6ddf8838\": container with ID starting with 84e36255d17f1d661e72d78dad8d8d7d580f6df006dde59a3a1c082c6ddf8838 not found: ID does not exist" Oct 06 10:32:17 crc kubenswrapper[4824]: I1006 10:32:17.809377 4824 scope.go:117] "RemoveContainer" containerID="1a075a645e57650f2e96218fda40a74390cf304ddaa0a4c643b45b738a0d5db5" Oct 06 10:32:17 crc kubenswrapper[4824]: E1006 10:32:17.809849 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1a075a645e57650f2e96218fda40a74390cf304ddaa0a4c643b45b738a0d5db5\": container with ID starting with 1a075a645e57650f2e96218fda40a74390cf304ddaa0a4c643b45b738a0d5db5 not found: ID does not exist" containerID="1a075a645e57650f2e96218fda40a74390cf304ddaa0a4c643b45b738a0d5db5" Oct 06 10:32:17 crc kubenswrapper[4824]: I1006 10:32:17.809928 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1a075a645e57650f2e96218fda40a74390cf304ddaa0a4c643b45b738a0d5db5"} err="failed to get container status \"1a075a645e57650f2e96218fda40a74390cf304ddaa0a4c643b45b738a0d5db5\": rpc error: code = NotFound desc = could not find container \"1a075a645e57650f2e96218fda40a74390cf304ddaa0a4c643b45b738a0d5db5\": container with ID starting with 1a075a645e57650f2e96218fda40a74390cf304ddaa0a4c643b45b738a0d5db5 not found: ID does not exist" Oct 06 10:32:19 crc kubenswrapper[4824]: I1006 10:32:19.289609 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="16176c02-b5b6-4c01-926a-bf54896090c1" path="/var/lib/kubelet/pods/16176c02-b5b6-4c01-926a-bf54896090c1/volumes" Oct 06 10:32:28 crc kubenswrapper[4824]: I1006 10:32:28.831157 4824 generic.go:334] "Generic (PLEG): container finished" podID="97deb005-9b94-45a5-8d7f-84147ae5807c" containerID="f65097274bd3872a6704d968a2bc05ff5315c2d701eeac6fc0d27eee50d68702" exitCode=0 Oct 06 10:32:28 crc kubenswrapper[4824]: I1006 10:32:28.831220 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-fn42c" event={"ID":"97deb005-9b94-45a5-8d7f-84147ae5807c","Type":"ContainerDied","Data":"f65097274bd3872a6704d968a2bc05ff5315c2d701eeac6fc0d27eee50d68702"} Oct 06 10:32:30 crc kubenswrapper[4824]: I1006 10:32:30.260254 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-fn42c" Oct 06 10:32:30 crc kubenswrapper[4824]: I1006 10:32:30.415698 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-87zqs"] Oct 06 10:32:30 crc kubenswrapper[4824]: E1006 10:32:30.416332 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97deb005-9b94-45a5-8d7f-84147ae5807c" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Oct 06 10:32:30 crc kubenswrapper[4824]: I1006 10:32:30.416366 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="97deb005-9b94-45a5-8d7f-84147ae5807c" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Oct 06 10:32:30 crc kubenswrapper[4824]: E1006 10:32:30.416383 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16176c02-b5b6-4c01-926a-bf54896090c1" containerName="extract-utilities" Oct 06 10:32:30 crc kubenswrapper[4824]: I1006 10:32:30.416397 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="16176c02-b5b6-4c01-926a-bf54896090c1" containerName="extract-utilities" Oct 06 10:32:30 crc kubenswrapper[4824]: E1006 10:32:30.416425 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16176c02-b5b6-4c01-926a-bf54896090c1" containerName="extract-content" Oct 06 10:32:30 crc kubenswrapper[4824]: I1006 10:32:30.416436 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="16176c02-b5b6-4c01-926a-bf54896090c1" containerName="extract-content" Oct 06 10:32:30 crc kubenswrapper[4824]: E1006 10:32:30.416459 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16176c02-b5b6-4c01-926a-bf54896090c1" containerName="registry-server" Oct 06 10:32:30 crc kubenswrapper[4824]: I1006 10:32:30.416470 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="16176c02-b5b6-4c01-926a-bf54896090c1" containerName="registry-server" Oct 06 10:32:30 crc kubenswrapper[4824]: I1006 10:32:30.416773 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="16176c02-b5b6-4c01-926a-bf54896090c1" containerName="registry-server" Oct 06 10:32:30 crc kubenswrapper[4824]: I1006 10:32:30.416819 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="97deb005-9b94-45a5-8d7f-84147ae5807c" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Oct 06 10:32:30 crc kubenswrapper[4824]: I1006 10:32:30.418880 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-87zqs" Oct 06 10:32:30 crc kubenswrapper[4824]: I1006 10:32:30.427424 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-87zqs"] Oct 06 10:32:30 crc kubenswrapper[4824]: I1006 10:32:30.434013 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/97deb005-9b94-45a5-8d7f-84147ae5807c-ovncontroller-config-0\") pod \"97deb005-9b94-45a5-8d7f-84147ae5807c\" (UID: \"97deb005-9b94-45a5-8d7f-84147ae5807c\") " Oct 06 10:32:30 crc kubenswrapper[4824]: I1006 10:32:30.434107 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/97deb005-9b94-45a5-8d7f-84147ae5807c-ssh-key\") pod \"97deb005-9b94-45a5-8d7f-84147ae5807c\" (UID: \"97deb005-9b94-45a5-8d7f-84147ae5807c\") " Oct 06 10:32:30 crc kubenswrapper[4824]: I1006 10:32:30.434147 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4rcwt\" (UniqueName: \"kubernetes.io/projected/97deb005-9b94-45a5-8d7f-84147ae5807c-kube-api-access-4rcwt\") pod \"97deb005-9b94-45a5-8d7f-84147ae5807c\" (UID: \"97deb005-9b94-45a5-8d7f-84147ae5807c\") " Oct 06 10:32:30 crc kubenswrapper[4824]: I1006 10:32:30.434289 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/97deb005-9b94-45a5-8d7f-84147ae5807c-inventory\") pod \"97deb005-9b94-45a5-8d7f-84147ae5807c\" (UID: \"97deb005-9b94-45a5-8d7f-84147ae5807c\") " Oct 06 10:32:30 crc kubenswrapper[4824]: I1006 10:32:30.434325 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97deb005-9b94-45a5-8d7f-84147ae5807c-ovn-combined-ca-bundle\") pod \"97deb005-9b94-45a5-8d7f-84147ae5807c\" (UID: \"97deb005-9b94-45a5-8d7f-84147ae5807c\") " Oct 06 10:32:30 crc kubenswrapper[4824]: I1006 10:32:30.449340 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/97deb005-9b94-45a5-8d7f-84147ae5807c-kube-api-access-4rcwt" (OuterVolumeSpecName: "kube-api-access-4rcwt") pod "97deb005-9b94-45a5-8d7f-84147ae5807c" (UID: "97deb005-9b94-45a5-8d7f-84147ae5807c"). InnerVolumeSpecName "kube-api-access-4rcwt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:32:30 crc kubenswrapper[4824]: I1006 10:32:30.450612 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/97deb005-9b94-45a5-8d7f-84147ae5807c-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "97deb005-9b94-45a5-8d7f-84147ae5807c" (UID: "97deb005-9b94-45a5-8d7f-84147ae5807c"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:32:30 crc kubenswrapper[4824]: I1006 10:32:30.476129 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/97deb005-9b94-45a5-8d7f-84147ae5807c-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "97deb005-9b94-45a5-8d7f-84147ae5807c" (UID: "97deb005-9b94-45a5-8d7f-84147ae5807c"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:32:30 crc kubenswrapper[4824]: I1006 10:32:30.478381 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/97deb005-9b94-45a5-8d7f-84147ae5807c-inventory" (OuterVolumeSpecName: "inventory") pod "97deb005-9b94-45a5-8d7f-84147ae5807c" (UID: "97deb005-9b94-45a5-8d7f-84147ae5807c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:32:30 crc kubenswrapper[4824]: I1006 10:32:30.492153 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/97deb005-9b94-45a5-8d7f-84147ae5807c-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "97deb005-9b94-45a5-8d7f-84147ae5807c" (UID: "97deb005-9b94-45a5-8d7f-84147ae5807c"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:32:30 crc kubenswrapper[4824]: I1006 10:32:30.541624 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7ea07f3a-f269-4fab-ac00-a9c78fd19865-catalog-content\") pod \"redhat-marketplace-87zqs\" (UID: \"7ea07f3a-f269-4fab-ac00-a9c78fd19865\") " pod="openshift-marketplace/redhat-marketplace-87zqs" Oct 06 10:32:30 crc kubenswrapper[4824]: I1006 10:32:30.541755 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ldsmh\" (UniqueName: \"kubernetes.io/projected/7ea07f3a-f269-4fab-ac00-a9c78fd19865-kube-api-access-ldsmh\") pod \"redhat-marketplace-87zqs\" (UID: \"7ea07f3a-f269-4fab-ac00-a9c78fd19865\") " pod="openshift-marketplace/redhat-marketplace-87zqs" Oct 06 10:32:30 crc kubenswrapper[4824]: I1006 10:32:30.541859 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7ea07f3a-f269-4fab-ac00-a9c78fd19865-utilities\") pod \"redhat-marketplace-87zqs\" (UID: \"7ea07f3a-f269-4fab-ac00-a9c78fd19865\") " pod="openshift-marketplace/redhat-marketplace-87zqs" Oct 06 10:32:30 crc kubenswrapper[4824]: I1006 10:32:30.542008 4824 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/97deb005-9b94-45a5-8d7f-84147ae5807c-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 06 10:32:30 crc kubenswrapper[4824]: I1006 10:32:30.542066 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4rcwt\" (UniqueName: \"kubernetes.io/projected/97deb005-9b94-45a5-8d7f-84147ae5807c-kube-api-access-4rcwt\") on node \"crc\" DevicePath \"\"" Oct 06 10:32:30 crc kubenswrapper[4824]: I1006 10:32:30.542710 4824 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97deb005-9b94-45a5-8d7f-84147ae5807c-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 10:32:30 crc kubenswrapper[4824]: I1006 10:32:30.542755 4824 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/97deb005-9b94-45a5-8d7f-84147ae5807c-inventory\") on node \"crc\" DevicePath \"\"" Oct 06 10:32:30 crc kubenswrapper[4824]: I1006 10:32:30.542766 4824 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/97deb005-9b94-45a5-8d7f-84147ae5807c-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Oct 06 10:32:30 crc kubenswrapper[4824]: I1006 10:32:30.644660 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7ea07f3a-f269-4fab-ac00-a9c78fd19865-utilities\") pod \"redhat-marketplace-87zqs\" (UID: \"7ea07f3a-f269-4fab-ac00-a9c78fd19865\") " pod="openshift-marketplace/redhat-marketplace-87zqs" Oct 06 10:32:30 crc kubenswrapper[4824]: I1006 10:32:30.644823 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7ea07f3a-f269-4fab-ac00-a9c78fd19865-catalog-content\") pod \"redhat-marketplace-87zqs\" (UID: \"7ea07f3a-f269-4fab-ac00-a9c78fd19865\") " pod="openshift-marketplace/redhat-marketplace-87zqs" Oct 06 10:32:30 crc kubenswrapper[4824]: I1006 10:32:30.644877 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ldsmh\" (UniqueName: \"kubernetes.io/projected/7ea07f3a-f269-4fab-ac00-a9c78fd19865-kube-api-access-ldsmh\") pod \"redhat-marketplace-87zqs\" (UID: \"7ea07f3a-f269-4fab-ac00-a9c78fd19865\") " pod="openshift-marketplace/redhat-marketplace-87zqs" Oct 06 10:32:30 crc kubenswrapper[4824]: I1006 10:32:30.645509 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7ea07f3a-f269-4fab-ac00-a9c78fd19865-utilities\") pod \"redhat-marketplace-87zqs\" (UID: \"7ea07f3a-f269-4fab-ac00-a9c78fd19865\") " pod="openshift-marketplace/redhat-marketplace-87zqs" Oct 06 10:32:30 crc kubenswrapper[4824]: I1006 10:32:30.645536 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7ea07f3a-f269-4fab-ac00-a9c78fd19865-catalog-content\") pod \"redhat-marketplace-87zqs\" (UID: \"7ea07f3a-f269-4fab-ac00-a9c78fd19865\") " pod="openshift-marketplace/redhat-marketplace-87zqs" Oct 06 10:32:30 crc kubenswrapper[4824]: I1006 10:32:30.661183 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ldsmh\" (UniqueName: \"kubernetes.io/projected/7ea07f3a-f269-4fab-ac00-a9c78fd19865-kube-api-access-ldsmh\") pod \"redhat-marketplace-87zqs\" (UID: \"7ea07f3a-f269-4fab-ac00-a9c78fd19865\") " pod="openshift-marketplace/redhat-marketplace-87zqs" Oct 06 10:32:30 crc kubenswrapper[4824]: I1006 10:32:30.743166 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-87zqs" Oct 06 10:32:30 crc kubenswrapper[4824]: I1006 10:32:30.917755 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-fn42c" event={"ID":"97deb005-9b94-45a5-8d7f-84147ae5807c","Type":"ContainerDied","Data":"44fdf78201a1b3f5fd961a30659b65407e3cbdd4e4fc0ebed8634e2f4d74380d"} Oct 06 10:32:30 crc kubenswrapper[4824]: I1006 10:32:30.917797 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="44fdf78201a1b3f5fd961a30659b65407e3cbdd4e4fc0ebed8634e2f4d74380d" Oct 06 10:32:30 crc kubenswrapper[4824]: I1006 10:32:30.917801 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-fn42c" Oct 06 10:32:31 crc kubenswrapper[4824]: I1006 10:32:31.026315 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-kcljj"] Oct 06 10:32:31 crc kubenswrapper[4824]: I1006 10:32:31.037264 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-kcljj"] Oct 06 10:32:31 crc kubenswrapper[4824]: I1006 10:32:31.037365 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-kcljj" Oct 06 10:32:31 crc kubenswrapper[4824]: I1006 10:32:31.041564 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-f886w" Oct 06 10:32:31 crc kubenswrapper[4824]: I1006 10:32:31.041795 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Oct 06 10:32:31 crc kubenswrapper[4824]: I1006 10:32:31.041905 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Oct 06 10:32:31 crc kubenswrapper[4824]: I1006 10:32:31.042030 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 06 10:32:31 crc kubenswrapper[4824]: I1006 10:32:31.042162 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 06 10:32:31 crc kubenswrapper[4824]: I1006 10:32:31.045742 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 06 10:32:31 crc kubenswrapper[4824]: I1006 10:32:31.060956 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g2jdh\" (UniqueName: \"kubernetes.io/projected/7e9c21f0-458e-493c-8dee-3ca1b591c60b-kube-api-access-g2jdh\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-kcljj\" (UID: \"7e9c21f0-458e-493c-8dee-3ca1b591c60b\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-kcljj" Oct 06 10:32:31 crc kubenswrapper[4824]: I1006 10:32:31.061143 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7e9c21f0-458e-493c-8dee-3ca1b591c60b-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-kcljj\" (UID: \"7e9c21f0-458e-493c-8dee-3ca1b591c60b\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-kcljj" Oct 06 10:32:31 crc kubenswrapper[4824]: I1006 10:32:31.061182 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/7e9c21f0-458e-493c-8dee-3ca1b591c60b-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-kcljj\" (UID: \"7e9c21f0-458e-493c-8dee-3ca1b591c60b\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-kcljj" Oct 06 10:32:31 crc kubenswrapper[4824]: I1006 10:32:31.061224 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7e9c21f0-458e-493c-8dee-3ca1b591c60b-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-kcljj\" (UID: \"7e9c21f0-458e-493c-8dee-3ca1b591c60b\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-kcljj" Oct 06 10:32:31 crc kubenswrapper[4824]: I1006 10:32:31.061299 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e9c21f0-458e-493c-8dee-3ca1b591c60b-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-kcljj\" (UID: \"7e9c21f0-458e-493c-8dee-3ca1b591c60b\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-kcljj" Oct 06 10:32:31 crc kubenswrapper[4824]: I1006 10:32:31.061318 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/7e9c21f0-458e-493c-8dee-3ca1b591c60b-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-kcljj\" (UID: \"7e9c21f0-458e-493c-8dee-3ca1b591c60b\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-kcljj" Oct 06 10:32:31 crc kubenswrapper[4824]: I1006 10:32:31.173158 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/7e9c21f0-458e-493c-8dee-3ca1b591c60b-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-kcljj\" (UID: \"7e9c21f0-458e-493c-8dee-3ca1b591c60b\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-kcljj" Oct 06 10:32:31 crc kubenswrapper[4824]: I1006 10:32:31.189588 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g2jdh\" (UniqueName: \"kubernetes.io/projected/7e9c21f0-458e-493c-8dee-3ca1b591c60b-kube-api-access-g2jdh\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-kcljj\" (UID: \"7e9c21f0-458e-493c-8dee-3ca1b591c60b\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-kcljj" Oct 06 10:32:31 crc kubenswrapper[4824]: I1006 10:32:31.189874 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7e9c21f0-458e-493c-8dee-3ca1b591c60b-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-kcljj\" (UID: \"7e9c21f0-458e-493c-8dee-3ca1b591c60b\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-kcljj" Oct 06 10:32:31 crc kubenswrapper[4824]: I1006 10:32:31.190049 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/7e9c21f0-458e-493c-8dee-3ca1b591c60b-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-kcljj\" (UID: \"7e9c21f0-458e-493c-8dee-3ca1b591c60b\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-kcljj" Oct 06 10:32:31 crc kubenswrapper[4824]: I1006 10:32:31.190213 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7e9c21f0-458e-493c-8dee-3ca1b591c60b-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-kcljj\" (UID: \"7e9c21f0-458e-493c-8dee-3ca1b591c60b\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-kcljj" Oct 06 10:32:31 crc kubenswrapper[4824]: I1006 10:32:31.190484 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e9c21f0-458e-493c-8dee-3ca1b591c60b-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-kcljj\" (UID: \"7e9c21f0-458e-493c-8dee-3ca1b591c60b\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-kcljj" Oct 06 10:32:31 crc kubenswrapper[4824]: I1006 10:32:31.218927 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e9c21f0-458e-493c-8dee-3ca1b591c60b-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-kcljj\" (UID: \"7e9c21f0-458e-493c-8dee-3ca1b591c60b\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-kcljj" Oct 06 10:32:31 crc kubenswrapper[4824]: I1006 10:32:31.219597 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7e9c21f0-458e-493c-8dee-3ca1b591c60b-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-kcljj\" (UID: \"7e9c21f0-458e-493c-8dee-3ca1b591c60b\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-kcljj" Oct 06 10:32:31 crc kubenswrapper[4824]: I1006 10:32:31.221506 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/7e9c21f0-458e-493c-8dee-3ca1b591c60b-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-kcljj\" (UID: \"7e9c21f0-458e-493c-8dee-3ca1b591c60b\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-kcljj" Oct 06 10:32:31 crc kubenswrapper[4824]: I1006 10:32:31.225158 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/7e9c21f0-458e-493c-8dee-3ca1b591c60b-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-kcljj\" (UID: \"7e9c21f0-458e-493c-8dee-3ca1b591c60b\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-kcljj" Oct 06 10:32:31 crc kubenswrapper[4824]: I1006 10:32:31.232832 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g2jdh\" (UniqueName: \"kubernetes.io/projected/7e9c21f0-458e-493c-8dee-3ca1b591c60b-kube-api-access-g2jdh\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-kcljj\" (UID: \"7e9c21f0-458e-493c-8dee-3ca1b591c60b\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-kcljj" Oct 06 10:32:31 crc kubenswrapper[4824]: I1006 10:32:31.238968 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7e9c21f0-458e-493c-8dee-3ca1b591c60b-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-kcljj\" (UID: \"7e9c21f0-458e-493c-8dee-3ca1b591c60b\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-kcljj" Oct 06 10:32:31 crc kubenswrapper[4824]: I1006 10:32:31.357152 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-87zqs"] Oct 06 10:32:31 crc kubenswrapper[4824]: I1006 10:32:31.360592 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-kcljj" Oct 06 10:32:31 crc kubenswrapper[4824]: W1006 10:32:31.362801 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7ea07f3a_f269_4fab_ac00_a9c78fd19865.slice/crio-df1099a23a0f27180010af0d8b8b1d16dd37d8a8efef499154723e1b62942b20 WatchSource:0}: Error finding container df1099a23a0f27180010af0d8b8b1d16dd37d8a8efef499154723e1b62942b20: Status 404 returned error can't find the container with id df1099a23a0f27180010af0d8b8b1d16dd37d8a8efef499154723e1b62942b20 Oct 06 10:32:31 crc kubenswrapper[4824]: I1006 10:32:31.932459 4824 generic.go:334] "Generic (PLEG): container finished" podID="7ea07f3a-f269-4fab-ac00-a9c78fd19865" containerID="7a311f89522e3f1c32e7a5acab70278a403f2f57451afcfe482376e9517252a6" exitCode=0 Oct 06 10:32:31 crc kubenswrapper[4824]: I1006 10:32:31.932555 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-87zqs" event={"ID":"7ea07f3a-f269-4fab-ac00-a9c78fd19865","Type":"ContainerDied","Data":"7a311f89522e3f1c32e7a5acab70278a403f2f57451afcfe482376e9517252a6"} Oct 06 10:32:31 crc kubenswrapper[4824]: I1006 10:32:31.932908 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-87zqs" event={"ID":"7ea07f3a-f269-4fab-ac00-a9c78fd19865","Type":"ContainerStarted","Data":"df1099a23a0f27180010af0d8b8b1d16dd37d8a8efef499154723e1b62942b20"} Oct 06 10:32:31 crc kubenswrapper[4824]: I1006 10:32:31.982402 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-kcljj"] Oct 06 10:32:31 crc kubenswrapper[4824]: W1006 10:32:31.990870 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7e9c21f0_458e_493c_8dee_3ca1b591c60b.slice/crio-aec764b74847058ffa86cc0f738f09ea976fd8cc545ade699c81646fd45b41d8 WatchSource:0}: Error finding container aec764b74847058ffa86cc0f738f09ea976fd8cc545ade699c81646fd45b41d8: Status 404 returned error can't find the container with id aec764b74847058ffa86cc0f738f09ea976fd8cc545ade699c81646fd45b41d8 Oct 06 10:32:32 crc kubenswrapper[4824]: I1006 10:32:32.948428 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-87zqs" event={"ID":"7ea07f3a-f269-4fab-ac00-a9c78fd19865","Type":"ContainerStarted","Data":"90ae0a0ff48929601b0c6b0a7fa10713d0f592994eb8388910dc5e55c8952923"} Oct 06 10:32:32 crc kubenswrapper[4824]: I1006 10:32:32.956581 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-kcljj" event={"ID":"7e9c21f0-458e-493c-8dee-3ca1b591c60b","Type":"ContainerStarted","Data":"8ba21fb73de0fc3b16387b7f6fec1ca60a6b9a6f4e50bed4ef2bc280c2fd1ee2"} Oct 06 10:32:32 crc kubenswrapper[4824]: I1006 10:32:32.956658 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-kcljj" event={"ID":"7e9c21f0-458e-493c-8dee-3ca1b591c60b","Type":"ContainerStarted","Data":"aec764b74847058ffa86cc0f738f09ea976fd8cc545ade699c81646fd45b41d8"} Oct 06 10:32:32 crc kubenswrapper[4824]: I1006 10:32:32.989948 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-kcljj" podStartSLOduration=1.426227085 podStartE2EDuration="1.989923924s" podCreationTimestamp="2025-10-06 10:32:31 +0000 UTC" firstStartedPulling="2025-10-06 10:32:31.993509798 +0000 UTC m=+2061.357932659" lastFinishedPulling="2025-10-06 10:32:32.557206637 +0000 UTC m=+2061.921629498" observedRunningTime="2025-10-06 10:32:32.985685386 +0000 UTC m=+2062.350108257" watchObservedRunningTime="2025-10-06 10:32:32.989923924 +0000 UTC m=+2062.354346805" Oct 06 10:32:33 crc kubenswrapper[4824]: I1006 10:32:33.962933 4824 generic.go:334] "Generic (PLEG): container finished" podID="7ea07f3a-f269-4fab-ac00-a9c78fd19865" containerID="90ae0a0ff48929601b0c6b0a7fa10713d0f592994eb8388910dc5e55c8952923" exitCode=0 Oct 06 10:32:33 crc kubenswrapper[4824]: I1006 10:32:33.963031 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-87zqs" event={"ID":"7ea07f3a-f269-4fab-ac00-a9c78fd19865","Type":"ContainerDied","Data":"90ae0a0ff48929601b0c6b0a7fa10713d0f592994eb8388910dc5e55c8952923"} Oct 06 10:32:33 crc kubenswrapper[4824]: I1006 10:32:33.963622 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-87zqs" event={"ID":"7ea07f3a-f269-4fab-ac00-a9c78fd19865","Type":"ContainerStarted","Data":"b4603ee926864f2753e4b1c63f84ce9ff8c2cf2d5dd3ce7b46dbf2d18b60e252"} Oct 06 10:32:33 crc kubenswrapper[4824]: I1006 10:32:33.987241 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-87zqs" podStartSLOduration=2.550155692 podStartE2EDuration="3.9872193s" podCreationTimestamp="2025-10-06 10:32:30 +0000 UTC" firstStartedPulling="2025-10-06 10:32:31.936122562 +0000 UTC m=+2061.300545463" lastFinishedPulling="2025-10-06 10:32:33.37318621 +0000 UTC m=+2062.737609071" observedRunningTime="2025-10-06 10:32:33.98250301 +0000 UTC m=+2063.346925881" watchObservedRunningTime="2025-10-06 10:32:33.9872193 +0000 UTC m=+2063.351642161" Oct 06 10:32:40 crc kubenswrapper[4824]: I1006 10:32:40.744163 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-87zqs" Oct 06 10:32:40 crc kubenswrapper[4824]: I1006 10:32:40.745246 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-87zqs" Oct 06 10:32:40 crc kubenswrapper[4824]: I1006 10:32:40.805399 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-87zqs" Oct 06 10:32:41 crc kubenswrapper[4824]: I1006 10:32:41.119106 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-87zqs" Oct 06 10:32:41 crc kubenswrapper[4824]: I1006 10:32:41.200923 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-87zqs"] Oct 06 10:32:43 crc kubenswrapper[4824]: I1006 10:32:43.062769 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-87zqs" podUID="7ea07f3a-f269-4fab-ac00-a9c78fd19865" containerName="registry-server" containerID="cri-o://b4603ee926864f2753e4b1c63f84ce9ff8c2cf2d5dd3ce7b46dbf2d18b60e252" gracePeriod=2 Oct 06 10:32:43 crc kubenswrapper[4824]: I1006 10:32:43.516625 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-87zqs" Oct 06 10:32:43 crc kubenswrapper[4824]: I1006 10:32:43.578320 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7ea07f3a-f269-4fab-ac00-a9c78fd19865-catalog-content\") pod \"7ea07f3a-f269-4fab-ac00-a9c78fd19865\" (UID: \"7ea07f3a-f269-4fab-ac00-a9c78fd19865\") " Oct 06 10:32:43 crc kubenswrapper[4824]: I1006 10:32:43.578895 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7ea07f3a-f269-4fab-ac00-a9c78fd19865-utilities\") pod \"7ea07f3a-f269-4fab-ac00-a9c78fd19865\" (UID: \"7ea07f3a-f269-4fab-ac00-a9c78fd19865\") " Oct 06 10:32:43 crc kubenswrapper[4824]: I1006 10:32:43.579145 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ldsmh\" (UniqueName: \"kubernetes.io/projected/7ea07f3a-f269-4fab-ac00-a9c78fd19865-kube-api-access-ldsmh\") pod \"7ea07f3a-f269-4fab-ac00-a9c78fd19865\" (UID: \"7ea07f3a-f269-4fab-ac00-a9c78fd19865\") " Oct 06 10:32:43 crc kubenswrapper[4824]: I1006 10:32:43.581173 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7ea07f3a-f269-4fab-ac00-a9c78fd19865-utilities" (OuterVolumeSpecName: "utilities") pod "7ea07f3a-f269-4fab-ac00-a9c78fd19865" (UID: "7ea07f3a-f269-4fab-ac00-a9c78fd19865"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:32:43 crc kubenswrapper[4824]: I1006 10:32:43.595782 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7ea07f3a-f269-4fab-ac00-a9c78fd19865-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7ea07f3a-f269-4fab-ac00-a9c78fd19865" (UID: "7ea07f3a-f269-4fab-ac00-a9c78fd19865"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:32:43 crc kubenswrapper[4824]: I1006 10:32:43.599406 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ea07f3a-f269-4fab-ac00-a9c78fd19865-kube-api-access-ldsmh" (OuterVolumeSpecName: "kube-api-access-ldsmh") pod "7ea07f3a-f269-4fab-ac00-a9c78fd19865" (UID: "7ea07f3a-f269-4fab-ac00-a9c78fd19865"). InnerVolumeSpecName "kube-api-access-ldsmh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:32:43 crc kubenswrapper[4824]: I1006 10:32:43.681538 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7ea07f3a-f269-4fab-ac00-a9c78fd19865-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 06 10:32:43 crc kubenswrapper[4824]: I1006 10:32:43.681586 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7ea07f3a-f269-4fab-ac00-a9c78fd19865-utilities\") on node \"crc\" DevicePath \"\"" Oct 06 10:32:43 crc kubenswrapper[4824]: I1006 10:32:43.681600 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ldsmh\" (UniqueName: \"kubernetes.io/projected/7ea07f3a-f269-4fab-ac00-a9c78fd19865-kube-api-access-ldsmh\") on node \"crc\" DevicePath \"\"" Oct 06 10:32:44 crc kubenswrapper[4824]: I1006 10:32:44.076324 4824 generic.go:334] "Generic (PLEG): container finished" podID="7ea07f3a-f269-4fab-ac00-a9c78fd19865" containerID="b4603ee926864f2753e4b1c63f84ce9ff8c2cf2d5dd3ce7b46dbf2d18b60e252" exitCode=0 Oct 06 10:32:44 crc kubenswrapper[4824]: I1006 10:32:44.076380 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-87zqs" event={"ID":"7ea07f3a-f269-4fab-ac00-a9c78fd19865","Type":"ContainerDied","Data":"b4603ee926864f2753e4b1c63f84ce9ff8c2cf2d5dd3ce7b46dbf2d18b60e252"} Oct 06 10:32:44 crc kubenswrapper[4824]: I1006 10:32:44.076415 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-87zqs" event={"ID":"7ea07f3a-f269-4fab-ac00-a9c78fd19865","Type":"ContainerDied","Data":"df1099a23a0f27180010af0d8b8b1d16dd37d8a8efef499154723e1b62942b20"} Oct 06 10:32:44 crc kubenswrapper[4824]: I1006 10:32:44.076440 4824 scope.go:117] "RemoveContainer" containerID="b4603ee926864f2753e4b1c63f84ce9ff8c2cf2d5dd3ce7b46dbf2d18b60e252" Oct 06 10:32:44 crc kubenswrapper[4824]: I1006 10:32:44.076438 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-87zqs" Oct 06 10:32:44 crc kubenswrapper[4824]: I1006 10:32:44.113617 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-87zqs"] Oct 06 10:32:44 crc kubenswrapper[4824]: I1006 10:32:44.116858 4824 scope.go:117] "RemoveContainer" containerID="90ae0a0ff48929601b0c6b0a7fa10713d0f592994eb8388910dc5e55c8952923" Oct 06 10:32:44 crc kubenswrapper[4824]: I1006 10:32:44.122928 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-87zqs"] Oct 06 10:32:44 crc kubenswrapper[4824]: I1006 10:32:44.141686 4824 scope.go:117] "RemoveContainer" containerID="7a311f89522e3f1c32e7a5acab70278a403f2f57451afcfe482376e9517252a6" Oct 06 10:32:44 crc kubenswrapper[4824]: I1006 10:32:44.193679 4824 scope.go:117] "RemoveContainer" containerID="b4603ee926864f2753e4b1c63f84ce9ff8c2cf2d5dd3ce7b46dbf2d18b60e252" Oct 06 10:32:44 crc kubenswrapper[4824]: E1006 10:32:44.194138 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b4603ee926864f2753e4b1c63f84ce9ff8c2cf2d5dd3ce7b46dbf2d18b60e252\": container with ID starting with b4603ee926864f2753e4b1c63f84ce9ff8c2cf2d5dd3ce7b46dbf2d18b60e252 not found: ID does not exist" containerID="b4603ee926864f2753e4b1c63f84ce9ff8c2cf2d5dd3ce7b46dbf2d18b60e252" Oct 06 10:32:44 crc kubenswrapper[4824]: I1006 10:32:44.194180 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b4603ee926864f2753e4b1c63f84ce9ff8c2cf2d5dd3ce7b46dbf2d18b60e252"} err="failed to get container status \"b4603ee926864f2753e4b1c63f84ce9ff8c2cf2d5dd3ce7b46dbf2d18b60e252\": rpc error: code = NotFound desc = could not find container \"b4603ee926864f2753e4b1c63f84ce9ff8c2cf2d5dd3ce7b46dbf2d18b60e252\": container with ID starting with b4603ee926864f2753e4b1c63f84ce9ff8c2cf2d5dd3ce7b46dbf2d18b60e252 not found: ID does not exist" Oct 06 10:32:44 crc kubenswrapper[4824]: I1006 10:32:44.194205 4824 scope.go:117] "RemoveContainer" containerID="90ae0a0ff48929601b0c6b0a7fa10713d0f592994eb8388910dc5e55c8952923" Oct 06 10:32:44 crc kubenswrapper[4824]: E1006 10:32:44.194490 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"90ae0a0ff48929601b0c6b0a7fa10713d0f592994eb8388910dc5e55c8952923\": container with ID starting with 90ae0a0ff48929601b0c6b0a7fa10713d0f592994eb8388910dc5e55c8952923 not found: ID does not exist" containerID="90ae0a0ff48929601b0c6b0a7fa10713d0f592994eb8388910dc5e55c8952923" Oct 06 10:32:44 crc kubenswrapper[4824]: I1006 10:32:44.194521 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"90ae0a0ff48929601b0c6b0a7fa10713d0f592994eb8388910dc5e55c8952923"} err="failed to get container status \"90ae0a0ff48929601b0c6b0a7fa10713d0f592994eb8388910dc5e55c8952923\": rpc error: code = NotFound desc = could not find container \"90ae0a0ff48929601b0c6b0a7fa10713d0f592994eb8388910dc5e55c8952923\": container with ID starting with 90ae0a0ff48929601b0c6b0a7fa10713d0f592994eb8388910dc5e55c8952923 not found: ID does not exist" Oct 06 10:32:44 crc kubenswrapper[4824]: I1006 10:32:44.194540 4824 scope.go:117] "RemoveContainer" containerID="7a311f89522e3f1c32e7a5acab70278a403f2f57451afcfe482376e9517252a6" Oct 06 10:32:44 crc kubenswrapper[4824]: E1006 10:32:44.194999 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7a311f89522e3f1c32e7a5acab70278a403f2f57451afcfe482376e9517252a6\": container with ID starting with 7a311f89522e3f1c32e7a5acab70278a403f2f57451afcfe482376e9517252a6 not found: ID does not exist" containerID="7a311f89522e3f1c32e7a5acab70278a403f2f57451afcfe482376e9517252a6" Oct 06 10:32:44 crc kubenswrapper[4824]: I1006 10:32:44.195056 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7a311f89522e3f1c32e7a5acab70278a403f2f57451afcfe482376e9517252a6"} err="failed to get container status \"7a311f89522e3f1c32e7a5acab70278a403f2f57451afcfe482376e9517252a6\": rpc error: code = NotFound desc = could not find container \"7a311f89522e3f1c32e7a5acab70278a403f2f57451afcfe482376e9517252a6\": container with ID starting with 7a311f89522e3f1c32e7a5acab70278a403f2f57451afcfe482376e9517252a6 not found: ID does not exist" Oct 06 10:32:45 crc kubenswrapper[4824]: I1006 10:32:45.289787 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7ea07f3a-f269-4fab-ac00-a9c78fd19865" path="/var/lib/kubelet/pods/7ea07f3a-f269-4fab-ac00-a9c78fd19865/volumes" Oct 06 10:33:23 crc kubenswrapper[4824]: I1006 10:33:23.474449 4824 generic.go:334] "Generic (PLEG): container finished" podID="7e9c21f0-458e-493c-8dee-3ca1b591c60b" containerID="8ba21fb73de0fc3b16387b7f6fec1ca60a6b9a6f4e50bed4ef2bc280c2fd1ee2" exitCode=0 Oct 06 10:33:23 crc kubenswrapper[4824]: I1006 10:33:23.474565 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-kcljj" event={"ID":"7e9c21f0-458e-493c-8dee-3ca1b591c60b","Type":"ContainerDied","Data":"8ba21fb73de0fc3b16387b7f6fec1ca60a6b9a6f4e50bed4ef2bc280c2fd1ee2"} Oct 06 10:33:24 crc kubenswrapper[4824]: I1006 10:33:24.898831 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-kcljj" Oct 06 10:33:24 crc kubenswrapper[4824]: I1006 10:33:24.967793 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7e9c21f0-458e-493c-8dee-3ca1b591c60b-ssh-key\") pod \"7e9c21f0-458e-493c-8dee-3ca1b591c60b\" (UID: \"7e9c21f0-458e-493c-8dee-3ca1b591c60b\") " Oct 06 10:33:24 crc kubenswrapper[4824]: I1006 10:33:24.967887 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e9c21f0-458e-493c-8dee-3ca1b591c60b-neutron-metadata-combined-ca-bundle\") pod \"7e9c21f0-458e-493c-8dee-3ca1b591c60b\" (UID: \"7e9c21f0-458e-493c-8dee-3ca1b591c60b\") " Oct 06 10:33:24 crc kubenswrapper[4824]: I1006 10:33:24.967926 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7e9c21f0-458e-493c-8dee-3ca1b591c60b-inventory\") pod \"7e9c21f0-458e-493c-8dee-3ca1b591c60b\" (UID: \"7e9c21f0-458e-493c-8dee-3ca1b591c60b\") " Oct 06 10:33:24 crc kubenswrapper[4824]: I1006 10:33:24.968131 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g2jdh\" (UniqueName: \"kubernetes.io/projected/7e9c21f0-458e-493c-8dee-3ca1b591c60b-kube-api-access-g2jdh\") pod \"7e9c21f0-458e-493c-8dee-3ca1b591c60b\" (UID: \"7e9c21f0-458e-493c-8dee-3ca1b591c60b\") " Oct 06 10:33:24 crc kubenswrapper[4824]: I1006 10:33:24.968178 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/7e9c21f0-458e-493c-8dee-3ca1b591c60b-neutron-ovn-metadata-agent-neutron-config-0\") pod \"7e9c21f0-458e-493c-8dee-3ca1b591c60b\" (UID: \"7e9c21f0-458e-493c-8dee-3ca1b591c60b\") " Oct 06 10:33:24 crc kubenswrapper[4824]: I1006 10:33:24.968220 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/7e9c21f0-458e-493c-8dee-3ca1b591c60b-nova-metadata-neutron-config-0\") pod \"7e9c21f0-458e-493c-8dee-3ca1b591c60b\" (UID: \"7e9c21f0-458e-493c-8dee-3ca1b591c60b\") " Oct 06 10:33:24 crc kubenswrapper[4824]: I1006 10:33:24.973812 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7e9c21f0-458e-493c-8dee-3ca1b591c60b-kube-api-access-g2jdh" (OuterVolumeSpecName: "kube-api-access-g2jdh") pod "7e9c21f0-458e-493c-8dee-3ca1b591c60b" (UID: "7e9c21f0-458e-493c-8dee-3ca1b591c60b"). InnerVolumeSpecName "kube-api-access-g2jdh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:33:24 crc kubenswrapper[4824]: I1006 10:33:24.974855 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7e9c21f0-458e-493c-8dee-3ca1b591c60b-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "7e9c21f0-458e-493c-8dee-3ca1b591c60b" (UID: "7e9c21f0-458e-493c-8dee-3ca1b591c60b"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:33:24 crc kubenswrapper[4824]: I1006 10:33:24.998855 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7e9c21f0-458e-493c-8dee-3ca1b591c60b-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "7e9c21f0-458e-493c-8dee-3ca1b591c60b" (UID: "7e9c21f0-458e-493c-8dee-3ca1b591c60b"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:33:25 crc kubenswrapper[4824]: I1006 10:33:25.000151 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7e9c21f0-458e-493c-8dee-3ca1b591c60b-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "7e9c21f0-458e-493c-8dee-3ca1b591c60b" (UID: "7e9c21f0-458e-493c-8dee-3ca1b591c60b"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:33:25 crc kubenswrapper[4824]: I1006 10:33:25.002961 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7e9c21f0-458e-493c-8dee-3ca1b591c60b-inventory" (OuterVolumeSpecName: "inventory") pod "7e9c21f0-458e-493c-8dee-3ca1b591c60b" (UID: "7e9c21f0-458e-493c-8dee-3ca1b591c60b"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:33:25 crc kubenswrapper[4824]: I1006 10:33:25.015083 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7e9c21f0-458e-493c-8dee-3ca1b591c60b-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "7e9c21f0-458e-493c-8dee-3ca1b591c60b" (UID: "7e9c21f0-458e-493c-8dee-3ca1b591c60b"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:33:25 crc kubenswrapper[4824]: I1006 10:33:25.070040 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g2jdh\" (UniqueName: \"kubernetes.io/projected/7e9c21f0-458e-493c-8dee-3ca1b591c60b-kube-api-access-g2jdh\") on node \"crc\" DevicePath \"\"" Oct 06 10:33:25 crc kubenswrapper[4824]: I1006 10:33:25.070616 4824 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/7e9c21f0-458e-493c-8dee-3ca1b591c60b-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Oct 06 10:33:25 crc kubenswrapper[4824]: I1006 10:33:25.070746 4824 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/7e9c21f0-458e-493c-8dee-3ca1b591c60b-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Oct 06 10:33:25 crc kubenswrapper[4824]: I1006 10:33:25.070872 4824 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7e9c21f0-458e-493c-8dee-3ca1b591c60b-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 06 10:33:25 crc kubenswrapper[4824]: I1006 10:33:25.070973 4824 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e9c21f0-458e-493c-8dee-3ca1b591c60b-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 10:33:25 crc kubenswrapper[4824]: I1006 10:33:25.071121 4824 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7e9c21f0-458e-493c-8dee-3ca1b591c60b-inventory\") on node \"crc\" DevicePath \"\"" Oct 06 10:33:25 crc kubenswrapper[4824]: I1006 10:33:25.496802 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-kcljj" event={"ID":"7e9c21f0-458e-493c-8dee-3ca1b591c60b","Type":"ContainerDied","Data":"aec764b74847058ffa86cc0f738f09ea976fd8cc545ade699c81646fd45b41d8"} Oct 06 10:33:25 crc kubenswrapper[4824]: I1006 10:33:25.496860 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="aec764b74847058ffa86cc0f738f09ea976fd8cc545ade699c81646fd45b41d8" Oct 06 10:33:25 crc kubenswrapper[4824]: I1006 10:33:25.497015 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-kcljj" Oct 06 10:33:25 crc kubenswrapper[4824]: I1006 10:33:25.592655 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8s745"] Oct 06 10:33:25 crc kubenswrapper[4824]: E1006 10:33:25.593387 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ea07f3a-f269-4fab-ac00-a9c78fd19865" containerName="registry-server" Oct 06 10:33:25 crc kubenswrapper[4824]: I1006 10:33:25.593416 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ea07f3a-f269-4fab-ac00-a9c78fd19865" containerName="registry-server" Oct 06 10:33:25 crc kubenswrapper[4824]: E1006 10:33:25.593463 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ea07f3a-f269-4fab-ac00-a9c78fd19865" containerName="extract-utilities" Oct 06 10:33:25 crc kubenswrapper[4824]: I1006 10:33:25.593473 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ea07f3a-f269-4fab-ac00-a9c78fd19865" containerName="extract-utilities" Oct 06 10:33:25 crc kubenswrapper[4824]: E1006 10:33:25.593490 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e9c21f0-458e-493c-8dee-3ca1b591c60b" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Oct 06 10:33:25 crc kubenswrapper[4824]: I1006 10:33:25.593500 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e9c21f0-458e-493c-8dee-3ca1b591c60b" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Oct 06 10:33:25 crc kubenswrapper[4824]: E1006 10:33:25.593525 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ea07f3a-f269-4fab-ac00-a9c78fd19865" containerName="extract-content" Oct 06 10:33:25 crc kubenswrapper[4824]: I1006 10:33:25.593534 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ea07f3a-f269-4fab-ac00-a9c78fd19865" containerName="extract-content" Oct 06 10:33:25 crc kubenswrapper[4824]: I1006 10:33:25.593773 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ea07f3a-f269-4fab-ac00-a9c78fd19865" containerName="registry-server" Oct 06 10:33:25 crc kubenswrapper[4824]: I1006 10:33:25.593806 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="7e9c21f0-458e-493c-8dee-3ca1b591c60b" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Oct 06 10:33:25 crc kubenswrapper[4824]: I1006 10:33:25.594595 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8s745" Oct 06 10:33:25 crc kubenswrapper[4824]: I1006 10:33:25.603489 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-f886w" Oct 06 10:33:25 crc kubenswrapper[4824]: I1006 10:33:25.603576 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 06 10:33:25 crc kubenswrapper[4824]: I1006 10:33:25.603761 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Oct 06 10:33:25 crc kubenswrapper[4824]: I1006 10:33:25.603938 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 06 10:33:25 crc kubenswrapper[4824]: I1006 10:33:25.604080 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 06 10:33:25 crc kubenswrapper[4824]: I1006 10:33:25.617055 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8s745"] Oct 06 10:33:25 crc kubenswrapper[4824]: I1006 10:33:25.683544 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4555aa06-f941-492e-9a1c-89a7c1a5c50f-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-8s745\" (UID: \"4555aa06-f941-492e-9a1c-89a7c1a5c50f\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8s745" Oct 06 10:33:25 crc kubenswrapper[4824]: I1006 10:33:25.683608 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6ctqj\" (UniqueName: \"kubernetes.io/projected/4555aa06-f941-492e-9a1c-89a7c1a5c50f-kube-api-access-6ctqj\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-8s745\" (UID: \"4555aa06-f941-492e-9a1c-89a7c1a5c50f\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8s745" Oct 06 10:33:25 crc kubenswrapper[4824]: I1006 10:33:25.683645 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4555aa06-f941-492e-9a1c-89a7c1a5c50f-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-8s745\" (UID: \"4555aa06-f941-492e-9a1c-89a7c1a5c50f\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8s745" Oct 06 10:33:25 crc kubenswrapper[4824]: I1006 10:33:25.683700 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4555aa06-f941-492e-9a1c-89a7c1a5c50f-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-8s745\" (UID: \"4555aa06-f941-492e-9a1c-89a7c1a5c50f\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8s745" Oct 06 10:33:25 crc kubenswrapper[4824]: I1006 10:33:25.683852 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/4555aa06-f941-492e-9a1c-89a7c1a5c50f-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-8s745\" (UID: \"4555aa06-f941-492e-9a1c-89a7c1a5c50f\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8s745" Oct 06 10:33:25 crc kubenswrapper[4824]: I1006 10:33:25.785872 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/4555aa06-f941-492e-9a1c-89a7c1a5c50f-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-8s745\" (UID: \"4555aa06-f941-492e-9a1c-89a7c1a5c50f\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8s745" Oct 06 10:33:25 crc kubenswrapper[4824]: I1006 10:33:25.785914 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4555aa06-f941-492e-9a1c-89a7c1a5c50f-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-8s745\" (UID: \"4555aa06-f941-492e-9a1c-89a7c1a5c50f\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8s745" Oct 06 10:33:25 crc kubenswrapper[4824]: I1006 10:33:25.785941 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6ctqj\" (UniqueName: \"kubernetes.io/projected/4555aa06-f941-492e-9a1c-89a7c1a5c50f-kube-api-access-6ctqj\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-8s745\" (UID: \"4555aa06-f941-492e-9a1c-89a7c1a5c50f\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8s745" Oct 06 10:33:25 crc kubenswrapper[4824]: I1006 10:33:25.785962 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4555aa06-f941-492e-9a1c-89a7c1a5c50f-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-8s745\" (UID: \"4555aa06-f941-492e-9a1c-89a7c1a5c50f\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8s745" Oct 06 10:33:25 crc kubenswrapper[4824]: I1006 10:33:25.786008 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4555aa06-f941-492e-9a1c-89a7c1a5c50f-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-8s745\" (UID: \"4555aa06-f941-492e-9a1c-89a7c1a5c50f\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8s745" Oct 06 10:33:25 crc kubenswrapper[4824]: I1006 10:33:25.791136 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/4555aa06-f941-492e-9a1c-89a7c1a5c50f-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-8s745\" (UID: \"4555aa06-f941-492e-9a1c-89a7c1a5c50f\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8s745" Oct 06 10:33:25 crc kubenswrapper[4824]: I1006 10:33:25.791597 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4555aa06-f941-492e-9a1c-89a7c1a5c50f-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-8s745\" (UID: \"4555aa06-f941-492e-9a1c-89a7c1a5c50f\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8s745" Oct 06 10:33:25 crc kubenswrapper[4824]: I1006 10:33:25.796660 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4555aa06-f941-492e-9a1c-89a7c1a5c50f-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-8s745\" (UID: \"4555aa06-f941-492e-9a1c-89a7c1a5c50f\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8s745" Oct 06 10:33:25 crc kubenswrapper[4824]: I1006 10:33:25.798680 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4555aa06-f941-492e-9a1c-89a7c1a5c50f-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-8s745\" (UID: \"4555aa06-f941-492e-9a1c-89a7c1a5c50f\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8s745" Oct 06 10:33:25 crc kubenswrapper[4824]: I1006 10:33:25.811196 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6ctqj\" (UniqueName: \"kubernetes.io/projected/4555aa06-f941-492e-9a1c-89a7c1a5c50f-kube-api-access-6ctqj\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-8s745\" (UID: \"4555aa06-f941-492e-9a1c-89a7c1a5c50f\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8s745" Oct 06 10:33:25 crc kubenswrapper[4824]: I1006 10:33:25.929343 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8s745" Oct 06 10:33:26 crc kubenswrapper[4824]: I1006 10:33:26.364357 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8s745"] Oct 06 10:33:26 crc kubenswrapper[4824]: I1006 10:33:26.508374 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8s745" event={"ID":"4555aa06-f941-492e-9a1c-89a7c1a5c50f","Type":"ContainerStarted","Data":"ab77fec3da2c3f4cf1540862ce0c4867b0ca77ca34867306bec12a1236ba7d86"} Oct 06 10:33:27 crc kubenswrapper[4824]: I1006 10:33:27.522122 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8s745" event={"ID":"4555aa06-f941-492e-9a1c-89a7c1a5c50f","Type":"ContainerStarted","Data":"d5e9b3fdf42b3aae98b95483972d106d334f9889ec76c5f4df490a19e1aa3de8"} Oct 06 10:33:27 crc kubenswrapper[4824]: I1006 10:33:27.543534 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8s745" podStartSLOduration=2.0400596 podStartE2EDuration="2.543517375s" podCreationTimestamp="2025-10-06 10:33:25 +0000 UTC" firstStartedPulling="2025-10-06 10:33:26.369254038 +0000 UTC m=+2115.733676899" lastFinishedPulling="2025-10-06 10:33:26.872711813 +0000 UTC m=+2116.237134674" observedRunningTime="2025-10-06 10:33:27.5372469 +0000 UTC m=+2116.901669761" watchObservedRunningTime="2025-10-06 10:33:27.543517375 +0000 UTC m=+2116.907940236" Oct 06 10:33:43 crc kubenswrapper[4824]: I1006 10:33:43.915398 4824 patch_prober.go:28] interesting pod/machine-config-daemon-khgzw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 10:33:43 crc kubenswrapper[4824]: I1006 10:33:43.915895 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 10:34:13 crc kubenswrapper[4824]: I1006 10:34:13.915086 4824 patch_prober.go:28] interesting pod/machine-config-daemon-khgzw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 10:34:13 crc kubenswrapper[4824]: I1006 10:34:13.915687 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 10:34:43 crc kubenswrapper[4824]: I1006 10:34:43.915202 4824 patch_prober.go:28] interesting pod/machine-config-daemon-khgzw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 10:34:43 crc kubenswrapper[4824]: I1006 10:34:43.915789 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 10:34:43 crc kubenswrapper[4824]: I1006 10:34:43.915837 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" Oct 06 10:34:43 crc kubenswrapper[4824]: I1006 10:34:43.916464 4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2e9a092782b8371eeb30f1c342e8665ca42002786b61ad0319bff781e0406a48"} pod="openshift-machine-config-operator/machine-config-daemon-khgzw" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 06 10:34:43 crc kubenswrapper[4824]: I1006 10:34:43.916531 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" containerName="machine-config-daemon" containerID="cri-o://2e9a092782b8371eeb30f1c342e8665ca42002786b61ad0319bff781e0406a48" gracePeriod=600 Oct 06 10:34:44 crc kubenswrapper[4824]: I1006 10:34:44.322661 4824 generic.go:334] "Generic (PLEG): container finished" podID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" containerID="2e9a092782b8371eeb30f1c342e8665ca42002786b61ad0319bff781e0406a48" exitCode=0 Oct 06 10:34:44 crc kubenswrapper[4824]: I1006 10:34:44.322734 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" event={"ID":"e1140c8c-93f1-4fce-8c06-a69561ff8a6d","Type":"ContainerDied","Data":"2e9a092782b8371eeb30f1c342e8665ca42002786b61ad0319bff781e0406a48"} Oct 06 10:34:44 crc kubenswrapper[4824]: I1006 10:34:44.323135 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" event={"ID":"e1140c8c-93f1-4fce-8c06-a69561ff8a6d","Type":"ContainerStarted","Data":"defbb9e5d617c326a5dda4c68f7311eec28a02ec245ca3527fbdf00979129a82"} Oct 06 10:34:44 crc kubenswrapper[4824]: I1006 10:34:44.323158 4824 scope.go:117] "RemoveContainer" containerID="9de43c7f48ed9ad071a24adf937829ab92276bd505bb72d5b89bac4922aac3b0" Oct 06 10:37:13 crc kubenswrapper[4824]: I1006 10:37:13.625303 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-bbq5l"] Oct 06 10:37:13 crc kubenswrapper[4824]: I1006 10:37:13.627798 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bbq5l" Oct 06 10:37:13 crc kubenswrapper[4824]: I1006 10:37:13.640996 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bbq5l"] Oct 06 10:37:13 crc kubenswrapper[4824]: I1006 10:37:13.747334 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f4b6ee8e-9e36-422c-95ab-35d166c03844-utilities\") pod \"redhat-operators-bbq5l\" (UID: \"f4b6ee8e-9e36-422c-95ab-35d166c03844\") " pod="openshift-marketplace/redhat-operators-bbq5l" Oct 06 10:37:13 crc kubenswrapper[4824]: I1006 10:37:13.747687 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9gftx\" (UniqueName: \"kubernetes.io/projected/f4b6ee8e-9e36-422c-95ab-35d166c03844-kube-api-access-9gftx\") pod \"redhat-operators-bbq5l\" (UID: \"f4b6ee8e-9e36-422c-95ab-35d166c03844\") " pod="openshift-marketplace/redhat-operators-bbq5l" Oct 06 10:37:13 crc kubenswrapper[4824]: I1006 10:37:13.747934 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f4b6ee8e-9e36-422c-95ab-35d166c03844-catalog-content\") pod \"redhat-operators-bbq5l\" (UID: \"f4b6ee8e-9e36-422c-95ab-35d166c03844\") " pod="openshift-marketplace/redhat-operators-bbq5l" Oct 06 10:37:13 crc kubenswrapper[4824]: I1006 10:37:13.850315 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f4b6ee8e-9e36-422c-95ab-35d166c03844-catalog-content\") pod \"redhat-operators-bbq5l\" (UID: \"f4b6ee8e-9e36-422c-95ab-35d166c03844\") " pod="openshift-marketplace/redhat-operators-bbq5l" Oct 06 10:37:13 crc kubenswrapper[4824]: I1006 10:37:13.850424 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f4b6ee8e-9e36-422c-95ab-35d166c03844-utilities\") pod \"redhat-operators-bbq5l\" (UID: \"f4b6ee8e-9e36-422c-95ab-35d166c03844\") " pod="openshift-marketplace/redhat-operators-bbq5l" Oct 06 10:37:13 crc kubenswrapper[4824]: I1006 10:37:13.850511 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9gftx\" (UniqueName: \"kubernetes.io/projected/f4b6ee8e-9e36-422c-95ab-35d166c03844-kube-api-access-9gftx\") pod \"redhat-operators-bbq5l\" (UID: \"f4b6ee8e-9e36-422c-95ab-35d166c03844\") " pod="openshift-marketplace/redhat-operators-bbq5l" Oct 06 10:37:13 crc kubenswrapper[4824]: I1006 10:37:13.850877 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f4b6ee8e-9e36-422c-95ab-35d166c03844-catalog-content\") pod \"redhat-operators-bbq5l\" (UID: \"f4b6ee8e-9e36-422c-95ab-35d166c03844\") " pod="openshift-marketplace/redhat-operators-bbq5l" Oct 06 10:37:13 crc kubenswrapper[4824]: I1006 10:37:13.850971 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f4b6ee8e-9e36-422c-95ab-35d166c03844-utilities\") pod \"redhat-operators-bbq5l\" (UID: \"f4b6ee8e-9e36-422c-95ab-35d166c03844\") " pod="openshift-marketplace/redhat-operators-bbq5l" Oct 06 10:37:13 crc kubenswrapper[4824]: I1006 10:37:13.880018 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9gftx\" (UniqueName: \"kubernetes.io/projected/f4b6ee8e-9e36-422c-95ab-35d166c03844-kube-api-access-9gftx\") pod \"redhat-operators-bbq5l\" (UID: \"f4b6ee8e-9e36-422c-95ab-35d166c03844\") " pod="openshift-marketplace/redhat-operators-bbq5l" Oct 06 10:37:13 crc kubenswrapper[4824]: I1006 10:37:13.915745 4824 patch_prober.go:28] interesting pod/machine-config-daemon-khgzw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 10:37:13 crc kubenswrapper[4824]: I1006 10:37:13.915797 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 10:37:13 crc kubenswrapper[4824]: I1006 10:37:13.976182 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bbq5l" Oct 06 10:37:14 crc kubenswrapper[4824]: I1006 10:37:14.460553 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bbq5l"] Oct 06 10:37:14 crc kubenswrapper[4824]: I1006 10:37:14.996003 4824 generic.go:334] "Generic (PLEG): container finished" podID="f4b6ee8e-9e36-422c-95ab-35d166c03844" containerID="dc856aae7130b2fa93cccc4e3cf29258e2be7a14af3f29a38d088398b6dbfdb6" exitCode=0 Oct 06 10:37:15 crc kubenswrapper[4824]: I1006 10:37:14.996093 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bbq5l" event={"ID":"f4b6ee8e-9e36-422c-95ab-35d166c03844","Type":"ContainerDied","Data":"dc856aae7130b2fa93cccc4e3cf29258e2be7a14af3f29a38d088398b6dbfdb6"} Oct 06 10:37:15 crc kubenswrapper[4824]: I1006 10:37:14.996937 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bbq5l" event={"ID":"f4b6ee8e-9e36-422c-95ab-35d166c03844","Type":"ContainerStarted","Data":"0529842ead62cfb39b0031bec2fc3a979b949d9e763d0e519a3ead978ebc9545"} Oct 06 10:37:15 crc kubenswrapper[4824]: I1006 10:37:14.999144 4824 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 06 10:37:17 crc kubenswrapper[4824]: I1006 10:37:17.025886 4824 generic.go:334] "Generic (PLEG): container finished" podID="f4b6ee8e-9e36-422c-95ab-35d166c03844" containerID="3a361dd7fc00103945f09b2ffc4864ba8f1aaec25e046e353e2f93b4d9c2cf35" exitCode=0 Oct 06 10:37:17 crc kubenswrapper[4824]: I1006 10:37:17.025959 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bbq5l" event={"ID":"f4b6ee8e-9e36-422c-95ab-35d166c03844","Type":"ContainerDied","Data":"3a361dd7fc00103945f09b2ffc4864ba8f1aaec25e046e353e2f93b4d9c2cf35"} Oct 06 10:37:18 crc kubenswrapper[4824]: I1006 10:37:18.036436 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bbq5l" event={"ID":"f4b6ee8e-9e36-422c-95ab-35d166c03844","Type":"ContainerStarted","Data":"3a703a7f87ec11dc9ef4d8308bbc544c68c9ee431448f97b626d43b69b7690a3"} Oct 06 10:37:18 crc kubenswrapper[4824]: I1006 10:37:18.066398 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-bbq5l" podStartSLOduration=2.402082055 podStartE2EDuration="5.066350022s" podCreationTimestamp="2025-10-06 10:37:13 +0000 UTC" firstStartedPulling="2025-10-06 10:37:14.998818712 +0000 UTC m=+2344.363241573" lastFinishedPulling="2025-10-06 10:37:17.663086669 +0000 UTC m=+2347.027509540" observedRunningTime="2025-10-06 10:37:18.062854698 +0000 UTC m=+2347.427277579" watchObservedRunningTime="2025-10-06 10:37:18.066350022 +0000 UTC m=+2347.430772903" Oct 06 10:37:23 crc kubenswrapper[4824]: I1006 10:37:23.977267 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-bbq5l" Oct 06 10:37:23 crc kubenswrapper[4824]: I1006 10:37:23.977817 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-bbq5l" Oct 06 10:37:24 crc kubenswrapper[4824]: I1006 10:37:24.047975 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-bbq5l" Oct 06 10:37:24 crc kubenswrapper[4824]: I1006 10:37:24.170065 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-bbq5l" Oct 06 10:37:24 crc kubenswrapper[4824]: I1006 10:37:24.314881 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-bbq5l"] Oct 06 10:37:26 crc kubenswrapper[4824]: I1006 10:37:26.123612 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-bbq5l" podUID="f4b6ee8e-9e36-422c-95ab-35d166c03844" containerName="registry-server" containerID="cri-o://3a703a7f87ec11dc9ef4d8308bbc544c68c9ee431448f97b626d43b69b7690a3" gracePeriod=2 Oct 06 10:37:26 crc kubenswrapper[4824]: I1006 10:37:26.682387 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bbq5l" Oct 06 10:37:26 crc kubenswrapper[4824]: I1006 10:37:26.722870 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9gftx\" (UniqueName: \"kubernetes.io/projected/f4b6ee8e-9e36-422c-95ab-35d166c03844-kube-api-access-9gftx\") pod \"f4b6ee8e-9e36-422c-95ab-35d166c03844\" (UID: \"f4b6ee8e-9e36-422c-95ab-35d166c03844\") " Oct 06 10:37:26 crc kubenswrapper[4824]: I1006 10:37:26.723052 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f4b6ee8e-9e36-422c-95ab-35d166c03844-utilities\") pod \"f4b6ee8e-9e36-422c-95ab-35d166c03844\" (UID: \"f4b6ee8e-9e36-422c-95ab-35d166c03844\") " Oct 06 10:37:26 crc kubenswrapper[4824]: I1006 10:37:26.723123 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f4b6ee8e-9e36-422c-95ab-35d166c03844-catalog-content\") pod \"f4b6ee8e-9e36-422c-95ab-35d166c03844\" (UID: \"f4b6ee8e-9e36-422c-95ab-35d166c03844\") " Oct 06 10:37:26 crc kubenswrapper[4824]: I1006 10:37:26.725291 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f4b6ee8e-9e36-422c-95ab-35d166c03844-utilities" (OuterVolumeSpecName: "utilities") pod "f4b6ee8e-9e36-422c-95ab-35d166c03844" (UID: "f4b6ee8e-9e36-422c-95ab-35d166c03844"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:37:26 crc kubenswrapper[4824]: I1006 10:37:26.733149 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f4b6ee8e-9e36-422c-95ab-35d166c03844-kube-api-access-9gftx" (OuterVolumeSpecName: "kube-api-access-9gftx") pod "f4b6ee8e-9e36-422c-95ab-35d166c03844" (UID: "f4b6ee8e-9e36-422c-95ab-35d166c03844"). InnerVolumeSpecName "kube-api-access-9gftx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:37:26 crc kubenswrapper[4824]: I1006 10:37:26.825947 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f4b6ee8e-9e36-422c-95ab-35d166c03844-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f4b6ee8e-9e36-422c-95ab-35d166c03844" (UID: "f4b6ee8e-9e36-422c-95ab-35d166c03844"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:37:26 crc kubenswrapper[4824]: I1006 10:37:26.826902 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f4b6ee8e-9e36-422c-95ab-35d166c03844-utilities\") on node \"crc\" DevicePath \"\"" Oct 06 10:37:26 crc kubenswrapper[4824]: I1006 10:37:26.826925 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f4b6ee8e-9e36-422c-95ab-35d166c03844-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 06 10:37:26 crc kubenswrapper[4824]: I1006 10:37:26.826938 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9gftx\" (UniqueName: \"kubernetes.io/projected/f4b6ee8e-9e36-422c-95ab-35d166c03844-kube-api-access-9gftx\") on node \"crc\" DevicePath \"\"" Oct 06 10:37:27 crc kubenswrapper[4824]: I1006 10:37:27.141646 4824 generic.go:334] "Generic (PLEG): container finished" podID="f4b6ee8e-9e36-422c-95ab-35d166c03844" containerID="3a703a7f87ec11dc9ef4d8308bbc544c68c9ee431448f97b626d43b69b7690a3" exitCode=0 Oct 06 10:37:27 crc kubenswrapper[4824]: I1006 10:37:27.141720 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bbq5l" Oct 06 10:37:27 crc kubenswrapper[4824]: I1006 10:37:27.141741 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bbq5l" event={"ID":"f4b6ee8e-9e36-422c-95ab-35d166c03844","Type":"ContainerDied","Data":"3a703a7f87ec11dc9ef4d8308bbc544c68c9ee431448f97b626d43b69b7690a3"} Oct 06 10:37:27 crc kubenswrapper[4824]: I1006 10:37:27.142940 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bbq5l" event={"ID":"f4b6ee8e-9e36-422c-95ab-35d166c03844","Type":"ContainerDied","Data":"0529842ead62cfb39b0031bec2fc3a979b949d9e763d0e519a3ead978ebc9545"} Oct 06 10:37:27 crc kubenswrapper[4824]: I1006 10:37:27.143014 4824 scope.go:117] "RemoveContainer" containerID="3a703a7f87ec11dc9ef4d8308bbc544c68c9ee431448f97b626d43b69b7690a3" Oct 06 10:37:27 crc kubenswrapper[4824]: I1006 10:37:27.189654 4824 scope.go:117] "RemoveContainer" containerID="3a361dd7fc00103945f09b2ffc4864ba8f1aaec25e046e353e2f93b4d9c2cf35" Oct 06 10:37:27 crc kubenswrapper[4824]: I1006 10:37:27.196170 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-bbq5l"] Oct 06 10:37:27 crc kubenswrapper[4824]: I1006 10:37:27.208650 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-bbq5l"] Oct 06 10:37:27 crc kubenswrapper[4824]: I1006 10:37:27.222733 4824 scope.go:117] "RemoveContainer" containerID="dc856aae7130b2fa93cccc4e3cf29258e2be7a14af3f29a38d088398b6dbfdb6" Oct 06 10:37:27 crc kubenswrapper[4824]: I1006 10:37:27.276265 4824 scope.go:117] "RemoveContainer" containerID="3a703a7f87ec11dc9ef4d8308bbc544c68c9ee431448f97b626d43b69b7690a3" Oct 06 10:37:27 crc kubenswrapper[4824]: E1006 10:37:27.276957 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3a703a7f87ec11dc9ef4d8308bbc544c68c9ee431448f97b626d43b69b7690a3\": container with ID starting with 3a703a7f87ec11dc9ef4d8308bbc544c68c9ee431448f97b626d43b69b7690a3 not found: ID does not exist" containerID="3a703a7f87ec11dc9ef4d8308bbc544c68c9ee431448f97b626d43b69b7690a3" Oct 06 10:37:27 crc kubenswrapper[4824]: I1006 10:37:27.277024 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3a703a7f87ec11dc9ef4d8308bbc544c68c9ee431448f97b626d43b69b7690a3"} err="failed to get container status \"3a703a7f87ec11dc9ef4d8308bbc544c68c9ee431448f97b626d43b69b7690a3\": rpc error: code = NotFound desc = could not find container \"3a703a7f87ec11dc9ef4d8308bbc544c68c9ee431448f97b626d43b69b7690a3\": container with ID starting with 3a703a7f87ec11dc9ef4d8308bbc544c68c9ee431448f97b626d43b69b7690a3 not found: ID does not exist" Oct 06 10:37:27 crc kubenswrapper[4824]: I1006 10:37:27.277046 4824 scope.go:117] "RemoveContainer" containerID="3a361dd7fc00103945f09b2ffc4864ba8f1aaec25e046e353e2f93b4d9c2cf35" Oct 06 10:37:27 crc kubenswrapper[4824]: E1006 10:37:27.285441 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3a361dd7fc00103945f09b2ffc4864ba8f1aaec25e046e353e2f93b4d9c2cf35\": container with ID starting with 3a361dd7fc00103945f09b2ffc4864ba8f1aaec25e046e353e2f93b4d9c2cf35 not found: ID does not exist" containerID="3a361dd7fc00103945f09b2ffc4864ba8f1aaec25e046e353e2f93b4d9c2cf35" Oct 06 10:37:27 crc kubenswrapper[4824]: I1006 10:37:27.285571 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3a361dd7fc00103945f09b2ffc4864ba8f1aaec25e046e353e2f93b4d9c2cf35"} err="failed to get container status \"3a361dd7fc00103945f09b2ffc4864ba8f1aaec25e046e353e2f93b4d9c2cf35\": rpc error: code = NotFound desc = could not find container \"3a361dd7fc00103945f09b2ffc4864ba8f1aaec25e046e353e2f93b4d9c2cf35\": container with ID starting with 3a361dd7fc00103945f09b2ffc4864ba8f1aaec25e046e353e2f93b4d9c2cf35 not found: ID does not exist" Oct 06 10:37:27 crc kubenswrapper[4824]: I1006 10:37:27.285620 4824 scope.go:117] "RemoveContainer" containerID="dc856aae7130b2fa93cccc4e3cf29258e2be7a14af3f29a38d088398b6dbfdb6" Oct 06 10:37:27 crc kubenswrapper[4824]: E1006 10:37:27.287238 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dc856aae7130b2fa93cccc4e3cf29258e2be7a14af3f29a38d088398b6dbfdb6\": container with ID starting with dc856aae7130b2fa93cccc4e3cf29258e2be7a14af3f29a38d088398b6dbfdb6 not found: ID does not exist" containerID="dc856aae7130b2fa93cccc4e3cf29258e2be7a14af3f29a38d088398b6dbfdb6" Oct 06 10:37:27 crc kubenswrapper[4824]: I1006 10:37:27.287348 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dc856aae7130b2fa93cccc4e3cf29258e2be7a14af3f29a38d088398b6dbfdb6"} err="failed to get container status \"dc856aae7130b2fa93cccc4e3cf29258e2be7a14af3f29a38d088398b6dbfdb6\": rpc error: code = NotFound desc = could not find container \"dc856aae7130b2fa93cccc4e3cf29258e2be7a14af3f29a38d088398b6dbfdb6\": container with ID starting with dc856aae7130b2fa93cccc4e3cf29258e2be7a14af3f29a38d088398b6dbfdb6 not found: ID does not exist" Oct 06 10:37:27 crc kubenswrapper[4824]: I1006 10:37:27.310584 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b6ee8e-9e36-422c-95ab-35d166c03844" path="/var/lib/kubelet/pods/f4b6ee8e-9e36-422c-95ab-35d166c03844/volumes" Oct 06 10:37:43 crc kubenswrapper[4824]: I1006 10:37:43.915622 4824 patch_prober.go:28] interesting pod/machine-config-daemon-khgzw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 10:37:43 crc kubenswrapper[4824]: I1006 10:37:43.916498 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 10:38:02 crc kubenswrapper[4824]: I1006 10:38:02.545167 4824 generic.go:334] "Generic (PLEG): container finished" podID="4555aa06-f941-492e-9a1c-89a7c1a5c50f" containerID="d5e9b3fdf42b3aae98b95483972d106d334f9889ec76c5f4df490a19e1aa3de8" exitCode=0 Oct 06 10:38:02 crc kubenswrapper[4824]: I1006 10:38:02.545274 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8s745" event={"ID":"4555aa06-f941-492e-9a1c-89a7c1a5c50f","Type":"ContainerDied","Data":"d5e9b3fdf42b3aae98b95483972d106d334f9889ec76c5f4df490a19e1aa3de8"} Oct 06 10:38:04 crc kubenswrapper[4824]: I1006 10:38:04.114918 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8s745" Oct 06 10:38:04 crc kubenswrapper[4824]: I1006 10:38:04.248496 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ctqj\" (UniqueName: \"kubernetes.io/projected/4555aa06-f941-492e-9a1c-89a7c1a5c50f-kube-api-access-6ctqj\") pod \"4555aa06-f941-492e-9a1c-89a7c1a5c50f\" (UID: \"4555aa06-f941-492e-9a1c-89a7c1a5c50f\") " Oct 06 10:38:04 crc kubenswrapper[4824]: I1006 10:38:04.248612 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4555aa06-f941-492e-9a1c-89a7c1a5c50f-ssh-key\") pod \"4555aa06-f941-492e-9a1c-89a7c1a5c50f\" (UID: \"4555aa06-f941-492e-9a1c-89a7c1a5c50f\") " Oct 06 10:38:04 crc kubenswrapper[4824]: I1006 10:38:04.248704 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4555aa06-f941-492e-9a1c-89a7c1a5c50f-libvirt-combined-ca-bundle\") pod \"4555aa06-f941-492e-9a1c-89a7c1a5c50f\" (UID: \"4555aa06-f941-492e-9a1c-89a7c1a5c50f\") " Oct 06 10:38:04 crc kubenswrapper[4824]: I1006 10:38:04.248788 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4555aa06-f941-492e-9a1c-89a7c1a5c50f-inventory\") pod \"4555aa06-f941-492e-9a1c-89a7c1a5c50f\" (UID: \"4555aa06-f941-492e-9a1c-89a7c1a5c50f\") " Oct 06 10:38:04 crc kubenswrapper[4824]: I1006 10:38:04.248857 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/4555aa06-f941-492e-9a1c-89a7c1a5c50f-libvirt-secret-0\") pod \"4555aa06-f941-492e-9a1c-89a7c1a5c50f\" (UID: \"4555aa06-f941-492e-9a1c-89a7c1a5c50f\") " Oct 06 10:38:04 crc kubenswrapper[4824]: I1006 10:38:04.257349 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4555aa06-f941-492e-9a1c-89a7c1a5c50f-kube-api-access-6ctqj" (OuterVolumeSpecName: "kube-api-access-6ctqj") pod "4555aa06-f941-492e-9a1c-89a7c1a5c50f" (UID: "4555aa06-f941-492e-9a1c-89a7c1a5c50f"). InnerVolumeSpecName "kube-api-access-6ctqj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:38:04 crc kubenswrapper[4824]: I1006 10:38:04.257473 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4555aa06-f941-492e-9a1c-89a7c1a5c50f-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "4555aa06-f941-492e-9a1c-89a7c1a5c50f" (UID: "4555aa06-f941-492e-9a1c-89a7c1a5c50f"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:38:04 crc kubenswrapper[4824]: I1006 10:38:04.282079 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4555aa06-f941-492e-9a1c-89a7c1a5c50f-inventory" (OuterVolumeSpecName: "inventory") pod "4555aa06-f941-492e-9a1c-89a7c1a5c50f" (UID: "4555aa06-f941-492e-9a1c-89a7c1a5c50f"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:38:04 crc kubenswrapper[4824]: I1006 10:38:04.282189 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4555aa06-f941-492e-9a1c-89a7c1a5c50f-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "4555aa06-f941-492e-9a1c-89a7c1a5c50f" (UID: "4555aa06-f941-492e-9a1c-89a7c1a5c50f"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:38:04 crc kubenswrapper[4824]: I1006 10:38:04.297181 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4555aa06-f941-492e-9a1c-89a7c1a5c50f-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "4555aa06-f941-492e-9a1c-89a7c1a5c50f" (UID: "4555aa06-f941-492e-9a1c-89a7c1a5c50f"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:38:04 crc kubenswrapper[4824]: I1006 10:38:04.352361 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ctqj\" (UniqueName: \"kubernetes.io/projected/4555aa06-f941-492e-9a1c-89a7c1a5c50f-kube-api-access-6ctqj\") on node \"crc\" DevicePath \"\"" Oct 06 10:38:04 crc kubenswrapper[4824]: I1006 10:38:04.352393 4824 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4555aa06-f941-492e-9a1c-89a7c1a5c50f-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 06 10:38:04 crc kubenswrapper[4824]: I1006 10:38:04.352407 4824 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4555aa06-f941-492e-9a1c-89a7c1a5c50f-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 10:38:04 crc kubenswrapper[4824]: I1006 10:38:04.352419 4824 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4555aa06-f941-492e-9a1c-89a7c1a5c50f-inventory\") on node \"crc\" DevicePath \"\"" Oct 06 10:38:04 crc kubenswrapper[4824]: I1006 10:38:04.352431 4824 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/4555aa06-f941-492e-9a1c-89a7c1a5c50f-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Oct 06 10:38:04 crc kubenswrapper[4824]: I1006 10:38:04.574442 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8s745" event={"ID":"4555aa06-f941-492e-9a1c-89a7c1a5c50f","Type":"ContainerDied","Data":"ab77fec3da2c3f4cf1540862ce0c4867b0ca77ca34867306bec12a1236ba7d86"} Oct 06 10:38:04 crc kubenswrapper[4824]: I1006 10:38:04.574482 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ab77fec3da2c3f4cf1540862ce0c4867b0ca77ca34867306bec12a1236ba7d86" Oct 06 10:38:04 crc kubenswrapper[4824]: I1006 10:38:04.574561 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8s745" Oct 06 10:38:04 crc kubenswrapper[4824]: I1006 10:38:04.671870 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-59jnp"] Oct 06 10:38:04 crc kubenswrapper[4824]: E1006 10:38:04.672635 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b6ee8e-9e36-422c-95ab-35d166c03844" containerName="extract-utilities" Oct 06 10:38:04 crc kubenswrapper[4824]: I1006 10:38:04.672652 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b6ee8e-9e36-422c-95ab-35d166c03844" containerName="extract-utilities" Oct 06 10:38:04 crc kubenswrapper[4824]: E1006 10:38:04.672660 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4555aa06-f941-492e-9a1c-89a7c1a5c50f" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Oct 06 10:38:04 crc kubenswrapper[4824]: I1006 10:38:04.672668 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="4555aa06-f941-492e-9a1c-89a7c1a5c50f" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Oct 06 10:38:04 crc kubenswrapper[4824]: E1006 10:38:04.672696 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b6ee8e-9e36-422c-95ab-35d166c03844" containerName="extract-content" Oct 06 10:38:04 crc kubenswrapper[4824]: I1006 10:38:04.672703 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b6ee8e-9e36-422c-95ab-35d166c03844" containerName="extract-content" Oct 06 10:38:04 crc kubenswrapper[4824]: E1006 10:38:04.672713 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b6ee8e-9e36-422c-95ab-35d166c03844" containerName="registry-server" Oct 06 10:38:04 crc kubenswrapper[4824]: I1006 10:38:04.672719 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b6ee8e-9e36-422c-95ab-35d166c03844" containerName="registry-server" Oct 06 10:38:04 crc kubenswrapper[4824]: I1006 10:38:04.672881 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="4555aa06-f941-492e-9a1c-89a7c1a5c50f" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Oct 06 10:38:04 crc kubenswrapper[4824]: I1006 10:38:04.672940 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b6ee8e-9e36-422c-95ab-35d166c03844" containerName="registry-server" Oct 06 10:38:04 crc kubenswrapper[4824]: I1006 10:38:04.673566 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-59jnp" Oct 06 10:38:04 crc kubenswrapper[4824]: I1006 10:38:04.677786 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 06 10:38:04 crc kubenswrapper[4824]: I1006 10:38:04.678095 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Oct 06 10:38:04 crc kubenswrapper[4824]: I1006 10:38:04.678747 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-f886w" Oct 06 10:38:04 crc kubenswrapper[4824]: I1006 10:38:04.679233 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Oct 06 10:38:04 crc kubenswrapper[4824]: I1006 10:38:04.680027 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 06 10:38:04 crc kubenswrapper[4824]: I1006 10:38:04.680250 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Oct 06 10:38:04 crc kubenswrapper[4824]: I1006 10:38:04.684173 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 06 10:38:04 crc kubenswrapper[4824]: I1006 10:38:04.697773 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-59jnp"] Oct 06 10:38:04 crc kubenswrapper[4824]: I1006 10:38:04.759508 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pp844\" (UniqueName: \"kubernetes.io/projected/910d1ffe-8746-4412-bba0-74bf7ff0e1ea-kube-api-access-pp844\") pod \"nova-edpm-deployment-openstack-edpm-ipam-59jnp\" (UID: \"910d1ffe-8746-4412-bba0-74bf7ff0e1ea\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-59jnp" Oct 06 10:38:04 crc kubenswrapper[4824]: I1006 10:38:04.759576 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/910d1ffe-8746-4412-bba0-74bf7ff0e1ea-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-59jnp\" (UID: \"910d1ffe-8746-4412-bba0-74bf7ff0e1ea\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-59jnp" Oct 06 10:38:04 crc kubenswrapper[4824]: I1006 10:38:04.759650 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/910d1ffe-8746-4412-bba0-74bf7ff0e1ea-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-59jnp\" (UID: \"910d1ffe-8746-4412-bba0-74bf7ff0e1ea\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-59jnp" Oct 06 10:38:04 crc kubenswrapper[4824]: I1006 10:38:04.759673 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/910d1ffe-8746-4412-bba0-74bf7ff0e1ea-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-59jnp\" (UID: \"910d1ffe-8746-4412-bba0-74bf7ff0e1ea\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-59jnp" Oct 06 10:38:04 crc kubenswrapper[4824]: I1006 10:38:04.759698 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/910d1ffe-8746-4412-bba0-74bf7ff0e1ea-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-59jnp\" (UID: \"910d1ffe-8746-4412-bba0-74bf7ff0e1ea\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-59jnp" Oct 06 10:38:04 crc kubenswrapper[4824]: I1006 10:38:04.759743 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/910d1ffe-8746-4412-bba0-74bf7ff0e1ea-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-59jnp\" (UID: \"910d1ffe-8746-4412-bba0-74bf7ff0e1ea\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-59jnp" Oct 06 10:38:04 crc kubenswrapper[4824]: I1006 10:38:04.759777 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/910d1ffe-8746-4412-bba0-74bf7ff0e1ea-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-59jnp\" (UID: \"910d1ffe-8746-4412-bba0-74bf7ff0e1ea\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-59jnp" Oct 06 10:38:04 crc kubenswrapper[4824]: I1006 10:38:04.759971 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/910d1ffe-8746-4412-bba0-74bf7ff0e1ea-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-59jnp\" (UID: \"910d1ffe-8746-4412-bba0-74bf7ff0e1ea\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-59jnp" Oct 06 10:38:04 crc kubenswrapper[4824]: I1006 10:38:04.760147 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/910d1ffe-8746-4412-bba0-74bf7ff0e1ea-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-59jnp\" (UID: \"910d1ffe-8746-4412-bba0-74bf7ff0e1ea\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-59jnp" Oct 06 10:38:04 crc kubenswrapper[4824]: I1006 10:38:04.862320 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pp844\" (UniqueName: \"kubernetes.io/projected/910d1ffe-8746-4412-bba0-74bf7ff0e1ea-kube-api-access-pp844\") pod \"nova-edpm-deployment-openstack-edpm-ipam-59jnp\" (UID: \"910d1ffe-8746-4412-bba0-74bf7ff0e1ea\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-59jnp" Oct 06 10:38:04 crc kubenswrapper[4824]: I1006 10:38:04.862392 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/910d1ffe-8746-4412-bba0-74bf7ff0e1ea-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-59jnp\" (UID: \"910d1ffe-8746-4412-bba0-74bf7ff0e1ea\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-59jnp" Oct 06 10:38:04 crc kubenswrapper[4824]: I1006 10:38:04.862445 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/910d1ffe-8746-4412-bba0-74bf7ff0e1ea-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-59jnp\" (UID: \"910d1ffe-8746-4412-bba0-74bf7ff0e1ea\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-59jnp" Oct 06 10:38:04 crc kubenswrapper[4824]: I1006 10:38:04.862469 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/910d1ffe-8746-4412-bba0-74bf7ff0e1ea-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-59jnp\" (UID: \"910d1ffe-8746-4412-bba0-74bf7ff0e1ea\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-59jnp" Oct 06 10:38:04 crc kubenswrapper[4824]: I1006 10:38:04.862493 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/910d1ffe-8746-4412-bba0-74bf7ff0e1ea-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-59jnp\" (UID: \"910d1ffe-8746-4412-bba0-74bf7ff0e1ea\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-59jnp" Oct 06 10:38:04 crc kubenswrapper[4824]: I1006 10:38:04.862538 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/910d1ffe-8746-4412-bba0-74bf7ff0e1ea-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-59jnp\" (UID: \"910d1ffe-8746-4412-bba0-74bf7ff0e1ea\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-59jnp" Oct 06 10:38:04 crc kubenswrapper[4824]: I1006 10:38:04.862573 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/910d1ffe-8746-4412-bba0-74bf7ff0e1ea-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-59jnp\" (UID: \"910d1ffe-8746-4412-bba0-74bf7ff0e1ea\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-59jnp" Oct 06 10:38:04 crc kubenswrapper[4824]: I1006 10:38:04.862599 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/910d1ffe-8746-4412-bba0-74bf7ff0e1ea-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-59jnp\" (UID: \"910d1ffe-8746-4412-bba0-74bf7ff0e1ea\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-59jnp" Oct 06 10:38:04 crc kubenswrapper[4824]: I1006 10:38:04.862627 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/910d1ffe-8746-4412-bba0-74bf7ff0e1ea-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-59jnp\" (UID: \"910d1ffe-8746-4412-bba0-74bf7ff0e1ea\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-59jnp" Oct 06 10:38:04 crc kubenswrapper[4824]: I1006 10:38:04.864533 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/910d1ffe-8746-4412-bba0-74bf7ff0e1ea-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-59jnp\" (UID: \"910d1ffe-8746-4412-bba0-74bf7ff0e1ea\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-59jnp" Oct 06 10:38:04 crc kubenswrapper[4824]: I1006 10:38:04.866571 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/910d1ffe-8746-4412-bba0-74bf7ff0e1ea-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-59jnp\" (UID: \"910d1ffe-8746-4412-bba0-74bf7ff0e1ea\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-59jnp" Oct 06 10:38:04 crc kubenswrapper[4824]: I1006 10:38:04.867102 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/910d1ffe-8746-4412-bba0-74bf7ff0e1ea-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-59jnp\" (UID: \"910d1ffe-8746-4412-bba0-74bf7ff0e1ea\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-59jnp" Oct 06 10:38:04 crc kubenswrapper[4824]: I1006 10:38:04.868076 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/910d1ffe-8746-4412-bba0-74bf7ff0e1ea-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-59jnp\" (UID: \"910d1ffe-8746-4412-bba0-74bf7ff0e1ea\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-59jnp" Oct 06 10:38:04 crc kubenswrapper[4824]: I1006 10:38:04.868405 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/910d1ffe-8746-4412-bba0-74bf7ff0e1ea-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-59jnp\" (UID: \"910d1ffe-8746-4412-bba0-74bf7ff0e1ea\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-59jnp" Oct 06 10:38:04 crc kubenswrapper[4824]: I1006 10:38:04.868576 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/910d1ffe-8746-4412-bba0-74bf7ff0e1ea-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-59jnp\" (UID: \"910d1ffe-8746-4412-bba0-74bf7ff0e1ea\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-59jnp" Oct 06 10:38:04 crc kubenswrapper[4824]: I1006 10:38:04.869034 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/910d1ffe-8746-4412-bba0-74bf7ff0e1ea-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-59jnp\" (UID: \"910d1ffe-8746-4412-bba0-74bf7ff0e1ea\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-59jnp" Oct 06 10:38:04 crc kubenswrapper[4824]: I1006 10:38:04.871256 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/910d1ffe-8746-4412-bba0-74bf7ff0e1ea-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-59jnp\" (UID: \"910d1ffe-8746-4412-bba0-74bf7ff0e1ea\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-59jnp" Oct 06 10:38:04 crc kubenswrapper[4824]: I1006 10:38:04.879673 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pp844\" (UniqueName: \"kubernetes.io/projected/910d1ffe-8746-4412-bba0-74bf7ff0e1ea-kube-api-access-pp844\") pod \"nova-edpm-deployment-openstack-edpm-ipam-59jnp\" (UID: \"910d1ffe-8746-4412-bba0-74bf7ff0e1ea\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-59jnp" Oct 06 10:38:04 crc kubenswrapper[4824]: I1006 10:38:04.993432 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-59jnp" Oct 06 10:38:05 crc kubenswrapper[4824]: I1006 10:38:05.647410 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-59jnp"] Oct 06 10:38:06 crc kubenswrapper[4824]: I1006 10:38:06.611482 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-59jnp" event={"ID":"910d1ffe-8746-4412-bba0-74bf7ff0e1ea","Type":"ContainerStarted","Data":"e4388ad67a18c5f1e3481df8acbc412ced18e612dd0fcd4a9540cc01df747935"} Oct 06 10:38:06 crc kubenswrapper[4824]: I1006 10:38:06.612590 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-59jnp" event={"ID":"910d1ffe-8746-4412-bba0-74bf7ff0e1ea","Type":"ContainerStarted","Data":"63301ce24b43dbc109213d81460fd9b1f8b79b9aabe9b370ec2dcb5e8092fd76"} Oct 06 10:38:06 crc kubenswrapper[4824]: I1006 10:38:06.648708 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-59jnp" podStartSLOduration=2.195377019 podStartE2EDuration="2.648677732s" podCreationTimestamp="2025-10-06 10:38:04 +0000 UTC" firstStartedPulling="2025-10-06 10:38:05.654158605 +0000 UTC m=+2395.018581476" lastFinishedPulling="2025-10-06 10:38:06.107459288 +0000 UTC m=+2395.471882189" observedRunningTime="2025-10-06 10:38:06.645367783 +0000 UTC m=+2396.009790664" watchObservedRunningTime="2025-10-06 10:38:06.648677732 +0000 UTC m=+2396.013100593" Oct 06 10:38:13 crc kubenswrapper[4824]: I1006 10:38:13.915708 4824 patch_prober.go:28] interesting pod/machine-config-daemon-khgzw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 10:38:13 crc kubenswrapper[4824]: I1006 10:38:13.916496 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 10:38:13 crc kubenswrapper[4824]: I1006 10:38:13.916564 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" Oct 06 10:38:13 crc kubenswrapper[4824]: I1006 10:38:13.917604 4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"defbb9e5d617c326a5dda4c68f7311eec28a02ec245ca3527fbdf00979129a82"} pod="openshift-machine-config-operator/machine-config-daemon-khgzw" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 06 10:38:13 crc kubenswrapper[4824]: I1006 10:38:13.917677 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" containerName="machine-config-daemon" containerID="cri-o://defbb9e5d617c326a5dda4c68f7311eec28a02ec245ca3527fbdf00979129a82" gracePeriod=600 Oct 06 10:38:14 crc kubenswrapper[4824]: E1006 10:38:14.038025 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-khgzw_openshift-machine-config-operator(e1140c8c-93f1-4fce-8c06-a69561ff8a6d)\"" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" Oct 06 10:38:14 crc kubenswrapper[4824]: I1006 10:38:14.714089 4824 generic.go:334] "Generic (PLEG): container finished" podID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" containerID="defbb9e5d617c326a5dda4c68f7311eec28a02ec245ca3527fbdf00979129a82" exitCode=0 Oct 06 10:38:14 crc kubenswrapper[4824]: I1006 10:38:14.714149 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" event={"ID":"e1140c8c-93f1-4fce-8c06-a69561ff8a6d","Type":"ContainerDied","Data":"defbb9e5d617c326a5dda4c68f7311eec28a02ec245ca3527fbdf00979129a82"} Oct 06 10:38:14 crc kubenswrapper[4824]: I1006 10:38:14.714610 4824 scope.go:117] "RemoveContainer" containerID="2e9a092782b8371eeb30f1c342e8665ca42002786b61ad0319bff781e0406a48" Oct 06 10:38:14 crc kubenswrapper[4824]: I1006 10:38:14.715368 4824 scope.go:117] "RemoveContainer" containerID="defbb9e5d617c326a5dda4c68f7311eec28a02ec245ca3527fbdf00979129a82" Oct 06 10:38:14 crc kubenswrapper[4824]: E1006 10:38:14.715941 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-khgzw_openshift-machine-config-operator(e1140c8c-93f1-4fce-8c06-a69561ff8a6d)\"" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" Oct 06 10:38:29 crc kubenswrapper[4824]: I1006 10:38:29.183646 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-grp5h"] Oct 06 10:38:29 crc kubenswrapper[4824]: I1006 10:38:29.190237 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-grp5h" Oct 06 10:38:29 crc kubenswrapper[4824]: I1006 10:38:29.204299 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-grp5h"] Oct 06 10:38:29 crc kubenswrapper[4824]: I1006 10:38:29.247376 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b66b0f49-bb9a-49b8-9a5a-2aca8d5b620e-utilities\") pod \"certified-operators-grp5h\" (UID: \"b66b0f49-bb9a-49b8-9a5a-2aca8d5b620e\") " pod="openshift-marketplace/certified-operators-grp5h" Oct 06 10:38:29 crc kubenswrapper[4824]: I1006 10:38:29.247515 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q5gfv\" (UniqueName: \"kubernetes.io/projected/b66b0f49-bb9a-49b8-9a5a-2aca8d5b620e-kube-api-access-q5gfv\") pod \"certified-operators-grp5h\" (UID: \"b66b0f49-bb9a-49b8-9a5a-2aca8d5b620e\") " pod="openshift-marketplace/certified-operators-grp5h" Oct 06 10:38:29 crc kubenswrapper[4824]: I1006 10:38:29.247545 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b66b0f49-bb9a-49b8-9a5a-2aca8d5b620e-catalog-content\") pod \"certified-operators-grp5h\" (UID: \"b66b0f49-bb9a-49b8-9a5a-2aca8d5b620e\") " pod="openshift-marketplace/certified-operators-grp5h" Oct 06 10:38:29 crc kubenswrapper[4824]: I1006 10:38:29.275442 4824 scope.go:117] "RemoveContainer" containerID="defbb9e5d617c326a5dda4c68f7311eec28a02ec245ca3527fbdf00979129a82" Oct 06 10:38:29 crc kubenswrapper[4824]: E1006 10:38:29.276362 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-khgzw_openshift-machine-config-operator(e1140c8c-93f1-4fce-8c06-a69561ff8a6d)\"" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" Oct 06 10:38:29 crc kubenswrapper[4824]: I1006 10:38:29.348948 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q5gfv\" (UniqueName: \"kubernetes.io/projected/b66b0f49-bb9a-49b8-9a5a-2aca8d5b620e-kube-api-access-q5gfv\") pod \"certified-operators-grp5h\" (UID: \"b66b0f49-bb9a-49b8-9a5a-2aca8d5b620e\") " pod="openshift-marketplace/certified-operators-grp5h" Oct 06 10:38:29 crc kubenswrapper[4824]: I1006 10:38:29.349034 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b66b0f49-bb9a-49b8-9a5a-2aca8d5b620e-catalog-content\") pod \"certified-operators-grp5h\" (UID: \"b66b0f49-bb9a-49b8-9a5a-2aca8d5b620e\") " pod="openshift-marketplace/certified-operators-grp5h" Oct 06 10:38:29 crc kubenswrapper[4824]: I1006 10:38:29.349112 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b66b0f49-bb9a-49b8-9a5a-2aca8d5b620e-utilities\") pod \"certified-operators-grp5h\" (UID: \"b66b0f49-bb9a-49b8-9a5a-2aca8d5b620e\") " pod="openshift-marketplace/certified-operators-grp5h" Oct 06 10:38:29 crc kubenswrapper[4824]: I1006 10:38:29.349618 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b66b0f49-bb9a-49b8-9a5a-2aca8d5b620e-utilities\") pod \"certified-operators-grp5h\" (UID: \"b66b0f49-bb9a-49b8-9a5a-2aca8d5b620e\") " pod="openshift-marketplace/certified-operators-grp5h" Oct 06 10:38:29 crc kubenswrapper[4824]: I1006 10:38:29.349834 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b66b0f49-bb9a-49b8-9a5a-2aca8d5b620e-catalog-content\") pod \"certified-operators-grp5h\" (UID: \"b66b0f49-bb9a-49b8-9a5a-2aca8d5b620e\") " pod="openshift-marketplace/certified-operators-grp5h" Oct 06 10:38:29 crc kubenswrapper[4824]: I1006 10:38:29.375909 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q5gfv\" (UniqueName: \"kubernetes.io/projected/b66b0f49-bb9a-49b8-9a5a-2aca8d5b620e-kube-api-access-q5gfv\") pod \"certified-operators-grp5h\" (UID: \"b66b0f49-bb9a-49b8-9a5a-2aca8d5b620e\") " pod="openshift-marketplace/certified-operators-grp5h" Oct 06 10:38:29 crc kubenswrapper[4824]: I1006 10:38:29.515516 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-grp5h" Oct 06 10:38:30 crc kubenswrapper[4824]: I1006 10:38:30.061484 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-grp5h"] Oct 06 10:38:30 crc kubenswrapper[4824]: I1006 10:38:30.918281 4824 generic.go:334] "Generic (PLEG): container finished" podID="b66b0f49-bb9a-49b8-9a5a-2aca8d5b620e" containerID="f9a25d0249f8e6be21f66182d479c02914adf09a341b637ed0d290d963ac4e45" exitCode=0 Oct 06 10:38:30 crc kubenswrapper[4824]: I1006 10:38:30.918401 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-grp5h" event={"ID":"b66b0f49-bb9a-49b8-9a5a-2aca8d5b620e","Type":"ContainerDied","Data":"f9a25d0249f8e6be21f66182d479c02914adf09a341b637ed0d290d963ac4e45"} Oct 06 10:38:30 crc kubenswrapper[4824]: I1006 10:38:30.918704 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-grp5h" event={"ID":"b66b0f49-bb9a-49b8-9a5a-2aca8d5b620e","Type":"ContainerStarted","Data":"9f1aecba8c9b0c189a401e3b2588eede6764770d310a6c4c41ce8651a20d8d97"} Oct 06 10:38:31 crc kubenswrapper[4824]: I1006 10:38:31.935729 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-grp5h" event={"ID":"b66b0f49-bb9a-49b8-9a5a-2aca8d5b620e","Type":"ContainerStarted","Data":"753d23f3b3ebeea35c4477b4b54e63c5de117a619ff4f0102dfb1419c5bb7906"} Oct 06 10:38:32 crc kubenswrapper[4824]: I1006 10:38:32.949577 4824 generic.go:334] "Generic (PLEG): container finished" podID="b66b0f49-bb9a-49b8-9a5a-2aca8d5b620e" containerID="753d23f3b3ebeea35c4477b4b54e63c5de117a619ff4f0102dfb1419c5bb7906" exitCode=0 Oct 06 10:38:32 crc kubenswrapper[4824]: I1006 10:38:32.949688 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-grp5h" event={"ID":"b66b0f49-bb9a-49b8-9a5a-2aca8d5b620e","Type":"ContainerDied","Data":"753d23f3b3ebeea35c4477b4b54e63c5de117a619ff4f0102dfb1419c5bb7906"} Oct 06 10:38:33 crc kubenswrapper[4824]: I1006 10:38:33.968326 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-grp5h" event={"ID":"b66b0f49-bb9a-49b8-9a5a-2aca8d5b620e","Type":"ContainerStarted","Data":"7de21155c716356f5c5d4f5a73ad154566803420cb9980a0cc1706e52f07e482"} Oct 06 10:38:33 crc kubenswrapper[4824]: I1006 10:38:33.991280 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-grp5h" podStartSLOduration=2.491393943 podStartE2EDuration="4.991245248s" podCreationTimestamp="2025-10-06 10:38:29 +0000 UTC" firstStartedPulling="2025-10-06 10:38:30.922318868 +0000 UTC m=+2420.286741769" lastFinishedPulling="2025-10-06 10:38:33.422170213 +0000 UTC m=+2422.786593074" observedRunningTime="2025-10-06 10:38:33.987845005 +0000 UTC m=+2423.352267886" watchObservedRunningTime="2025-10-06 10:38:33.991245248 +0000 UTC m=+2423.355668139" Oct 06 10:38:39 crc kubenswrapper[4824]: I1006 10:38:39.516706 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-grp5h" Oct 06 10:38:39 crc kubenswrapper[4824]: I1006 10:38:39.517500 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-grp5h" Oct 06 10:38:39 crc kubenswrapper[4824]: I1006 10:38:39.583881 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-grp5h" Oct 06 10:38:40 crc kubenswrapper[4824]: I1006 10:38:40.137536 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-grp5h" Oct 06 10:38:40 crc kubenswrapper[4824]: I1006 10:38:40.210814 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-grp5h"] Oct 06 10:38:40 crc kubenswrapper[4824]: I1006 10:38:40.274531 4824 scope.go:117] "RemoveContainer" containerID="defbb9e5d617c326a5dda4c68f7311eec28a02ec245ca3527fbdf00979129a82" Oct 06 10:38:40 crc kubenswrapper[4824]: E1006 10:38:40.274826 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-khgzw_openshift-machine-config-operator(e1140c8c-93f1-4fce-8c06-a69561ff8a6d)\"" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" Oct 06 10:38:42 crc kubenswrapper[4824]: I1006 10:38:42.088620 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-grp5h" podUID="b66b0f49-bb9a-49b8-9a5a-2aca8d5b620e" containerName="registry-server" containerID="cri-o://7de21155c716356f5c5d4f5a73ad154566803420cb9980a0cc1706e52f07e482" gracePeriod=2 Oct 06 10:38:43 crc kubenswrapper[4824]: I1006 10:38:43.103154 4824 generic.go:334] "Generic (PLEG): container finished" podID="b66b0f49-bb9a-49b8-9a5a-2aca8d5b620e" containerID="7de21155c716356f5c5d4f5a73ad154566803420cb9980a0cc1706e52f07e482" exitCode=0 Oct 06 10:38:43 crc kubenswrapper[4824]: I1006 10:38:43.103237 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-grp5h" event={"ID":"b66b0f49-bb9a-49b8-9a5a-2aca8d5b620e","Type":"ContainerDied","Data":"7de21155c716356f5c5d4f5a73ad154566803420cb9980a0cc1706e52f07e482"} Oct 06 10:38:43 crc kubenswrapper[4824]: I1006 10:38:43.103786 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-grp5h" event={"ID":"b66b0f49-bb9a-49b8-9a5a-2aca8d5b620e","Type":"ContainerDied","Data":"9f1aecba8c9b0c189a401e3b2588eede6764770d310a6c4c41ce8651a20d8d97"} Oct 06 10:38:43 crc kubenswrapper[4824]: I1006 10:38:43.103842 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9f1aecba8c9b0c189a401e3b2588eede6764770d310a6c4c41ce8651a20d8d97" Oct 06 10:38:43 crc kubenswrapper[4824]: I1006 10:38:43.113105 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-grp5h" Oct 06 10:38:43 crc kubenswrapper[4824]: I1006 10:38:43.216576 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q5gfv\" (UniqueName: \"kubernetes.io/projected/b66b0f49-bb9a-49b8-9a5a-2aca8d5b620e-kube-api-access-q5gfv\") pod \"b66b0f49-bb9a-49b8-9a5a-2aca8d5b620e\" (UID: \"b66b0f49-bb9a-49b8-9a5a-2aca8d5b620e\") " Oct 06 10:38:43 crc kubenswrapper[4824]: I1006 10:38:43.216792 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b66b0f49-bb9a-49b8-9a5a-2aca8d5b620e-utilities\") pod \"b66b0f49-bb9a-49b8-9a5a-2aca8d5b620e\" (UID: \"b66b0f49-bb9a-49b8-9a5a-2aca8d5b620e\") " Oct 06 10:38:43 crc kubenswrapper[4824]: I1006 10:38:43.217000 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b66b0f49-bb9a-49b8-9a5a-2aca8d5b620e-catalog-content\") pod \"b66b0f49-bb9a-49b8-9a5a-2aca8d5b620e\" (UID: \"b66b0f49-bb9a-49b8-9a5a-2aca8d5b620e\") " Oct 06 10:38:43 crc kubenswrapper[4824]: I1006 10:38:43.218871 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b66b0f49-bb9a-49b8-9a5a-2aca8d5b620e-utilities" (OuterVolumeSpecName: "utilities") pod "b66b0f49-bb9a-49b8-9a5a-2aca8d5b620e" (UID: "b66b0f49-bb9a-49b8-9a5a-2aca8d5b620e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:38:43 crc kubenswrapper[4824]: I1006 10:38:43.227040 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b66b0f49-bb9a-49b8-9a5a-2aca8d5b620e-kube-api-access-q5gfv" (OuterVolumeSpecName: "kube-api-access-q5gfv") pod "b66b0f49-bb9a-49b8-9a5a-2aca8d5b620e" (UID: "b66b0f49-bb9a-49b8-9a5a-2aca8d5b620e"). InnerVolumeSpecName "kube-api-access-q5gfv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:38:43 crc kubenswrapper[4824]: I1006 10:38:43.266309 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b66b0f49-bb9a-49b8-9a5a-2aca8d5b620e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b66b0f49-bb9a-49b8-9a5a-2aca8d5b620e" (UID: "b66b0f49-bb9a-49b8-9a5a-2aca8d5b620e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:38:43 crc kubenswrapper[4824]: I1006 10:38:43.320003 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b66b0f49-bb9a-49b8-9a5a-2aca8d5b620e-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 06 10:38:43 crc kubenswrapper[4824]: I1006 10:38:43.320041 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q5gfv\" (UniqueName: \"kubernetes.io/projected/b66b0f49-bb9a-49b8-9a5a-2aca8d5b620e-kube-api-access-q5gfv\") on node \"crc\" DevicePath \"\"" Oct 06 10:38:43 crc kubenswrapper[4824]: I1006 10:38:43.320054 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b66b0f49-bb9a-49b8-9a5a-2aca8d5b620e-utilities\") on node \"crc\" DevicePath \"\"" Oct 06 10:38:44 crc kubenswrapper[4824]: I1006 10:38:44.115922 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-grp5h" Oct 06 10:38:44 crc kubenswrapper[4824]: I1006 10:38:44.145105 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-grp5h"] Oct 06 10:38:44 crc kubenswrapper[4824]: I1006 10:38:44.155897 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-grp5h"] Oct 06 10:38:45 crc kubenswrapper[4824]: I1006 10:38:45.293902 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b66b0f49-bb9a-49b8-9a5a-2aca8d5b620e" path="/var/lib/kubelet/pods/b66b0f49-bb9a-49b8-9a5a-2aca8d5b620e/volumes" Oct 06 10:38:55 crc kubenswrapper[4824]: I1006 10:38:55.274302 4824 scope.go:117] "RemoveContainer" containerID="defbb9e5d617c326a5dda4c68f7311eec28a02ec245ca3527fbdf00979129a82" Oct 06 10:38:55 crc kubenswrapper[4824]: E1006 10:38:55.275333 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-khgzw_openshift-machine-config-operator(e1140c8c-93f1-4fce-8c06-a69561ff8a6d)\"" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" Oct 06 10:39:09 crc kubenswrapper[4824]: I1006 10:39:09.274290 4824 scope.go:117] "RemoveContainer" containerID="defbb9e5d617c326a5dda4c68f7311eec28a02ec245ca3527fbdf00979129a82" Oct 06 10:39:09 crc kubenswrapper[4824]: E1006 10:39:09.275149 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-khgzw_openshift-machine-config-operator(e1140c8c-93f1-4fce-8c06-a69561ff8a6d)\"" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" Oct 06 10:39:24 crc kubenswrapper[4824]: I1006 10:39:24.274537 4824 scope.go:117] "RemoveContainer" containerID="defbb9e5d617c326a5dda4c68f7311eec28a02ec245ca3527fbdf00979129a82" Oct 06 10:39:24 crc kubenswrapper[4824]: E1006 10:39:24.275798 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-khgzw_openshift-machine-config-operator(e1140c8c-93f1-4fce-8c06-a69561ff8a6d)\"" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" Oct 06 10:39:39 crc kubenswrapper[4824]: I1006 10:39:39.274768 4824 scope.go:117] "RemoveContainer" containerID="defbb9e5d617c326a5dda4c68f7311eec28a02ec245ca3527fbdf00979129a82" Oct 06 10:39:39 crc kubenswrapper[4824]: E1006 10:39:39.275925 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-khgzw_openshift-machine-config-operator(e1140c8c-93f1-4fce-8c06-a69561ff8a6d)\"" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" Oct 06 10:39:50 crc kubenswrapper[4824]: I1006 10:39:50.275259 4824 scope.go:117] "RemoveContainer" containerID="defbb9e5d617c326a5dda4c68f7311eec28a02ec245ca3527fbdf00979129a82" Oct 06 10:39:50 crc kubenswrapper[4824]: E1006 10:39:50.276699 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-khgzw_openshift-machine-config-operator(e1140c8c-93f1-4fce-8c06-a69561ff8a6d)\"" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" Oct 06 10:40:02 crc kubenswrapper[4824]: I1006 10:40:02.274927 4824 scope.go:117] "RemoveContainer" containerID="defbb9e5d617c326a5dda4c68f7311eec28a02ec245ca3527fbdf00979129a82" Oct 06 10:40:02 crc kubenswrapper[4824]: E1006 10:40:02.275993 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-khgzw_openshift-machine-config-operator(e1140c8c-93f1-4fce-8c06-a69561ff8a6d)\"" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" Oct 06 10:40:16 crc kubenswrapper[4824]: I1006 10:40:16.274246 4824 scope.go:117] "RemoveContainer" containerID="defbb9e5d617c326a5dda4c68f7311eec28a02ec245ca3527fbdf00979129a82" Oct 06 10:40:16 crc kubenswrapper[4824]: E1006 10:40:16.275184 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-khgzw_openshift-machine-config-operator(e1140c8c-93f1-4fce-8c06-a69561ff8a6d)\"" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" Oct 06 10:40:27 crc kubenswrapper[4824]: I1006 10:40:27.275103 4824 scope.go:117] "RemoveContainer" containerID="defbb9e5d617c326a5dda4c68f7311eec28a02ec245ca3527fbdf00979129a82" Oct 06 10:40:27 crc kubenswrapper[4824]: E1006 10:40:27.276301 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-khgzw_openshift-machine-config-operator(e1140c8c-93f1-4fce-8c06-a69561ff8a6d)\"" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" Oct 06 10:40:40 crc kubenswrapper[4824]: I1006 10:40:40.274610 4824 scope.go:117] "RemoveContainer" containerID="defbb9e5d617c326a5dda4c68f7311eec28a02ec245ca3527fbdf00979129a82" Oct 06 10:40:40 crc kubenswrapper[4824]: E1006 10:40:40.276847 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-khgzw_openshift-machine-config-operator(e1140c8c-93f1-4fce-8c06-a69561ff8a6d)\"" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" Oct 06 10:40:55 crc kubenswrapper[4824]: I1006 10:40:55.275035 4824 scope.go:117] "RemoveContainer" containerID="defbb9e5d617c326a5dda4c68f7311eec28a02ec245ca3527fbdf00979129a82" Oct 06 10:40:55 crc kubenswrapper[4824]: E1006 10:40:55.275864 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-khgzw_openshift-machine-config-operator(e1140c8c-93f1-4fce-8c06-a69561ff8a6d)\"" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" Oct 06 10:41:08 crc kubenswrapper[4824]: I1006 10:41:08.274236 4824 scope.go:117] "RemoveContainer" containerID="defbb9e5d617c326a5dda4c68f7311eec28a02ec245ca3527fbdf00979129a82" Oct 06 10:41:08 crc kubenswrapper[4824]: E1006 10:41:08.275305 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-khgzw_openshift-machine-config-operator(e1140c8c-93f1-4fce-8c06-a69561ff8a6d)\"" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" Oct 06 10:41:19 crc kubenswrapper[4824]: I1006 10:41:19.275105 4824 scope.go:117] "RemoveContainer" containerID="defbb9e5d617c326a5dda4c68f7311eec28a02ec245ca3527fbdf00979129a82" Oct 06 10:41:19 crc kubenswrapper[4824]: E1006 10:41:19.278127 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-khgzw_openshift-machine-config-operator(e1140c8c-93f1-4fce-8c06-a69561ff8a6d)\"" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" Oct 06 10:41:32 crc kubenswrapper[4824]: I1006 10:41:32.273917 4824 scope.go:117] "RemoveContainer" containerID="defbb9e5d617c326a5dda4c68f7311eec28a02ec245ca3527fbdf00979129a82" Oct 06 10:41:32 crc kubenswrapper[4824]: E1006 10:41:32.274907 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-khgzw_openshift-machine-config-operator(e1140c8c-93f1-4fce-8c06-a69561ff8a6d)\"" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" Oct 06 10:41:42 crc kubenswrapper[4824]: I1006 10:41:42.209419 4824 generic.go:334] "Generic (PLEG): container finished" podID="910d1ffe-8746-4412-bba0-74bf7ff0e1ea" containerID="e4388ad67a18c5f1e3481df8acbc412ced18e612dd0fcd4a9540cc01df747935" exitCode=0 Oct 06 10:41:42 crc kubenswrapper[4824]: I1006 10:41:42.209558 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-59jnp" event={"ID":"910d1ffe-8746-4412-bba0-74bf7ff0e1ea","Type":"ContainerDied","Data":"e4388ad67a18c5f1e3481df8acbc412ced18e612dd0fcd4a9540cc01df747935"} Oct 06 10:41:43 crc kubenswrapper[4824]: I1006 10:41:43.664151 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-59jnp" Oct 06 10:41:43 crc kubenswrapper[4824]: I1006 10:41:43.755398 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/910d1ffe-8746-4412-bba0-74bf7ff0e1ea-inventory\") pod \"910d1ffe-8746-4412-bba0-74bf7ff0e1ea\" (UID: \"910d1ffe-8746-4412-bba0-74bf7ff0e1ea\") " Oct 06 10:41:43 crc kubenswrapper[4824]: I1006 10:41:43.755903 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/910d1ffe-8746-4412-bba0-74bf7ff0e1ea-nova-cell1-compute-config-1\") pod \"910d1ffe-8746-4412-bba0-74bf7ff0e1ea\" (UID: \"910d1ffe-8746-4412-bba0-74bf7ff0e1ea\") " Oct 06 10:41:43 crc kubenswrapper[4824]: I1006 10:41:43.755963 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/910d1ffe-8746-4412-bba0-74bf7ff0e1ea-nova-cell1-compute-config-0\") pod \"910d1ffe-8746-4412-bba0-74bf7ff0e1ea\" (UID: \"910d1ffe-8746-4412-bba0-74bf7ff0e1ea\") " Oct 06 10:41:43 crc kubenswrapper[4824]: I1006 10:41:43.756010 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/910d1ffe-8746-4412-bba0-74bf7ff0e1ea-nova-extra-config-0\") pod \"910d1ffe-8746-4412-bba0-74bf7ff0e1ea\" (UID: \"910d1ffe-8746-4412-bba0-74bf7ff0e1ea\") " Oct 06 10:41:43 crc kubenswrapper[4824]: I1006 10:41:43.756037 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/910d1ffe-8746-4412-bba0-74bf7ff0e1ea-nova-migration-ssh-key-1\") pod \"910d1ffe-8746-4412-bba0-74bf7ff0e1ea\" (UID: \"910d1ffe-8746-4412-bba0-74bf7ff0e1ea\") " Oct 06 10:41:43 crc kubenswrapper[4824]: I1006 10:41:43.756118 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/910d1ffe-8746-4412-bba0-74bf7ff0e1ea-nova-migration-ssh-key-0\") pod \"910d1ffe-8746-4412-bba0-74bf7ff0e1ea\" (UID: \"910d1ffe-8746-4412-bba0-74bf7ff0e1ea\") " Oct 06 10:41:43 crc kubenswrapper[4824]: I1006 10:41:43.756161 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/910d1ffe-8746-4412-bba0-74bf7ff0e1ea-nova-combined-ca-bundle\") pod \"910d1ffe-8746-4412-bba0-74bf7ff0e1ea\" (UID: \"910d1ffe-8746-4412-bba0-74bf7ff0e1ea\") " Oct 06 10:41:43 crc kubenswrapper[4824]: I1006 10:41:43.756238 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pp844\" (UniqueName: \"kubernetes.io/projected/910d1ffe-8746-4412-bba0-74bf7ff0e1ea-kube-api-access-pp844\") pod \"910d1ffe-8746-4412-bba0-74bf7ff0e1ea\" (UID: \"910d1ffe-8746-4412-bba0-74bf7ff0e1ea\") " Oct 06 10:41:43 crc kubenswrapper[4824]: I1006 10:41:43.756261 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/910d1ffe-8746-4412-bba0-74bf7ff0e1ea-ssh-key\") pod \"910d1ffe-8746-4412-bba0-74bf7ff0e1ea\" (UID: \"910d1ffe-8746-4412-bba0-74bf7ff0e1ea\") " Oct 06 10:41:43 crc kubenswrapper[4824]: I1006 10:41:43.761737 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/910d1ffe-8746-4412-bba0-74bf7ff0e1ea-kube-api-access-pp844" (OuterVolumeSpecName: "kube-api-access-pp844") pod "910d1ffe-8746-4412-bba0-74bf7ff0e1ea" (UID: "910d1ffe-8746-4412-bba0-74bf7ff0e1ea"). InnerVolumeSpecName "kube-api-access-pp844". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:41:43 crc kubenswrapper[4824]: I1006 10:41:43.776757 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/910d1ffe-8746-4412-bba0-74bf7ff0e1ea-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "910d1ffe-8746-4412-bba0-74bf7ff0e1ea" (UID: "910d1ffe-8746-4412-bba0-74bf7ff0e1ea"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:41:43 crc kubenswrapper[4824]: I1006 10:41:43.786569 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/910d1ffe-8746-4412-bba0-74bf7ff0e1ea-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "910d1ffe-8746-4412-bba0-74bf7ff0e1ea" (UID: "910d1ffe-8746-4412-bba0-74bf7ff0e1ea"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:41:43 crc kubenswrapper[4824]: I1006 10:41:43.788260 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/910d1ffe-8746-4412-bba0-74bf7ff0e1ea-inventory" (OuterVolumeSpecName: "inventory") pod "910d1ffe-8746-4412-bba0-74bf7ff0e1ea" (UID: "910d1ffe-8746-4412-bba0-74bf7ff0e1ea"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:41:43 crc kubenswrapper[4824]: I1006 10:41:43.789828 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/910d1ffe-8746-4412-bba0-74bf7ff0e1ea-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "910d1ffe-8746-4412-bba0-74bf7ff0e1ea" (UID: "910d1ffe-8746-4412-bba0-74bf7ff0e1ea"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:41:43 crc kubenswrapper[4824]: I1006 10:41:43.790376 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/910d1ffe-8746-4412-bba0-74bf7ff0e1ea-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "910d1ffe-8746-4412-bba0-74bf7ff0e1ea" (UID: "910d1ffe-8746-4412-bba0-74bf7ff0e1ea"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:41:43 crc kubenswrapper[4824]: I1006 10:41:43.794086 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/910d1ffe-8746-4412-bba0-74bf7ff0e1ea-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "910d1ffe-8746-4412-bba0-74bf7ff0e1ea" (UID: "910d1ffe-8746-4412-bba0-74bf7ff0e1ea"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:41:43 crc kubenswrapper[4824]: I1006 10:41:43.797004 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/910d1ffe-8746-4412-bba0-74bf7ff0e1ea-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "910d1ffe-8746-4412-bba0-74bf7ff0e1ea" (UID: "910d1ffe-8746-4412-bba0-74bf7ff0e1ea"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:41:43 crc kubenswrapper[4824]: I1006 10:41:43.811818 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/910d1ffe-8746-4412-bba0-74bf7ff0e1ea-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "910d1ffe-8746-4412-bba0-74bf7ff0e1ea" (UID: "910d1ffe-8746-4412-bba0-74bf7ff0e1ea"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:41:43 crc kubenswrapper[4824]: I1006 10:41:43.859277 4824 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/910d1ffe-8746-4412-bba0-74bf7ff0e1ea-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Oct 06 10:41:43 crc kubenswrapper[4824]: I1006 10:41:43.859318 4824 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/910d1ffe-8746-4412-bba0-74bf7ff0e1ea-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 10:41:43 crc kubenswrapper[4824]: I1006 10:41:43.859327 4824 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/910d1ffe-8746-4412-bba0-74bf7ff0e1ea-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 06 10:41:43 crc kubenswrapper[4824]: I1006 10:41:43.859336 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pp844\" (UniqueName: \"kubernetes.io/projected/910d1ffe-8746-4412-bba0-74bf7ff0e1ea-kube-api-access-pp844\") on node \"crc\" DevicePath \"\"" Oct 06 10:41:43 crc kubenswrapper[4824]: I1006 10:41:43.859344 4824 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/910d1ffe-8746-4412-bba0-74bf7ff0e1ea-inventory\") on node \"crc\" DevicePath \"\"" Oct 06 10:41:43 crc kubenswrapper[4824]: I1006 10:41:43.859355 4824 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/910d1ffe-8746-4412-bba0-74bf7ff0e1ea-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Oct 06 10:41:43 crc kubenswrapper[4824]: I1006 10:41:43.859366 4824 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/910d1ffe-8746-4412-bba0-74bf7ff0e1ea-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Oct 06 10:41:43 crc kubenswrapper[4824]: I1006 10:41:43.859399 4824 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/910d1ffe-8746-4412-bba0-74bf7ff0e1ea-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Oct 06 10:41:43 crc kubenswrapper[4824]: I1006 10:41:43.859408 4824 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/910d1ffe-8746-4412-bba0-74bf7ff0e1ea-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Oct 06 10:41:44 crc kubenswrapper[4824]: I1006 10:41:44.241211 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-59jnp" event={"ID":"910d1ffe-8746-4412-bba0-74bf7ff0e1ea","Type":"ContainerDied","Data":"63301ce24b43dbc109213d81460fd9b1f8b79b9aabe9b370ec2dcb5e8092fd76"} Oct 06 10:41:44 crc kubenswrapper[4824]: I1006 10:41:44.241269 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="63301ce24b43dbc109213d81460fd9b1f8b79b9aabe9b370ec2dcb5e8092fd76" Oct 06 10:41:44 crc kubenswrapper[4824]: I1006 10:41:44.241305 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-59jnp" Oct 06 10:41:44 crc kubenswrapper[4824]: I1006 10:41:44.275561 4824 scope.go:117] "RemoveContainer" containerID="defbb9e5d617c326a5dda4c68f7311eec28a02ec245ca3527fbdf00979129a82" Oct 06 10:41:44 crc kubenswrapper[4824]: E1006 10:41:44.275892 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-khgzw_openshift-machine-config-operator(e1140c8c-93f1-4fce-8c06-a69561ff8a6d)\"" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" Oct 06 10:41:44 crc kubenswrapper[4824]: I1006 10:41:44.380747 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-4br4q"] Oct 06 10:41:44 crc kubenswrapper[4824]: E1006 10:41:44.381250 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b66b0f49-bb9a-49b8-9a5a-2aca8d5b620e" containerName="extract-utilities" Oct 06 10:41:44 crc kubenswrapper[4824]: I1006 10:41:44.381273 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="b66b0f49-bb9a-49b8-9a5a-2aca8d5b620e" containerName="extract-utilities" Oct 06 10:41:44 crc kubenswrapper[4824]: E1006 10:41:44.381290 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b66b0f49-bb9a-49b8-9a5a-2aca8d5b620e" containerName="registry-server" Oct 06 10:41:44 crc kubenswrapper[4824]: I1006 10:41:44.381298 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="b66b0f49-bb9a-49b8-9a5a-2aca8d5b620e" containerName="registry-server" Oct 06 10:41:44 crc kubenswrapper[4824]: E1006 10:41:44.381317 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="910d1ffe-8746-4412-bba0-74bf7ff0e1ea" containerName="nova-edpm-deployment-openstack-edpm-ipam" Oct 06 10:41:44 crc kubenswrapper[4824]: I1006 10:41:44.381323 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="910d1ffe-8746-4412-bba0-74bf7ff0e1ea" containerName="nova-edpm-deployment-openstack-edpm-ipam" Oct 06 10:41:44 crc kubenswrapper[4824]: E1006 10:41:44.381344 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b66b0f49-bb9a-49b8-9a5a-2aca8d5b620e" containerName="extract-content" Oct 06 10:41:44 crc kubenswrapper[4824]: I1006 10:41:44.381351 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="b66b0f49-bb9a-49b8-9a5a-2aca8d5b620e" containerName="extract-content" Oct 06 10:41:44 crc kubenswrapper[4824]: I1006 10:41:44.381530 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="b66b0f49-bb9a-49b8-9a5a-2aca8d5b620e" containerName="registry-server" Oct 06 10:41:44 crc kubenswrapper[4824]: I1006 10:41:44.381556 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="910d1ffe-8746-4412-bba0-74bf7ff0e1ea" containerName="nova-edpm-deployment-openstack-edpm-ipam" Oct 06 10:41:44 crc kubenswrapper[4824]: I1006 10:41:44.382391 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-4br4q" Oct 06 10:41:44 crc kubenswrapper[4824]: I1006 10:41:44.387302 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 06 10:41:44 crc kubenswrapper[4824]: I1006 10:41:44.387478 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 06 10:41:44 crc kubenswrapper[4824]: I1006 10:41:44.387791 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Oct 06 10:41:44 crc kubenswrapper[4824]: I1006 10:41:44.388003 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-f886w" Oct 06 10:41:44 crc kubenswrapper[4824]: I1006 10:41:44.388132 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 06 10:41:44 crc kubenswrapper[4824]: I1006 10:41:44.410697 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-4br4q"] Oct 06 10:41:44 crc kubenswrapper[4824]: I1006 10:41:44.469263 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a410146-5f2b-48a0-824f-ee424f685694-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-4br4q\" (UID: \"3a410146-5f2b-48a0-824f-ee424f685694\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-4br4q" Oct 06 10:41:44 crc kubenswrapper[4824]: I1006 10:41:44.469311 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3a410146-5f2b-48a0-824f-ee424f685694-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-4br4q\" (UID: \"3a410146-5f2b-48a0-824f-ee424f685694\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-4br4q" Oct 06 10:41:44 crc kubenswrapper[4824]: I1006 10:41:44.469658 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3a410146-5f2b-48a0-824f-ee424f685694-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-4br4q\" (UID: \"3a410146-5f2b-48a0-824f-ee424f685694\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-4br4q" Oct 06 10:41:44 crc kubenswrapper[4824]: I1006 10:41:44.469720 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/3a410146-5f2b-48a0-824f-ee424f685694-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-4br4q\" (UID: \"3a410146-5f2b-48a0-824f-ee424f685694\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-4br4q" Oct 06 10:41:44 crc kubenswrapper[4824]: I1006 10:41:44.469932 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/3a410146-5f2b-48a0-824f-ee424f685694-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-4br4q\" (UID: \"3a410146-5f2b-48a0-824f-ee424f685694\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-4br4q" Oct 06 10:41:44 crc kubenswrapper[4824]: I1006 10:41:44.470010 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/3a410146-5f2b-48a0-824f-ee424f685694-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-4br4q\" (UID: \"3a410146-5f2b-48a0-824f-ee424f685694\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-4br4q" Oct 06 10:41:44 crc kubenswrapper[4824]: I1006 10:41:44.470126 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9qwbc\" (UniqueName: \"kubernetes.io/projected/3a410146-5f2b-48a0-824f-ee424f685694-kube-api-access-9qwbc\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-4br4q\" (UID: \"3a410146-5f2b-48a0-824f-ee424f685694\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-4br4q" Oct 06 10:41:44 crc kubenswrapper[4824]: I1006 10:41:44.572066 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/3a410146-5f2b-48a0-824f-ee424f685694-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-4br4q\" (UID: \"3a410146-5f2b-48a0-824f-ee424f685694\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-4br4q" Oct 06 10:41:44 crc kubenswrapper[4824]: I1006 10:41:44.572149 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/3a410146-5f2b-48a0-824f-ee424f685694-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-4br4q\" (UID: \"3a410146-5f2b-48a0-824f-ee424f685694\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-4br4q" Oct 06 10:41:44 crc kubenswrapper[4824]: I1006 10:41:44.572216 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9qwbc\" (UniqueName: \"kubernetes.io/projected/3a410146-5f2b-48a0-824f-ee424f685694-kube-api-access-9qwbc\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-4br4q\" (UID: \"3a410146-5f2b-48a0-824f-ee424f685694\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-4br4q" Oct 06 10:41:44 crc kubenswrapper[4824]: I1006 10:41:44.572300 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a410146-5f2b-48a0-824f-ee424f685694-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-4br4q\" (UID: \"3a410146-5f2b-48a0-824f-ee424f685694\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-4br4q" Oct 06 10:41:44 crc kubenswrapper[4824]: I1006 10:41:44.572346 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3a410146-5f2b-48a0-824f-ee424f685694-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-4br4q\" (UID: \"3a410146-5f2b-48a0-824f-ee424f685694\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-4br4q" Oct 06 10:41:44 crc kubenswrapper[4824]: I1006 10:41:44.572503 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3a410146-5f2b-48a0-824f-ee424f685694-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-4br4q\" (UID: \"3a410146-5f2b-48a0-824f-ee424f685694\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-4br4q" Oct 06 10:41:44 crc kubenswrapper[4824]: I1006 10:41:44.572541 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/3a410146-5f2b-48a0-824f-ee424f685694-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-4br4q\" (UID: \"3a410146-5f2b-48a0-824f-ee424f685694\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-4br4q" Oct 06 10:41:44 crc kubenswrapper[4824]: I1006 10:41:44.577092 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/3a410146-5f2b-48a0-824f-ee424f685694-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-4br4q\" (UID: \"3a410146-5f2b-48a0-824f-ee424f685694\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-4br4q" Oct 06 10:41:44 crc kubenswrapper[4824]: I1006 10:41:44.577255 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/3a410146-5f2b-48a0-824f-ee424f685694-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-4br4q\" (UID: \"3a410146-5f2b-48a0-824f-ee424f685694\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-4br4q" Oct 06 10:41:44 crc kubenswrapper[4824]: I1006 10:41:44.577413 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3a410146-5f2b-48a0-824f-ee424f685694-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-4br4q\" (UID: \"3a410146-5f2b-48a0-824f-ee424f685694\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-4br4q" Oct 06 10:41:44 crc kubenswrapper[4824]: I1006 10:41:44.585126 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3a410146-5f2b-48a0-824f-ee424f685694-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-4br4q\" (UID: \"3a410146-5f2b-48a0-824f-ee424f685694\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-4br4q" Oct 06 10:41:44 crc kubenswrapper[4824]: I1006 10:41:44.585942 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/3a410146-5f2b-48a0-824f-ee424f685694-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-4br4q\" (UID: \"3a410146-5f2b-48a0-824f-ee424f685694\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-4br4q" Oct 06 10:41:44 crc kubenswrapper[4824]: I1006 10:41:44.585972 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a410146-5f2b-48a0-824f-ee424f685694-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-4br4q\" (UID: \"3a410146-5f2b-48a0-824f-ee424f685694\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-4br4q" Oct 06 10:41:44 crc kubenswrapper[4824]: I1006 10:41:44.591080 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9qwbc\" (UniqueName: \"kubernetes.io/projected/3a410146-5f2b-48a0-824f-ee424f685694-kube-api-access-9qwbc\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-4br4q\" (UID: \"3a410146-5f2b-48a0-824f-ee424f685694\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-4br4q" Oct 06 10:41:44 crc kubenswrapper[4824]: I1006 10:41:44.726768 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-4br4q" Oct 06 10:41:45 crc kubenswrapper[4824]: I1006 10:41:45.308713 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-4br4q"] Oct 06 10:41:46 crc kubenswrapper[4824]: I1006 10:41:46.270004 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-4br4q" event={"ID":"3a410146-5f2b-48a0-824f-ee424f685694","Type":"ContainerStarted","Data":"bab8163b6158aa354be67d6805f05751cf08f765ad57b2fc2f0afb9b002c8b16"} Oct 06 10:41:46 crc kubenswrapper[4824]: I1006 10:41:46.270704 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-4br4q" event={"ID":"3a410146-5f2b-48a0-824f-ee424f685694","Type":"ContainerStarted","Data":"5206ef25ab1fd2549f1cbe9f0b47dfb81bd3d8957f1235ff7819f5c4544c25a9"} Oct 06 10:41:46 crc kubenswrapper[4824]: I1006 10:41:46.293193 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-4br4q" podStartSLOduration=1.7164181630000002 podStartE2EDuration="2.293161475s" podCreationTimestamp="2025-10-06 10:41:44 +0000 UTC" firstStartedPulling="2025-10-06 10:41:45.327001215 +0000 UTC m=+2614.691424076" lastFinishedPulling="2025-10-06 10:41:45.903744487 +0000 UTC m=+2615.268167388" observedRunningTime="2025-10-06 10:41:46.291699979 +0000 UTC m=+2615.656122890" watchObservedRunningTime="2025-10-06 10:41:46.293161475 +0000 UTC m=+2615.657584386" Oct 06 10:41:59 crc kubenswrapper[4824]: I1006 10:41:59.274668 4824 scope.go:117] "RemoveContainer" containerID="defbb9e5d617c326a5dda4c68f7311eec28a02ec245ca3527fbdf00979129a82" Oct 06 10:41:59 crc kubenswrapper[4824]: E1006 10:41:59.275723 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-khgzw_openshift-machine-config-operator(e1140c8c-93f1-4fce-8c06-a69561ff8a6d)\"" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" Oct 06 10:42:12 crc kubenswrapper[4824]: I1006 10:42:12.274032 4824 scope.go:117] "RemoveContainer" containerID="defbb9e5d617c326a5dda4c68f7311eec28a02ec245ca3527fbdf00979129a82" Oct 06 10:42:12 crc kubenswrapper[4824]: E1006 10:42:12.275107 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-khgzw_openshift-machine-config-operator(e1140c8c-93f1-4fce-8c06-a69561ff8a6d)\"" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" Oct 06 10:42:26 crc kubenswrapper[4824]: I1006 10:42:26.274244 4824 scope.go:117] "RemoveContainer" containerID="defbb9e5d617c326a5dda4c68f7311eec28a02ec245ca3527fbdf00979129a82" Oct 06 10:42:26 crc kubenswrapper[4824]: E1006 10:42:26.275368 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-khgzw_openshift-machine-config-operator(e1140c8c-93f1-4fce-8c06-a69561ff8a6d)\"" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" Oct 06 10:42:40 crc kubenswrapper[4824]: I1006 10:42:40.273729 4824 scope.go:117] "RemoveContainer" containerID="defbb9e5d617c326a5dda4c68f7311eec28a02ec245ca3527fbdf00979129a82" Oct 06 10:42:40 crc kubenswrapper[4824]: E1006 10:42:40.275178 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-khgzw_openshift-machine-config-operator(e1140c8c-93f1-4fce-8c06-a69561ff8a6d)\"" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" Oct 06 10:42:52 crc kubenswrapper[4824]: I1006 10:42:52.274466 4824 scope.go:117] "RemoveContainer" containerID="defbb9e5d617c326a5dda4c68f7311eec28a02ec245ca3527fbdf00979129a82" Oct 06 10:42:52 crc kubenswrapper[4824]: E1006 10:42:52.275846 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-khgzw_openshift-machine-config-operator(e1140c8c-93f1-4fce-8c06-a69561ff8a6d)\"" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" Oct 06 10:43:04 crc kubenswrapper[4824]: I1006 10:43:04.274785 4824 scope.go:117] "RemoveContainer" containerID="defbb9e5d617c326a5dda4c68f7311eec28a02ec245ca3527fbdf00979129a82" Oct 06 10:43:04 crc kubenswrapper[4824]: E1006 10:43:04.275620 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-khgzw_openshift-machine-config-operator(e1140c8c-93f1-4fce-8c06-a69561ff8a6d)\"" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" Oct 06 10:43:18 crc kubenswrapper[4824]: I1006 10:43:18.274483 4824 scope.go:117] "RemoveContainer" containerID="defbb9e5d617c326a5dda4c68f7311eec28a02ec245ca3527fbdf00979129a82" Oct 06 10:43:19 crc kubenswrapper[4824]: I1006 10:43:19.451973 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" event={"ID":"e1140c8c-93f1-4fce-8c06-a69561ff8a6d","Type":"ContainerStarted","Data":"db15accdeb46a9039be894301153ff2ab7709ec808de959d20f2a048678d55cc"} Oct 06 10:43:27 crc kubenswrapper[4824]: I1006 10:43:27.922441 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-j4dff"] Oct 06 10:43:27 crc kubenswrapper[4824]: I1006 10:43:27.927492 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-j4dff" Oct 06 10:43:27 crc kubenswrapper[4824]: I1006 10:43:27.935080 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-j4dff"] Oct 06 10:43:27 crc kubenswrapper[4824]: I1006 10:43:27.956527 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c84fn\" (UniqueName: \"kubernetes.io/projected/5ae7f4db-4a50-466c-a56b-f45403a18178-kube-api-access-c84fn\") pod \"community-operators-j4dff\" (UID: \"5ae7f4db-4a50-466c-a56b-f45403a18178\") " pod="openshift-marketplace/community-operators-j4dff" Oct 06 10:43:27 crc kubenswrapper[4824]: I1006 10:43:27.956676 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ae7f4db-4a50-466c-a56b-f45403a18178-utilities\") pod \"community-operators-j4dff\" (UID: \"5ae7f4db-4a50-466c-a56b-f45403a18178\") " pod="openshift-marketplace/community-operators-j4dff" Oct 06 10:43:27 crc kubenswrapper[4824]: I1006 10:43:27.956737 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ae7f4db-4a50-466c-a56b-f45403a18178-catalog-content\") pod \"community-operators-j4dff\" (UID: \"5ae7f4db-4a50-466c-a56b-f45403a18178\") " pod="openshift-marketplace/community-operators-j4dff" Oct 06 10:43:28 crc kubenswrapper[4824]: I1006 10:43:28.058614 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ae7f4db-4a50-466c-a56b-f45403a18178-catalog-content\") pod \"community-operators-j4dff\" (UID: \"5ae7f4db-4a50-466c-a56b-f45403a18178\") " pod="openshift-marketplace/community-operators-j4dff" Oct 06 10:43:28 crc kubenswrapper[4824]: I1006 10:43:28.059093 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c84fn\" (UniqueName: \"kubernetes.io/projected/5ae7f4db-4a50-466c-a56b-f45403a18178-kube-api-access-c84fn\") pod \"community-operators-j4dff\" (UID: \"5ae7f4db-4a50-466c-a56b-f45403a18178\") " pod="openshift-marketplace/community-operators-j4dff" Oct 06 10:43:28 crc kubenswrapper[4824]: I1006 10:43:28.059241 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ae7f4db-4a50-466c-a56b-f45403a18178-utilities\") pod \"community-operators-j4dff\" (UID: \"5ae7f4db-4a50-466c-a56b-f45403a18178\") " pod="openshift-marketplace/community-operators-j4dff" Oct 06 10:43:28 crc kubenswrapper[4824]: I1006 10:43:28.059717 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ae7f4db-4a50-466c-a56b-f45403a18178-catalog-content\") pod \"community-operators-j4dff\" (UID: \"5ae7f4db-4a50-466c-a56b-f45403a18178\") " pod="openshift-marketplace/community-operators-j4dff" Oct 06 10:43:28 crc kubenswrapper[4824]: I1006 10:43:28.059822 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ae7f4db-4a50-466c-a56b-f45403a18178-utilities\") pod \"community-operators-j4dff\" (UID: \"5ae7f4db-4a50-466c-a56b-f45403a18178\") " pod="openshift-marketplace/community-operators-j4dff" Oct 06 10:43:28 crc kubenswrapper[4824]: I1006 10:43:28.088592 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c84fn\" (UniqueName: \"kubernetes.io/projected/5ae7f4db-4a50-466c-a56b-f45403a18178-kube-api-access-c84fn\") pod \"community-operators-j4dff\" (UID: \"5ae7f4db-4a50-466c-a56b-f45403a18178\") " pod="openshift-marketplace/community-operators-j4dff" Oct 06 10:43:28 crc kubenswrapper[4824]: I1006 10:43:28.270441 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-j4dff" Oct 06 10:43:28 crc kubenswrapper[4824]: I1006 10:43:28.868857 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-j4dff"] Oct 06 10:43:29 crc kubenswrapper[4824]: I1006 10:43:29.585360 4824 generic.go:334] "Generic (PLEG): container finished" podID="5ae7f4db-4a50-466c-a56b-f45403a18178" containerID="b0f95a82fc260aefa829dc938a26f6b660584f973a10ba2929d7a743d6dccc89" exitCode=0 Oct 06 10:43:29 crc kubenswrapper[4824]: I1006 10:43:29.585435 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j4dff" event={"ID":"5ae7f4db-4a50-466c-a56b-f45403a18178","Type":"ContainerDied","Data":"b0f95a82fc260aefa829dc938a26f6b660584f973a10ba2929d7a743d6dccc89"} Oct 06 10:43:29 crc kubenswrapper[4824]: I1006 10:43:29.585916 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j4dff" event={"ID":"5ae7f4db-4a50-466c-a56b-f45403a18178","Type":"ContainerStarted","Data":"fa1415f9fd4cabb3ef4302faa4d911b46ca4f8ab0fa8b9db6ac636750edefab5"} Oct 06 10:43:29 crc kubenswrapper[4824]: I1006 10:43:29.590356 4824 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 06 10:43:30 crc kubenswrapper[4824]: I1006 10:43:30.604631 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j4dff" event={"ID":"5ae7f4db-4a50-466c-a56b-f45403a18178","Type":"ContainerStarted","Data":"a76f3bdd80c8e9e7b79ecb00deb2ff8105842eaff3b6383604fbad1a70fe3752"} Oct 06 10:43:31 crc kubenswrapper[4824]: I1006 10:43:31.620295 4824 generic.go:334] "Generic (PLEG): container finished" podID="5ae7f4db-4a50-466c-a56b-f45403a18178" containerID="a76f3bdd80c8e9e7b79ecb00deb2ff8105842eaff3b6383604fbad1a70fe3752" exitCode=0 Oct 06 10:43:31 crc kubenswrapper[4824]: I1006 10:43:31.620351 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j4dff" event={"ID":"5ae7f4db-4a50-466c-a56b-f45403a18178","Type":"ContainerDied","Data":"a76f3bdd80c8e9e7b79ecb00deb2ff8105842eaff3b6383604fbad1a70fe3752"} Oct 06 10:43:32 crc kubenswrapper[4824]: I1006 10:43:32.633553 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j4dff" event={"ID":"5ae7f4db-4a50-466c-a56b-f45403a18178","Type":"ContainerStarted","Data":"53272273f27d8bf96a7c822f8aa6576206714283bfbc4ba74de436460b9b5e4d"} Oct 06 10:43:32 crc kubenswrapper[4824]: I1006 10:43:32.656928 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-j4dff" podStartSLOduration=3.169959141 podStartE2EDuration="5.656901472s" podCreationTimestamp="2025-10-06 10:43:27 +0000 UTC" firstStartedPulling="2025-10-06 10:43:29.590032762 +0000 UTC m=+2718.954455623" lastFinishedPulling="2025-10-06 10:43:32.076975073 +0000 UTC m=+2721.441397954" observedRunningTime="2025-10-06 10:43:32.650494255 +0000 UTC m=+2722.014917116" watchObservedRunningTime="2025-10-06 10:43:32.656901472 +0000 UTC m=+2722.021324333" Oct 06 10:43:38 crc kubenswrapper[4824]: I1006 10:43:38.271389 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-j4dff" Oct 06 10:43:38 crc kubenswrapper[4824]: I1006 10:43:38.272238 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-j4dff" Oct 06 10:43:38 crc kubenswrapper[4824]: I1006 10:43:38.367887 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-j4dff" Oct 06 10:43:38 crc kubenswrapper[4824]: I1006 10:43:38.797614 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-j4dff" Oct 06 10:43:38 crc kubenswrapper[4824]: I1006 10:43:38.903790 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-j4dff"] Oct 06 10:43:40 crc kubenswrapper[4824]: I1006 10:43:40.735513 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-j4dff" podUID="5ae7f4db-4a50-466c-a56b-f45403a18178" containerName="registry-server" containerID="cri-o://53272273f27d8bf96a7c822f8aa6576206714283bfbc4ba74de436460b9b5e4d" gracePeriod=2 Oct 06 10:43:41 crc kubenswrapper[4824]: I1006 10:43:41.054808 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-8vqd9"] Oct 06 10:43:41 crc kubenswrapper[4824]: I1006 10:43:41.058047 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8vqd9" Oct 06 10:43:41 crc kubenswrapper[4824]: I1006 10:43:41.068335 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8vqd9"] Oct 06 10:43:41 crc kubenswrapper[4824]: I1006 10:43:41.111426 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2psc8\" (UniqueName: \"kubernetes.io/projected/cfb03cbf-729e-4c5a-b1e5-d5e6ca02e497-kube-api-access-2psc8\") pod \"redhat-marketplace-8vqd9\" (UID: \"cfb03cbf-729e-4c5a-b1e5-d5e6ca02e497\") " pod="openshift-marketplace/redhat-marketplace-8vqd9" Oct 06 10:43:41 crc kubenswrapper[4824]: I1006 10:43:41.111736 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cfb03cbf-729e-4c5a-b1e5-d5e6ca02e497-utilities\") pod \"redhat-marketplace-8vqd9\" (UID: \"cfb03cbf-729e-4c5a-b1e5-d5e6ca02e497\") " pod="openshift-marketplace/redhat-marketplace-8vqd9" Oct 06 10:43:41 crc kubenswrapper[4824]: I1006 10:43:41.111782 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cfb03cbf-729e-4c5a-b1e5-d5e6ca02e497-catalog-content\") pod \"redhat-marketplace-8vqd9\" (UID: \"cfb03cbf-729e-4c5a-b1e5-d5e6ca02e497\") " pod="openshift-marketplace/redhat-marketplace-8vqd9" Oct 06 10:43:41 crc kubenswrapper[4824]: I1006 10:43:41.214525 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cfb03cbf-729e-4c5a-b1e5-d5e6ca02e497-utilities\") pod \"redhat-marketplace-8vqd9\" (UID: \"cfb03cbf-729e-4c5a-b1e5-d5e6ca02e497\") " pod="openshift-marketplace/redhat-marketplace-8vqd9" Oct 06 10:43:41 crc kubenswrapper[4824]: I1006 10:43:41.214578 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cfb03cbf-729e-4c5a-b1e5-d5e6ca02e497-catalog-content\") pod \"redhat-marketplace-8vqd9\" (UID: \"cfb03cbf-729e-4c5a-b1e5-d5e6ca02e497\") " pod="openshift-marketplace/redhat-marketplace-8vqd9" Oct 06 10:43:41 crc kubenswrapper[4824]: I1006 10:43:41.214703 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2psc8\" (UniqueName: \"kubernetes.io/projected/cfb03cbf-729e-4c5a-b1e5-d5e6ca02e497-kube-api-access-2psc8\") pod \"redhat-marketplace-8vqd9\" (UID: \"cfb03cbf-729e-4c5a-b1e5-d5e6ca02e497\") " pod="openshift-marketplace/redhat-marketplace-8vqd9" Oct 06 10:43:41 crc kubenswrapper[4824]: I1006 10:43:41.215184 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cfb03cbf-729e-4c5a-b1e5-d5e6ca02e497-utilities\") pod \"redhat-marketplace-8vqd9\" (UID: \"cfb03cbf-729e-4c5a-b1e5-d5e6ca02e497\") " pod="openshift-marketplace/redhat-marketplace-8vqd9" Oct 06 10:43:41 crc kubenswrapper[4824]: I1006 10:43:41.215408 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cfb03cbf-729e-4c5a-b1e5-d5e6ca02e497-catalog-content\") pod \"redhat-marketplace-8vqd9\" (UID: \"cfb03cbf-729e-4c5a-b1e5-d5e6ca02e497\") " pod="openshift-marketplace/redhat-marketplace-8vqd9" Oct 06 10:43:41 crc kubenswrapper[4824]: I1006 10:43:41.237565 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2psc8\" (UniqueName: \"kubernetes.io/projected/cfb03cbf-729e-4c5a-b1e5-d5e6ca02e497-kube-api-access-2psc8\") pod \"redhat-marketplace-8vqd9\" (UID: \"cfb03cbf-729e-4c5a-b1e5-d5e6ca02e497\") " pod="openshift-marketplace/redhat-marketplace-8vqd9" Oct 06 10:43:41 crc kubenswrapper[4824]: I1006 10:43:41.330312 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-j4dff" Oct 06 10:43:41 crc kubenswrapper[4824]: I1006 10:43:41.409901 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8vqd9" Oct 06 10:43:41 crc kubenswrapper[4824]: I1006 10:43:41.421949 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c84fn\" (UniqueName: \"kubernetes.io/projected/5ae7f4db-4a50-466c-a56b-f45403a18178-kube-api-access-c84fn\") pod \"5ae7f4db-4a50-466c-a56b-f45403a18178\" (UID: \"5ae7f4db-4a50-466c-a56b-f45403a18178\") " Oct 06 10:43:41 crc kubenswrapper[4824]: I1006 10:43:41.423128 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ae7f4db-4a50-466c-a56b-f45403a18178-utilities\") pod \"5ae7f4db-4a50-466c-a56b-f45403a18178\" (UID: \"5ae7f4db-4a50-466c-a56b-f45403a18178\") " Oct 06 10:43:41 crc kubenswrapper[4824]: I1006 10:43:41.423208 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ae7f4db-4a50-466c-a56b-f45403a18178-catalog-content\") pod \"5ae7f4db-4a50-466c-a56b-f45403a18178\" (UID: \"5ae7f4db-4a50-466c-a56b-f45403a18178\") " Oct 06 10:43:41 crc kubenswrapper[4824]: I1006 10:43:41.424123 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5ae7f4db-4a50-466c-a56b-f45403a18178-utilities" (OuterVolumeSpecName: "utilities") pod "5ae7f4db-4a50-466c-a56b-f45403a18178" (UID: "5ae7f4db-4a50-466c-a56b-f45403a18178"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:43:41 crc kubenswrapper[4824]: I1006 10:43:41.431017 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5ae7f4db-4a50-466c-a56b-f45403a18178-kube-api-access-c84fn" (OuterVolumeSpecName: "kube-api-access-c84fn") pod "5ae7f4db-4a50-466c-a56b-f45403a18178" (UID: "5ae7f4db-4a50-466c-a56b-f45403a18178"). InnerVolumeSpecName "kube-api-access-c84fn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:43:41 crc kubenswrapper[4824]: I1006 10:43:41.478707 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5ae7f4db-4a50-466c-a56b-f45403a18178-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5ae7f4db-4a50-466c-a56b-f45403a18178" (UID: "5ae7f4db-4a50-466c-a56b-f45403a18178"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:43:41 crc kubenswrapper[4824]: I1006 10:43:41.525768 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ae7f4db-4a50-466c-a56b-f45403a18178-utilities\") on node \"crc\" DevicePath \"\"" Oct 06 10:43:41 crc kubenswrapper[4824]: I1006 10:43:41.525801 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ae7f4db-4a50-466c-a56b-f45403a18178-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 06 10:43:41 crc kubenswrapper[4824]: I1006 10:43:41.525834 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c84fn\" (UniqueName: \"kubernetes.io/projected/5ae7f4db-4a50-466c-a56b-f45403a18178-kube-api-access-c84fn\") on node \"crc\" DevicePath \"\"" Oct 06 10:43:41 crc kubenswrapper[4824]: I1006 10:43:41.750025 4824 generic.go:334] "Generic (PLEG): container finished" podID="5ae7f4db-4a50-466c-a56b-f45403a18178" containerID="53272273f27d8bf96a7c822f8aa6576206714283bfbc4ba74de436460b9b5e4d" exitCode=0 Oct 06 10:43:41 crc kubenswrapper[4824]: I1006 10:43:41.750088 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j4dff" event={"ID":"5ae7f4db-4a50-466c-a56b-f45403a18178","Type":"ContainerDied","Data":"53272273f27d8bf96a7c822f8aa6576206714283bfbc4ba74de436460b9b5e4d"} Oct 06 10:43:41 crc kubenswrapper[4824]: I1006 10:43:41.750101 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-j4dff" Oct 06 10:43:41 crc kubenswrapper[4824]: I1006 10:43:41.750143 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j4dff" event={"ID":"5ae7f4db-4a50-466c-a56b-f45403a18178","Type":"ContainerDied","Data":"fa1415f9fd4cabb3ef4302faa4d911b46ca4f8ab0fa8b9db6ac636750edefab5"} Oct 06 10:43:41 crc kubenswrapper[4824]: I1006 10:43:41.750176 4824 scope.go:117] "RemoveContainer" containerID="53272273f27d8bf96a7c822f8aa6576206714283bfbc4ba74de436460b9b5e4d" Oct 06 10:43:41 crc kubenswrapper[4824]: I1006 10:43:41.789354 4824 scope.go:117] "RemoveContainer" containerID="a76f3bdd80c8e9e7b79ecb00deb2ff8105842eaff3b6383604fbad1a70fe3752" Oct 06 10:43:41 crc kubenswrapper[4824]: I1006 10:43:41.795949 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-j4dff"] Oct 06 10:43:41 crc kubenswrapper[4824]: I1006 10:43:41.806065 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-j4dff"] Oct 06 10:43:41 crc kubenswrapper[4824]: I1006 10:43:41.815670 4824 scope.go:117] "RemoveContainer" containerID="b0f95a82fc260aefa829dc938a26f6b660584f973a10ba2929d7a743d6dccc89" Oct 06 10:43:41 crc kubenswrapper[4824]: I1006 10:43:41.838556 4824 scope.go:117] "RemoveContainer" containerID="53272273f27d8bf96a7c822f8aa6576206714283bfbc4ba74de436460b9b5e4d" Oct 06 10:43:41 crc kubenswrapper[4824]: E1006 10:43:41.839345 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"53272273f27d8bf96a7c822f8aa6576206714283bfbc4ba74de436460b9b5e4d\": container with ID starting with 53272273f27d8bf96a7c822f8aa6576206714283bfbc4ba74de436460b9b5e4d not found: ID does not exist" containerID="53272273f27d8bf96a7c822f8aa6576206714283bfbc4ba74de436460b9b5e4d" Oct 06 10:43:41 crc kubenswrapper[4824]: I1006 10:43:41.839408 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"53272273f27d8bf96a7c822f8aa6576206714283bfbc4ba74de436460b9b5e4d"} err="failed to get container status \"53272273f27d8bf96a7c822f8aa6576206714283bfbc4ba74de436460b9b5e4d\": rpc error: code = NotFound desc = could not find container \"53272273f27d8bf96a7c822f8aa6576206714283bfbc4ba74de436460b9b5e4d\": container with ID starting with 53272273f27d8bf96a7c822f8aa6576206714283bfbc4ba74de436460b9b5e4d not found: ID does not exist" Oct 06 10:43:41 crc kubenswrapper[4824]: I1006 10:43:41.839455 4824 scope.go:117] "RemoveContainer" containerID="a76f3bdd80c8e9e7b79ecb00deb2ff8105842eaff3b6383604fbad1a70fe3752" Oct 06 10:43:41 crc kubenswrapper[4824]: E1006 10:43:41.840165 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a76f3bdd80c8e9e7b79ecb00deb2ff8105842eaff3b6383604fbad1a70fe3752\": container with ID starting with a76f3bdd80c8e9e7b79ecb00deb2ff8105842eaff3b6383604fbad1a70fe3752 not found: ID does not exist" containerID="a76f3bdd80c8e9e7b79ecb00deb2ff8105842eaff3b6383604fbad1a70fe3752" Oct 06 10:43:41 crc kubenswrapper[4824]: I1006 10:43:41.840212 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a76f3bdd80c8e9e7b79ecb00deb2ff8105842eaff3b6383604fbad1a70fe3752"} err="failed to get container status \"a76f3bdd80c8e9e7b79ecb00deb2ff8105842eaff3b6383604fbad1a70fe3752\": rpc error: code = NotFound desc = could not find container \"a76f3bdd80c8e9e7b79ecb00deb2ff8105842eaff3b6383604fbad1a70fe3752\": container with ID starting with a76f3bdd80c8e9e7b79ecb00deb2ff8105842eaff3b6383604fbad1a70fe3752 not found: ID does not exist" Oct 06 10:43:41 crc kubenswrapper[4824]: I1006 10:43:41.840261 4824 scope.go:117] "RemoveContainer" containerID="b0f95a82fc260aefa829dc938a26f6b660584f973a10ba2929d7a743d6dccc89" Oct 06 10:43:41 crc kubenswrapper[4824]: E1006 10:43:41.840689 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b0f95a82fc260aefa829dc938a26f6b660584f973a10ba2929d7a743d6dccc89\": container with ID starting with b0f95a82fc260aefa829dc938a26f6b660584f973a10ba2929d7a743d6dccc89 not found: ID does not exist" containerID="b0f95a82fc260aefa829dc938a26f6b660584f973a10ba2929d7a743d6dccc89" Oct 06 10:43:41 crc kubenswrapper[4824]: I1006 10:43:41.840773 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b0f95a82fc260aefa829dc938a26f6b660584f973a10ba2929d7a743d6dccc89"} err="failed to get container status \"b0f95a82fc260aefa829dc938a26f6b660584f973a10ba2929d7a743d6dccc89\": rpc error: code = NotFound desc = could not find container \"b0f95a82fc260aefa829dc938a26f6b660584f973a10ba2929d7a743d6dccc89\": container with ID starting with b0f95a82fc260aefa829dc938a26f6b660584f973a10ba2929d7a743d6dccc89 not found: ID does not exist" Oct 06 10:43:41 crc kubenswrapper[4824]: W1006 10:43:41.926623 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcfb03cbf_729e_4c5a_b1e5_d5e6ca02e497.slice/crio-5725a8e272a9d38c1e4caa0aa9084c5c27d91602c1bdc36668838a9db8248162 WatchSource:0}: Error finding container 5725a8e272a9d38c1e4caa0aa9084c5c27d91602c1bdc36668838a9db8248162: Status 404 returned error can't find the container with id 5725a8e272a9d38c1e4caa0aa9084c5c27d91602c1bdc36668838a9db8248162 Oct 06 10:43:41 crc kubenswrapper[4824]: I1006 10:43:41.935742 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8vqd9"] Oct 06 10:43:42 crc kubenswrapper[4824]: I1006 10:43:42.788404 4824 generic.go:334] "Generic (PLEG): container finished" podID="cfb03cbf-729e-4c5a-b1e5-d5e6ca02e497" containerID="a821817c7c09bde17a1072a065fec9fe9b26a9b9059ca387f12929056051d1d4" exitCode=0 Oct 06 10:43:42 crc kubenswrapper[4824]: I1006 10:43:42.788583 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8vqd9" event={"ID":"cfb03cbf-729e-4c5a-b1e5-d5e6ca02e497","Type":"ContainerDied","Data":"a821817c7c09bde17a1072a065fec9fe9b26a9b9059ca387f12929056051d1d4"} Oct 06 10:43:42 crc kubenswrapper[4824]: I1006 10:43:42.788938 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8vqd9" event={"ID":"cfb03cbf-729e-4c5a-b1e5-d5e6ca02e497","Type":"ContainerStarted","Data":"5725a8e272a9d38c1e4caa0aa9084c5c27d91602c1bdc36668838a9db8248162"} Oct 06 10:43:43 crc kubenswrapper[4824]: I1006 10:43:43.301319 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5ae7f4db-4a50-466c-a56b-f45403a18178" path="/var/lib/kubelet/pods/5ae7f4db-4a50-466c-a56b-f45403a18178/volumes" Oct 06 10:43:44 crc kubenswrapper[4824]: I1006 10:43:44.814255 4824 generic.go:334] "Generic (PLEG): container finished" podID="cfb03cbf-729e-4c5a-b1e5-d5e6ca02e497" containerID="2404ad3886244359893e8aac325397bd46270ef6506ad89f43408b2dfdd9b58b" exitCode=0 Oct 06 10:43:44 crc kubenswrapper[4824]: I1006 10:43:44.814357 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8vqd9" event={"ID":"cfb03cbf-729e-4c5a-b1e5-d5e6ca02e497","Type":"ContainerDied","Data":"2404ad3886244359893e8aac325397bd46270ef6506ad89f43408b2dfdd9b58b"} Oct 06 10:43:45 crc kubenswrapper[4824]: I1006 10:43:45.828944 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8vqd9" event={"ID":"cfb03cbf-729e-4c5a-b1e5-d5e6ca02e497","Type":"ContainerStarted","Data":"f2162519cdfe4d1e8ffcec6d522a6173492d372ddaeb5810278f0b2a47caf13d"} Oct 06 10:43:45 crc kubenswrapper[4824]: I1006 10:43:45.855405 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-8vqd9" podStartSLOduration=2.385033828 podStartE2EDuration="4.855380843s" podCreationTimestamp="2025-10-06 10:43:41 +0000 UTC" firstStartedPulling="2025-10-06 10:43:42.814913258 +0000 UTC m=+2732.179336139" lastFinishedPulling="2025-10-06 10:43:45.285260283 +0000 UTC m=+2734.649683154" observedRunningTime="2025-10-06 10:43:45.848846354 +0000 UTC m=+2735.213269235" watchObservedRunningTime="2025-10-06 10:43:45.855380843 +0000 UTC m=+2735.219803724" Oct 06 10:43:51 crc kubenswrapper[4824]: I1006 10:43:51.411849 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-8vqd9" Oct 06 10:43:51 crc kubenswrapper[4824]: I1006 10:43:51.412734 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-8vqd9" Oct 06 10:43:51 crc kubenswrapper[4824]: I1006 10:43:51.470335 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-8vqd9" Oct 06 10:43:51 crc kubenswrapper[4824]: I1006 10:43:51.980352 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-8vqd9" Oct 06 10:43:52 crc kubenswrapper[4824]: I1006 10:43:52.625695 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-8vqd9"] Oct 06 10:43:53 crc kubenswrapper[4824]: I1006 10:43:53.939331 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-8vqd9" podUID="cfb03cbf-729e-4c5a-b1e5-d5e6ca02e497" containerName="registry-server" containerID="cri-o://f2162519cdfe4d1e8ffcec6d522a6173492d372ddaeb5810278f0b2a47caf13d" gracePeriod=2 Oct 06 10:43:54 crc kubenswrapper[4824]: I1006 10:43:54.434952 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8vqd9" Oct 06 10:43:54 crc kubenswrapper[4824]: I1006 10:43:54.542627 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cfb03cbf-729e-4c5a-b1e5-d5e6ca02e497-utilities\") pod \"cfb03cbf-729e-4c5a-b1e5-d5e6ca02e497\" (UID: \"cfb03cbf-729e-4c5a-b1e5-d5e6ca02e497\") " Oct 06 10:43:54 crc kubenswrapper[4824]: I1006 10:43:54.542817 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cfb03cbf-729e-4c5a-b1e5-d5e6ca02e497-catalog-content\") pod \"cfb03cbf-729e-4c5a-b1e5-d5e6ca02e497\" (UID: \"cfb03cbf-729e-4c5a-b1e5-d5e6ca02e497\") " Oct 06 10:43:54 crc kubenswrapper[4824]: I1006 10:43:54.542922 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2psc8\" (UniqueName: \"kubernetes.io/projected/cfb03cbf-729e-4c5a-b1e5-d5e6ca02e497-kube-api-access-2psc8\") pod \"cfb03cbf-729e-4c5a-b1e5-d5e6ca02e497\" (UID: \"cfb03cbf-729e-4c5a-b1e5-d5e6ca02e497\") " Oct 06 10:43:54 crc kubenswrapper[4824]: I1006 10:43:54.544403 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cfb03cbf-729e-4c5a-b1e5-d5e6ca02e497-utilities" (OuterVolumeSpecName: "utilities") pod "cfb03cbf-729e-4c5a-b1e5-d5e6ca02e497" (UID: "cfb03cbf-729e-4c5a-b1e5-d5e6ca02e497"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:43:54 crc kubenswrapper[4824]: I1006 10:43:54.554449 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cfb03cbf-729e-4c5a-b1e5-d5e6ca02e497-kube-api-access-2psc8" (OuterVolumeSpecName: "kube-api-access-2psc8") pod "cfb03cbf-729e-4c5a-b1e5-d5e6ca02e497" (UID: "cfb03cbf-729e-4c5a-b1e5-d5e6ca02e497"). InnerVolumeSpecName "kube-api-access-2psc8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:43:54 crc kubenswrapper[4824]: I1006 10:43:54.572446 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cfb03cbf-729e-4c5a-b1e5-d5e6ca02e497-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cfb03cbf-729e-4c5a-b1e5-d5e6ca02e497" (UID: "cfb03cbf-729e-4c5a-b1e5-d5e6ca02e497"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:43:54 crc kubenswrapper[4824]: I1006 10:43:54.645363 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cfb03cbf-729e-4c5a-b1e5-d5e6ca02e497-utilities\") on node \"crc\" DevicePath \"\"" Oct 06 10:43:54 crc kubenswrapper[4824]: I1006 10:43:54.645430 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cfb03cbf-729e-4c5a-b1e5-d5e6ca02e497-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 06 10:43:54 crc kubenswrapper[4824]: I1006 10:43:54.645458 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2psc8\" (UniqueName: \"kubernetes.io/projected/cfb03cbf-729e-4c5a-b1e5-d5e6ca02e497-kube-api-access-2psc8\") on node \"crc\" DevicePath \"\"" Oct 06 10:43:54 crc kubenswrapper[4824]: I1006 10:43:54.951205 4824 generic.go:334] "Generic (PLEG): container finished" podID="cfb03cbf-729e-4c5a-b1e5-d5e6ca02e497" containerID="f2162519cdfe4d1e8ffcec6d522a6173492d372ddaeb5810278f0b2a47caf13d" exitCode=0 Oct 06 10:43:54 crc kubenswrapper[4824]: I1006 10:43:54.951264 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8vqd9" event={"ID":"cfb03cbf-729e-4c5a-b1e5-d5e6ca02e497","Type":"ContainerDied","Data":"f2162519cdfe4d1e8ffcec6d522a6173492d372ddaeb5810278f0b2a47caf13d"} Oct 06 10:43:54 crc kubenswrapper[4824]: I1006 10:43:54.953303 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8vqd9" event={"ID":"cfb03cbf-729e-4c5a-b1e5-d5e6ca02e497","Type":"ContainerDied","Data":"5725a8e272a9d38c1e4caa0aa9084c5c27d91602c1bdc36668838a9db8248162"} Oct 06 10:43:54 crc kubenswrapper[4824]: I1006 10:43:54.951291 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8vqd9" Oct 06 10:43:54 crc kubenswrapper[4824]: I1006 10:43:54.953350 4824 scope.go:117] "RemoveContainer" containerID="f2162519cdfe4d1e8ffcec6d522a6173492d372ddaeb5810278f0b2a47caf13d" Oct 06 10:43:54 crc kubenswrapper[4824]: I1006 10:43:54.986004 4824 scope.go:117] "RemoveContainer" containerID="2404ad3886244359893e8aac325397bd46270ef6506ad89f43408b2dfdd9b58b" Oct 06 10:43:54 crc kubenswrapper[4824]: I1006 10:43:54.991893 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-8vqd9"] Oct 06 10:43:54 crc kubenswrapper[4824]: I1006 10:43:54.999756 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-8vqd9"] Oct 06 10:43:55 crc kubenswrapper[4824]: I1006 10:43:55.011791 4824 scope.go:117] "RemoveContainer" containerID="a821817c7c09bde17a1072a065fec9fe9b26a9b9059ca387f12929056051d1d4" Oct 06 10:43:55 crc kubenswrapper[4824]: I1006 10:43:55.056700 4824 scope.go:117] "RemoveContainer" containerID="f2162519cdfe4d1e8ffcec6d522a6173492d372ddaeb5810278f0b2a47caf13d" Oct 06 10:43:55 crc kubenswrapper[4824]: E1006 10:43:55.057638 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f2162519cdfe4d1e8ffcec6d522a6173492d372ddaeb5810278f0b2a47caf13d\": container with ID starting with f2162519cdfe4d1e8ffcec6d522a6173492d372ddaeb5810278f0b2a47caf13d not found: ID does not exist" containerID="f2162519cdfe4d1e8ffcec6d522a6173492d372ddaeb5810278f0b2a47caf13d" Oct 06 10:43:55 crc kubenswrapper[4824]: I1006 10:43:55.057702 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f2162519cdfe4d1e8ffcec6d522a6173492d372ddaeb5810278f0b2a47caf13d"} err="failed to get container status \"f2162519cdfe4d1e8ffcec6d522a6173492d372ddaeb5810278f0b2a47caf13d\": rpc error: code = NotFound desc = could not find container \"f2162519cdfe4d1e8ffcec6d522a6173492d372ddaeb5810278f0b2a47caf13d\": container with ID starting with f2162519cdfe4d1e8ffcec6d522a6173492d372ddaeb5810278f0b2a47caf13d not found: ID does not exist" Oct 06 10:43:55 crc kubenswrapper[4824]: I1006 10:43:55.057739 4824 scope.go:117] "RemoveContainer" containerID="2404ad3886244359893e8aac325397bd46270ef6506ad89f43408b2dfdd9b58b" Oct 06 10:43:55 crc kubenswrapper[4824]: E1006 10:43:55.058292 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2404ad3886244359893e8aac325397bd46270ef6506ad89f43408b2dfdd9b58b\": container with ID starting with 2404ad3886244359893e8aac325397bd46270ef6506ad89f43408b2dfdd9b58b not found: ID does not exist" containerID="2404ad3886244359893e8aac325397bd46270ef6506ad89f43408b2dfdd9b58b" Oct 06 10:43:55 crc kubenswrapper[4824]: I1006 10:43:55.058312 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2404ad3886244359893e8aac325397bd46270ef6506ad89f43408b2dfdd9b58b"} err="failed to get container status \"2404ad3886244359893e8aac325397bd46270ef6506ad89f43408b2dfdd9b58b\": rpc error: code = NotFound desc = could not find container \"2404ad3886244359893e8aac325397bd46270ef6506ad89f43408b2dfdd9b58b\": container with ID starting with 2404ad3886244359893e8aac325397bd46270ef6506ad89f43408b2dfdd9b58b not found: ID does not exist" Oct 06 10:43:55 crc kubenswrapper[4824]: I1006 10:43:55.058327 4824 scope.go:117] "RemoveContainer" containerID="a821817c7c09bde17a1072a065fec9fe9b26a9b9059ca387f12929056051d1d4" Oct 06 10:43:55 crc kubenswrapper[4824]: E1006 10:43:55.059356 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a821817c7c09bde17a1072a065fec9fe9b26a9b9059ca387f12929056051d1d4\": container with ID starting with a821817c7c09bde17a1072a065fec9fe9b26a9b9059ca387f12929056051d1d4 not found: ID does not exist" containerID="a821817c7c09bde17a1072a065fec9fe9b26a9b9059ca387f12929056051d1d4" Oct 06 10:43:55 crc kubenswrapper[4824]: I1006 10:43:55.059396 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a821817c7c09bde17a1072a065fec9fe9b26a9b9059ca387f12929056051d1d4"} err="failed to get container status \"a821817c7c09bde17a1072a065fec9fe9b26a9b9059ca387f12929056051d1d4\": rpc error: code = NotFound desc = could not find container \"a821817c7c09bde17a1072a065fec9fe9b26a9b9059ca387f12929056051d1d4\": container with ID starting with a821817c7c09bde17a1072a065fec9fe9b26a9b9059ca387f12929056051d1d4 not found: ID does not exist" Oct 06 10:43:55 crc kubenswrapper[4824]: I1006 10:43:55.291913 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cfb03cbf-729e-4c5a-b1e5-d5e6ca02e497" path="/var/lib/kubelet/pods/cfb03cbf-729e-4c5a-b1e5-d5e6ca02e497/volumes" Oct 06 10:44:29 crc kubenswrapper[4824]: I1006 10:44:29.405032 4824 generic.go:334] "Generic (PLEG): container finished" podID="3a410146-5f2b-48a0-824f-ee424f685694" containerID="bab8163b6158aa354be67d6805f05751cf08f765ad57b2fc2f0afb9b002c8b16" exitCode=0 Oct 06 10:44:29 crc kubenswrapper[4824]: I1006 10:44:29.405209 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-4br4q" event={"ID":"3a410146-5f2b-48a0-824f-ee424f685694","Type":"ContainerDied","Data":"bab8163b6158aa354be67d6805f05751cf08f765ad57b2fc2f0afb9b002c8b16"} Oct 06 10:44:30 crc kubenswrapper[4824]: I1006 10:44:30.900723 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-4br4q" Oct 06 10:44:31 crc kubenswrapper[4824]: I1006 10:44:31.052235 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9qwbc\" (UniqueName: \"kubernetes.io/projected/3a410146-5f2b-48a0-824f-ee424f685694-kube-api-access-9qwbc\") pod \"3a410146-5f2b-48a0-824f-ee424f685694\" (UID: \"3a410146-5f2b-48a0-824f-ee424f685694\") " Oct 06 10:44:31 crc kubenswrapper[4824]: I1006 10:44:31.052326 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/3a410146-5f2b-48a0-824f-ee424f685694-ceilometer-compute-config-data-2\") pod \"3a410146-5f2b-48a0-824f-ee424f685694\" (UID: \"3a410146-5f2b-48a0-824f-ee424f685694\") " Oct 06 10:44:31 crc kubenswrapper[4824]: I1006 10:44:31.052389 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a410146-5f2b-48a0-824f-ee424f685694-telemetry-combined-ca-bundle\") pod \"3a410146-5f2b-48a0-824f-ee424f685694\" (UID: \"3a410146-5f2b-48a0-824f-ee424f685694\") " Oct 06 10:44:31 crc kubenswrapper[4824]: I1006 10:44:31.052603 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/3a410146-5f2b-48a0-824f-ee424f685694-ceilometer-compute-config-data-0\") pod \"3a410146-5f2b-48a0-824f-ee424f685694\" (UID: \"3a410146-5f2b-48a0-824f-ee424f685694\") " Oct 06 10:44:31 crc kubenswrapper[4824]: I1006 10:44:31.052676 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3a410146-5f2b-48a0-824f-ee424f685694-ssh-key\") pod \"3a410146-5f2b-48a0-824f-ee424f685694\" (UID: \"3a410146-5f2b-48a0-824f-ee424f685694\") " Oct 06 10:44:31 crc kubenswrapper[4824]: I1006 10:44:31.052738 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/3a410146-5f2b-48a0-824f-ee424f685694-ceilometer-compute-config-data-1\") pod \"3a410146-5f2b-48a0-824f-ee424f685694\" (UID: \"3a410146-5f2b-48a0-824f-ee424f685694\") " Oct 06 10:44:31 crc kubenswrapper[4824]: I1006 10:44:31.054168 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3a410146-5f2b-48a0-824f-ee424f685694-inventory\") pod \"3a410146-5f2b-48a0-824f-ee424f685694\" (UID: \"3a410146-5f2b-48a0-824f-ee424f685694\") " Oct 06 10:44:31 crc kubenswrapper[4824]: I1006 10:44:31.060173 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3a410146-5f2b-48a0-824f-ee424f685694-kube-api-access-9qwbc" (OuterVolumeSpecName: "kube-api-access-9qwbc") pod "3a410146-5f2b-48a0-824f-ee424f685694" (UID: "3a410146-5f2b-48a0-824f-ee424f685694"). InnerVolumeSpecName "kube-api-access-9qwbc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:44:31 crc kubenswrapper[4824]: I1006 10:44:31.060341 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a410146-5f2b-48a0-824f-ee424f685694-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "3a410146-5f2b-48a0-824f-ee424f685694" (UID: "3a410146-5f2b-48a0-824f-ee424f685694"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:44:31 crc kubenswrapper[4824]: I1006 10:44:31.085288 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a410146-5f2b-48a0-824f-ee424f685694-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "3a410146-5f2b-48a0-824f-ee424f685694" (UID: "3a410146-5f2b-48a0-824f-ee424f685694"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:44:31 crc kubenswrapper[4824]: I1006 10:44:31.099353 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a410146-5f2b-48a0-824f-ee424f685694-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "3a410146-5f2b-48a0-824f-ee424f685694" (UID: "3a410146-5f2b-48a0-824f-ee424f685694"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:44:31 crc kubenswrapper[4824]: I1006 10:44:31.101698 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a410146-5f2b-48a0-824f-ee424f685694-inventory" (OuterVolumeSpecName: "inventory") pod "3a410146-5f2b-48a0-824f-ee424f685694" (UID: "3a410146-5f2b-48a0-824f-ee424f685694"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:44:31 crc kubenswrapper[4824]: I1006 10:44:31.104156 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a410146-5f2b-48a0-824f-ee424f685694-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "3a410146-5f2b-48a0-824f-ee424f685694" (UID: "3a410146-5f2b-48a0-824f-ee424f685694"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:44:31 crc kubenswrapper[4824]: I1006 10:44:31.105820 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a410146-5f2b-48a0-824f-ee424f685694-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "3a410146-5f2b-48a0-824f-ee424f685694" (UID: "3a410146-5f2b-48a0-824f-ee424f685694"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:44:31 crc kubenswrapper[4824]: I1006 10:44:31.158233 4824 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a410146-5f2b-48a0-824f-ee424f685694-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 10:44:31 crc kubenswrapper[4824]: I1006 10:44:31.158288 4824 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/3a410146-5f2b-48a0-824f-ee424f685694-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Oct 06 10:44:31 crc kubenswrapper[4824]: I1006 10:44:31.158307 4824 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3a410146-5f2b-48a0-824f-ee424f685694-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 06 10:44:31 crc kubenswrapper[4824]: I1006 10:44:31.158323 4824 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/3a410146-5f2b-48a0-824f-ee424f685694-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Oct 06 10:44:31 crc kubenswrapper[4824]: I1006 10:44:31.158342 4824 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3a410146-5f2b-48a0-824f-ee424f685694-inventory\") on node \"crc\" DevicePath \"\"" Oct 06 10:44:31 crc kubenswrapper[4824]: I1006 10:44:31.158360 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9qwbc\" (UniqueName: \"kubernetes.io/projected/3a410146-5f2b-48a0-824f-ee424f685694-kube-api-access-9qwbc\") on node \"crc\" DevicePath \"\"" Oct 06 10:44:31 crc kubenswrapper[4824]: I1006 10:44:31.158376 4824 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/3a410146-5f2b-48a0-824f-ee424f685694-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Oct 06 10:44:31 crc kubenswrapper[4824]: I1006 10:44:31.430959 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-4br4q" event={"ID":"3a410146-5f2b-48a0-824f-ee424f685694","Type":"ContainerDied","Data":"5206ef25ab1fd2549f1cbe9f0b47dfb81bd3d8957f1235ff7819f5c4544c25a9"} Oct 06 10:44:31 crc kubenswrapper[4824]: I1006 10:44:31.431056 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5206ef25ab1fd2549f1cbe9f0b47dfb81bd3d8957f1235ff7819f5c4544c25a9" Oct 06 10:44:31 crc kubenswrapper[4824]: I1006 10:44:31.431144 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-4br4q" Oct 06 10:44:33 crc kubenswrapper[4824]: I1006 10:44:33.355950 4824 scope.go:117] "RemoveContainer" containerID="f9a25d0249f8e6be21f66182d479c02914adf09a341b637ed0d290d963ac4e45" Oct 06 10:44:33 crc kubenswrapper[4824]: I1006 10:44:33.395673 4824 scope.go:117] "RemoveContainer" containerID="753d23f3b3ebeea35c4477b4b54e63c5de117a619ff4f0102dfb1419c5bb7906" Oct 06 10:45:00 crc kubenswrapper[4824]: I1006 10:45:00.158994 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29329125-xxjmw"] Oct 06 10:45:00 crc kubenswrapper[4824]: E1006 10:45:00.160173 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ae7f4db-4a50-466c-a56b-f45403a18178" containerName="registry-server" Oct 06 10:45:00 crc kubenswrapper[4824]: I1006 10:45:00.160190 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ae7f4db-4a50-466c-a56b-f45403a18178" containerName="registry-server" Oct 06 10:45:00 crc kubenswrapper[4824]: E1006 10:45:00.160202 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cfb03cbf-729e-4c5a-b1e5-d5e6ca02e497" containerName="extract-utilities" Oct 06 10:45:00 crc kubenswrapper[4824]: I1006 10:45:00.160209 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="cfb03cbf-729e-4c5a-b1e5-d5e6ca02e497" containerName="extract-utilities" Oct 06 10:45:00 crc kubenswrapper[4824]: E1006 10:45:00.160222 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cfb03cbf-729e-4c5a-b1e5-d5e6ca02e497" containerName="extract-content" Oct 06 10:45:00 crc kubenswrapper[4824]: I1006 10:45:00.160228 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="cfb03cbf-729e-4c5a-b1e5-d5e6ca02e497" containerName="extract-content" Oct 06 10:45:00 crc kubenswrapper[4824]: E1006 10:45:00.160240 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a410146-5f2b-48a0-824f-ee424f685694" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Oct 06 10:45:00 crc kubenswrapper[4824]: I1006 10:45:00.160247 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a410146-5f2b-48a0-824f-ee424f685694" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Oct 06 10:45:00 crc kubenswrapper[4824]: E1006 10:45:00.160262 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cfb03cbf-729e-4c5a-b1e5-d5e6ca02e497" containerName="registry-server" Oct 06 10:45:00 crc kubenswrapper[4824]: I1006 10:45:00.160268 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="cfb03cbf-729e-4c5a-b1e5-d5e6ca02e497" containerName="registry-server" Oct 06 10:45:00 crc kubenswrapper[4824]: E1006 10:45:00.160284 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ae7f4db-4a50-466c-a56b-f45403a18178" containerName="extract-content" Oct 06 10:45:00 crc kubenswrapper[4824]: I1006 10:45:00.160290 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ae7f4db-4a50-466c-a56b-f45403a18178" containerName="extract-content" Oct 06 10:45:00 crc kubenswrapper[4824]: E1006 10:45:00.160308 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ae7f4db-4a50-466c-a56b-f45403a18178" containerName="extract-utilities" Oct 06 10:45:00 crc kubenswrapper[4824]: I1006 10:45:00.160314 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ae7f4db-4a50-466c-a56b-f45403a18178" containerName="extract-utilities" Oct 06 10:45:00 crc kubenswrapper[4824]: I1006 10:45:00.160542 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="cfb03cbf-729e-4c5a-b1e5-d5e6ca02e497" containerName="registry-server" Oct 06 10:45:00 crc kubenswrapper[4824]: I1006 10:45:00.160558 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a410146-5f2b-48a0-824f-ee424f685694" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Oct 06 10:45:00 crc kubenswrapper[4824]: I1006 10:45:00.160574 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ae7f4db-4a50-466c-a56b-f45403a18178" containerName="registry-server" Oct 06 10:45:00 crc kubenswrapper[4824]: I1006 10:45:00.161779 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29329125-xxjmw" Oct 06 10:45:00 crc kubenswrapper[4824]: I1006 10:45:00.164377 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 06 10:45:00 crc kubenswrapper[4824]: I1006 10:45:00.164497 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 06 10:45:00 crc kubenswrapper[4824]: I1006 10:45:00.182585 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29329125-xxjmw"] Oct 06 10:45:00 crc kubenswrapper[4824]: I1006 10:45:00.257174 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5c929685-b96b-4a59-b83c-78b5b5a1197e-secret-volume\") pod \"collect-profiles-29329125-xxjmw\" (UID: \"5c929685-b96b-4a59-b83c-78b5b5a1197e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329125-xxjmw" Oct 06 10:45:00 crc kubenswrapper[4824]: I1006 10:45:00.257241 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kfq67\" (UniqueName: \"kubernetes.io/projected/5c929685-b96b-4a59-b83c-78b5b5a1197e-kube-api-access-kfq67\") pod \"collect-profiles-29329125-xxjmw\" (UID: \"5c929685-b96b-4a59-b83c-78b5b5a1197e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329125-xxjmw" Oct 06 10:45:00 crc kubenswrapper[4824]: I1006 10:45:00.257387 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5c929685-b96b-4a59-b83c-78b5b5a1197e-config-volume\") pod \"collect-profiles-29329125-xxjmw\" (UID: \"5c929685-b96b-4a59-b83c-78b5b5a1197e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329125-xxjmw" Oct 06 10:45:00 crc kubenswrapper[4824]: I1006 10:45:00.359110 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5c929685-b96b-4a59-b83c-78b5b5a1197e-secret-volume\") pod \"collect-profiles-29329125-xxjmw\" (UID: \"5c929685-b96b-4a59-b83c-78b5b5a1197e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329125-xxjmw" Oct 06 10:45:00 crc kubenswrapper[4824]: I1006 10:45:00.359905 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kfq67\" (UniqueName: \"kubernetes.io/projected/5c929685-b96b-4a59-b83c-78b5b5a1197e-kube-api-access-kfq67\") pod \"collect-profiles-29329125-xxjmw\" (UID: \"5c929685-b96b-4a59-b83c-78b5b5a1197e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329125-xxjmw" Oct 06 10:45:00 crc kubenswrapper[4824]: I1006 10:45:00.361055 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5c929685-b96b-4a59-b83c-78b5b5a1197e-config-volume\") pod \"collect-profiles-29329125-xxjmw\" (UID: \"5c929685-b96b-4a59-b83c-78b5b5a1197e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329125-xxjmw" Oct 06 10:45:00 crc kubenswrapper[4824]: I1006 10:45:00.361998 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5c929685-b96b-4a59-b83c-78b5b5a1197e-config-volume\") pod \"collect-profiles-29329125-xxjmw\" (UID: \"5c929685-b96b-4a59-b83c-78b5b5a1197e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329125-xxjmw" Oct 06 10:45:00 crc kubenswrapper[4824]: I1006 10:45:00.370826 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5c929685-b96b-4a59-b83c-78b5b5a1197e-secret-volume\") pod \"collect-profiles-29329125-xxjmw\" (UID: \"5c929685-b96b-4a59-b83c-78b5b5a1197e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329125-xxjmw" Oct 06 10:45:00 crc kubenswrapper[4824]: I1006 10:45:00.378130 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kfq67\" (UniqueName: \"kubernetes.io/projected/5c929685-b96b-4a59-b83c-78b5b5a1197e-kube-api-access-kfq67\") pod \"collect-profiles-29329125-xxjmw\" (UID: \"5c929685-b96b-4a59-b83c-78b5b5a1197e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329125-xxjmw" Oct 06 10:45:00 crc kubenswrapper[4824]: I1006 10:45:00.494404 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29329125-xxjmw" Oct 06 10:45:00 crc kubenswrapper[4824]: I1006 10:45:00.969379 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29329125-xxjmw"] Oct 06 10:45:01 crc kubenswrapper[4824]: I1006 10:45:01.793808 4824 generic.go:334] "Generic (PLEG): container finished" podID="5c929685-b96b-4a59-b83c-78b5b5a1197e" containerID="7d1ec06a7c7d0bcb8eb0ef4b2777333be71187337cc94cf7b261b4bc43ab5b67" exitCode=0 Oct 06 10:45:01 crc kubenswrapper[4824]: I1006 10:45:01.794049 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29329125-xxjmw" event={"ID":"5c929685-b96b-4a59-b83c-78b5b5a1197e","Type":"ContainerDied","Data":"7d1ec06a7c7d0bcb8eb0ef4b2777333be71187337cc94cf7b261b4bc43ab5b67"} Oct 06 10:45:01 crc kubenswrapper[4824]: I1006 10:45:01.794288 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29329125-xxjmw" event={"ID":"5c929685-b96b-4a59-b83c-78b5b5a1197e","Type":"ContainerStarted","Data":"9a77d73196cc90973a6943cdba71f5c982169f20f52d48483cd75781acc8685f"} Oct 06 10:45:03 crc kubenswrapper[4824]: I1006 10:45:03.193044 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29329125-xxjmw" Oct 06 10:45:03 crc kubenswrapper[4824]: I1006 10:45:03.221322 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5c929685-b96b-4a59-b83c-78b5b5a1197e-config-volume\") pod \"5c929685-b96b-4a59-b83c-78b5b5a1197e\" (UID: \"5c929685-b96b-4a59-b83c-78b5b5a1197e\") " Oct 06 10:45:03 crc kubenswrapper[4824]: I1006 10:45:03.221804 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5c929685-b96b-4a59-b83c-78b5b5a1197e-secret-volume\") pod \"5c929685-b96b-4a59-b83c-78b5b5a1197e\" (UID: \"5c929685-b96b-4a59-b83c-78b5b5a1197e\") " Oct 06 10:45:03 crc kubenswrapper[4824]: I1006 10:45:03.221893 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfq67\" (UniqueName: \"kubernetes.io/projected/5c929685-b96b-4a59-b83c-78b5b5a1197e-kube-api-access-kfq67\") pod \"5c929685-b96b-4a59-b83c-78b5b5a1197e\" (UID: \"5c929685-b96b-4a59-b83c-78b5b5a1197e\") " Oct 06 10:45:03 crc kubenswrapper[4824]: I1006 10:45:03.222591 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5c929685-b96b-4a59-b83c-78b5b5a1197e-config-volume" (OuterVolumeSpecName: "config-volume") pod "5c929685-b96b-4a59-b83c-78b5b5a1197e" (UID: "5c929685-b96b-4a59-b83c-78b5b5a1197e"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:45:03 crc kubenswrapper[4824]: I1006 10:45:03.232363 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5c929685-b96b-4a59-b83c-78b5b5a1197e-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "5c929685-b96b-4a59-b83c-78b5b5a1197e" (UID: "5c929685-b96b-4a59-b83c-78b5b5a1197e"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:45:03 crc kubenswrapper[4824]: I1006 10:45:03.234141 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5c929685-b96b-4a59-b83c-78b5b5a1197e-kube-api-access-kfq67" (OuterVolumeSpecName: "kube-api-access-kfq67") pod "5c929685-b96b-4a59-b83c-78b5b5a1197e" (UID: "5c929685-b96b-4a59-b83c-78b5b5a1197e"). InnerVolumeSpecName "kube-api-access-kfq67". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:45:03 crc kubenswrapper[4824]: I1006 10:45:03.324191 4824 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5c929685-b96b-4a59-b83c-78b5b5a1197e-config-volume\") on node \"crc\" DevicePath \"\"" Oct 06 10:45:03 crc kubenswrapper[4824]: I1006 10:45:03.324241 4824 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5c929685-b96b-4a59-b83c-78b5b5a1197e-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 06 10:45:03 crc kubenswrapper[4824]: I1006 10:45:03.324259 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfq67\" (UniqueName: \"kubernetes.io/projected/5c929685-b96b-4a59-b83c-78b5b5a1197e-kube-api-access-kfq67\") on node \"crc\" DevicePath \"\"" Oct 06 10:45:03 crc kubenswrapper[4824]: I1006 10:45:03.814381 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29329125-xxjmw" event={"ID":"5c929685-b96b-4a59-b83c-78b5b5a1197e","Type":"ContainerDied","Data":"9a77d73196cc90973a6943cdba71f5c982169f20f52d48483cd75781acc8685f"} Oct 06 10:45:03 crc kubenswrapper[4824]: I1006 10:45:03.814436 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9a77d73196cc90973a6943cdba71f5c982169f20f52d48483cd75781acc8685f" Oct 06 10:45:03 crc kubenswrapper[4824]: I1006 10:45:03.814470 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29329125-xxjmw" Oct 06 10:45:04 crc kubenswrapper[4824]: I1006 10:45:04.283769 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29329080-9kpdv"] Oct 06 10:45:04 crc kubenswrapper[4824]: I1006 10:45:04.292412 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29329080-9kpdv"] Oct 06 10:45:05 crc kubenswrapper[4824]: I1006 10:45:05.288946 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="26536a6b-3639-46ad-b024-5e21b7118fe7" path="/var/lib/kubelet/pods/26536a6b-3639-46ad-b024-5e21b7118fe7/volumes" Oct 06 10:45:21 crc kubenswrapper[4824]: I1006 10:45:21.123243 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest"] Oct 06 10:45:21 crc kubenswrapper[4824]: E1006 10:45:21.124699 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c929685-b96b-4a59-b83c-78b5b5a1197e" containerName="collect-profiles" Oct 06 10:45:21 crc kubenswrapper[4824]: I1006 10:45:21.124732 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c929685-b96b-4a59-b83c-78b5b5a1197e" containerName="collect-profiles" Oct 06 10:45:21 crc kubenswrapper[4824]: I1006 10:45:21.125157 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c929685-b96b-4a59-b83c-78b5b5a1197e" containerName="collect-profiles" Oct 06 10:45:21 crc kubenswrapper[4824]: I1006 10:45:21.126467 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Oct 06 10:45:21 crc kubenswrapper[4824]: I1006 10:45:21.130046 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Oct 06 10:45:21 crc kubenswrapper[4824]: I1006 10:45:21.130956 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-5bmsf" Oct 06 10:45:21 crc kubenswrapper[4824]: I1006 10:45:21.130956 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Oct 06 10:45:21 crc kubenswrapper[4824]: I1006 10:45:21.131062 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Oct 06 10:45:21 crc kubenswrapper[4824]: I1006 10:45:21.153311 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/7fd677dd-afa0-45a9-8198-ea78e02f4fb7-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"7fd677dd-afa0-45a9-8198-ea78e02f4fb7\") " pod="openstack/tempest-tests-tempest" Oct 06 10:45:21 crc kubenswrapper[4824]: I1006 10:45:21.153577 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7fd677dd-afa0-45a9-8198-ea78e02f4fb7-config-data\") pod \"tempest-tests-tempest\" (UID: \"7fd677dd-afa0-45a9-8198-ea78e02f4fb7\") " pod="openstack/tempest-tests-tempest" Oct 06 10:45:21 crc kubenswrapper[4824]: I1006 10:45:21.153666 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/7fd677dd-afa0-45a9-8198-ea78e02f4fb7-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"7fd677dd-afa0-45a9-8198-ea78e02f4fb7\") " pod="openstack/tempest-tests-tempest" Oct 06 10:45:21 crc kubenswrapper[4824]: I1006 10:45:21.155793 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Oct 06 10:45:21 crc kubenswrapper[4824]: I1006 10:45:21.255949 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/7fd677dd-afa0-45a9-8198-ea78e02f4fb7-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"7fd677dd-afa0-45a9-8198-ea78e02f4fb7\") " pod="openstack/tempest-tests-tempest" Oct 06 10:45:21 crc kubenswrapper[4824]: I1006 10:45:21.256184 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7fd677dd-afa0-45a9-8198-ea78e02f4fb7-config-data\") pod \"tempest-tests-tempest\" (UID: \"7fd677dd-afa0-45a9-8198-ea78e02f4fb7\") " pod="openstack/tempest-tests-tempest" Oct 06 10:45:21 crc kubenswrapper[4824]: I1006 10:45:21.256250 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/7fd677dd-afa0-45a9-8198-ea78e02f4fb7-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"7fd677dd-afa0-45a9-8198-ea78e02f4fb7\") " pod="openstack/tempest-tests-tempest" Oct 06 10:45:21 crc kubenswrapper[4824]: I1006 10:45:21.256306 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/7fd677dd-afa0-45a9-8198-ea78e02f4fb7-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"7fd677dd-afa0-45a9-8198-ea78e02f4fb7\") " pod="openstack/tempest-tests-tempest" Oct 06 10:45:21 crc kubenswrapper[4824]: I1006 10:45:21.256358 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7fd677dd-afa0-45a9-8198-ea78e02f4fb7-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"7fd677dd-afa0-45a9-8198-ea78e02f4fb7\") " pod="openstack/tempest-tests-tempest" Oct 06 10:45:21 crc kubenswrapper[4824]: I1006 10:45:21.256474 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8l7lg\" (UniqueName: \"kubernetes.io/projected/7fd677dd-afa0-45a9-8198-ea78e02f4fb7-kube-api-access-8l7lg\") pod \"tempest-tests-tempest\" (UID: \"7fd677dd-afa0-45a9-8198-ea78e02f4fb7\") " pod="openstack/tempest-tests-tempest" Oct 06 10:45:21 crc kubenswrapper[4824]: I1006 10:45:21.256725 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"tempest-tests-tempest\" (UID: \"7fd677dd-afa0-45a9-8198-ea78e02f4fb7\") " pod="openstack/tempest-tests-tempest" Oct 06 10:45:21 crc kubenswrapper[4824]: I1006 10:45:21.256916 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/7fd677dd-afa0-45a9-8198-ea78e02f4fb7-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"7fd677dd-afa0-45a9-8198-ea78e02f4fb7\") " pod="openstack/tempest-tests-tempest" Oct 06 10:45:21 crc kubenswrapper[4824]: I1006 10:45:21.257046 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/7fd677dd-afa0-45a9-8198-ea78e02f4fb7-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"7fd677dd-afa0-45a9-8198-ea78e02f4fb7\") " pod="openstack/tempest-tests-tempest" Oct 06 10:45:21 crc kubenswrapper[4824]: I1006 10:45:21.258659 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/7fd677dd-afa0-45a9-8198-ea78e02f4fb7-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"7fd677dd-afa0-45a9-8198-ea78e02f4fb7\") " pod="openstack/tempest-tests-tempest" Oct 06 10:45:21 crc kubenswrapper[4824]: I1006 10:45:21.260973 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7fd677dd-afa0-45a9-8198-ea78e02f4fb7-config-data\") pod \"tempest-tests-tempest\" (UID: \"7fd677dd-afa0-45a9-8198-ea78e02f4fb7\") " pod="openstack/tempest-tests-tempest" Oct 06 10:45:21 crc kubenswrapper[4824]: I1006 10:45:21.275465 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/7fd677dd-afa0-45a9-8198-ea78e02f4fb7-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"7fd677dd-afa0-45a9-8198-ea78e02f4fb7\") " pod="openstack/tempest-tests-tempest" Oct 06 10:45:21 crc kubenswrapper[4824]: I1006 10:45:21.359504 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8l7lg\" (UniqueName: \"kubernetes.io/projected/7fd677dd-afa0-45a9-8198-ea78e02f4fb7-kube-api-access-8l7lg\") pod \"tempest-tests-tempest\" (UID: \"7fd677dd-afa0-45a9-8198-ea78e02f4fb7\") " pod="openstack/tempest-tests-tempest" Oct 06 10:45:21 crc kubenswrapper[4824]: I1006 10:45:21.360056 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"tempest-tests-tempest\" (UID: \"7fd677dd-afa0-45a9-8198-ea78e02f4fb7\") " pod="openstack/tempest-tests-tempest" Oct 06 10:45:21 crc kubenswrapper[4824]: I1006 10:45:21.360262 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/7fd677dd-afa0-45a9-8198-ea78e02f4fb7-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"7fd677dd-afa0-45a9-8198-ea78e02f4fb7\") " pod="openstack/tempest-tests-tempest" Oct 06 10:45:21 crc kubenswrapper[4824]: I1006 10:45:21.360502 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/7fd677dd-afa0-45a9-8198-ea78e02f4fb7-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"7fd677dd-afa0-45a9-8198-ea78e02f4fb7\") " pod="openstack/tempest-tests-tempest" Oct 06 10:45:21 crc kubenswrapper[4824]: I1006 10:45:21.360740 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/7fd677dd-afa0-45a9-8198-ea78e02f4fb7-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"7fd677dd-afa0-45a9-8198-ea78e02f4fb7\") " pod="openstack/tempest-tests-tempest" Oct 06 10:45:21 crc kubenswrapper[4824]: I1006 10:45:21.360911 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/7fd677dd-afa0-45a9-8198-ea78e02f4fb7-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"7fd677dd-afa0-45a9-8198-ea78e02f4fb7\") " pod="openstack/tempest-tests-tempest" Oct 06 10:45:21 crc kubenswrapper[4824]: I1006 10:45:21.361057 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/7fd677dd-afa0-45a9-8198-ea78e02f4fb7-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"7fd677dd-afa0-45a9-8198-ea78e02f4fb7\") " pod="openstack/tempest-tests-tempest" Oct 06 10:45:21 crc kubenswrapper[4824]: I1006 10:45:21.360545 4824 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"tempest-tests-tempest\" (UID: \"7fd677dd-afa0-45a9-8198-ea78e02f4fb7\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/tempest-tests-tempest" Oct 06 10:45:21 crc kubenswrapper[4824]: I1006 10:45:21.361217 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7fd677dd-afa0-45a9-8198-ea78e02f4fb7-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"7fd677dd-afa0-45a9-8198-ea78e02f4fb7\") " pod="openstack/tempest-tests-tempest" Oct 06 10:45:21 crc kubenswrapper[4824]: I1006 10:45:21.366693 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/7fd677dd-afa0-45a9-8198-ea78e02f4fb7-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"7fd677dd-afa0-45a9-8198-ea78e02f4fb7\") " pod="openstack/tempest-tests-tempest" Oct 06 10:45:21 crc kubenswrapper[4824]: I1006 10:45:21.367618 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7fd677dd-afa0-45a9-8198-ea78e02f4fb7-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"7fd677dd-afa0-45a9-8198-ea78e02f4fb7\") " pod="openstack/tempest-tests-tempest" Oct 06 10:45:21 crc kubenswrapper[4824]: I1006 10:45:21.382481 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8l7lg\" (UniqueName: \"kubernetes.io/projected/7fd677dd-afa0-45a9-8198-ea78e02f4fb7-kube-api-access-8l7lg\") pod \"tempest-tests-tempest\" (UID: \"7fd677dd-afa0-45a9-8198-ea78e02f4fb7\") " pod="openstack/tempest-tests-tempest" Oct 06 10:45:21 crc kubenswrapper[4824]: I1006 10:45:21.404402 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"tempest-tests-tempest\" (UID: \"7fd677dd-afa0-45a9-8198-ea78e02f4fb7\") " pod="openstack/tempest-tests-tempest" Oct 06 10:45:21 crc kubenswrapper[4824]: I1006 10:45:21.484667 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Oct 06 10:45:21 crc kubenswrapper[4824]: I1006 10:45:21.982730 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Oct 06 10:45:21 crc kubenswrapper[4824]: W1006 10:45:21.985812 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7fd677dd_afa0_45a9_8198_ea78e02f4fb7.slice/crio-3ac1dd564fcdc44b871406cb8ec2680f9e87759df1fc6c97e05f085c2b189d5d WatchSource:0}: Error finding container 3ac1dd564fcdc44b871406cb8ec2680f9e87759df1fc6c97e05f085c2b189d5d: Status 404 returned error can't find the container with id 3ac1dd564fcdc44b871406cb8ec2680f9e87759df1fc6c97e05f085c2b189d5d Oct 06 10:45:22 crc kubenswrapper[4824]: I1006 10:45:22.036251 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"7fd677dd-afa0-45a9-8198-ea78e02f4fb7","Type":"ContainerStarted","Data":"3ac1dd564fcdc44b871406cb8ec2680f9e87759df1fc6c97e05f085c2b189d5d"} Oct 06 10:45:33 crc kubenswrapper[4824]: I1006 10:45:33.493772 4824 scope.go:117] "RemoveContainer" containerID="04591b6d2b8e78446287d500ea05b4d620e1226e9a5b376c57340ea2b03971f9" Oct 06 10:45:43 crc kubenswrapper[4824]: I1006 10:45:43.921142 4824 patch_prober.go:28] interesting pod/machine-config-daemon-khgzw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 10:45:43 crc kubenswrapper[4824]: I1006 10:45:43.921746 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 10:45:50 crc kubenswrapper[4824]: I1006 10:45:50.613138 4824 scope.go:117] "RemoveContainer" containerID="7de21155c716356f5c5d4f5a73ad154566803420cb9980a0cc1706e52f07e482" Oct 06 10:45:50 crc kubenswrapper[4824]: E1006 10:45:50.647146 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified" Oct 06 10:45:50 crc kubenswrapper[4824]: E1006 10:45:50.647422 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:tempest-tests-tempest-tests-runner,Image:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/tempest,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/test_operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/tempest/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/var/lib/tempest/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/var/lib/tempest/id_ecdsa,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-8l7lg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42480,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42480,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-custom-data-s0,},Optional:nil,},SecretRef:nil,},EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-env-vars-s0,},Optional:nil,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod tempest-tests-tempest_openstack(7fd677dd-afa0-45a9-8198-ea78e02f4fb7): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 06 10:45:50 crc kubenswrapper[4824]: E1006 10:45:50.648858 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/tempest-tests-tempest" podUID="7fd677dd-afa0-45a9-8198-ea78e02f4fb7" Oct 06 10:45:51 crc kubenswrapper[4824]: E1006 10:45:51.491567 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified\\\"\"" pod="openstack/tempest-tests-tempest" podUID="7fd677dd-afa0-45a9-8198-ea78e02f4fb7" Oct 06 10:46:04 crc kubenswrapper[4824]: I1006 10:46:04.646564 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"7fd677dd-afa0-45a9-8198-ea78e02f4fb7","Type":"ContainerStarted","Data":"97e83913e74175b39ebaa5d81693fba451d2731ce38d61aae9ded9fa6eef80d9"} Oct 06 10:46:04 crc kubenswrapper[4824]: I1006 10:46:04.689046 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest" podStartSLOduration=3.672307061 podStartE2EDuration="44.689012183s" podCreationTimestamp="2025-10-06 10:45:20 +0000 UTC" firstStartedPulling="2025-10-06 10:45:21.989898485 +0000 UTC m=+2831.354321376" lastFinishedPulling="2025-10-06 10:46:03.006603637 +0000 UTC m=+2872.371026498" observedRunningTime="2025-10-06 10:46:04.675287909 +0000 UTC m=+2874.039710790" watchObservedRunningTime="2025-10-06 10:46:04.689012183 +0000 UTC m=+2874.053435084" Oct 06 10:46:13 crc kubenswrapper[4824]: I1006 10:46:13.915299 4824 patch_prober.go:28] interesting pod/machine-config-daemon-khgzw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 10:46:13 crc kubenswrapper[4824]: I1006 10:46:13.915885 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 10:46:43 crc kubenswrapper[4824]: I1006 10:46:43.915403 4824 patch_prober.go:28] interesting pod/machine-config-daemon-khgzw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 10:46:43 crc kubenswrapper[4824]: I1006 10:46:43.916073 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 10:46:43 crc kubenswrapper[4824]: I1006 10:46:43.916130 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" Oct 06 10:46:43 crc kubenswrapper[4824]: I1006 10:46:43.917024 4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"db15accdeb46a9039be894301153ff2ab7709ec808de959d20f2a048678d55cc"} pod="openshift-machine-config-operator/machine-config-daemon-khgzw" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 06 10:46:43 crc kubenswrapper[4824]: I1006 10:46:43.917085 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" containerName="machine-config-daemon" containerID="cri-o://db15accdeb46a9039be894301153ff2ab7709ec808de959d20f2a048678d55cc" gracePeriod=600 Oct 06 10:46:44 crc kubenswrapper[4824]: I1006 10:46:44.145942 4824 generic.go:334] "Generic (PLEG): container finished" podID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" containerID="db15accdeb46a9039be894301153ff2ab7709ec808de959d20f2a048678d55cc" exitCode=0 Oct 06 10:46:44 crc kubenswrapper[4824]: I1006 10:46:44.146015 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" event={"ID":"e1140c8c-93f1-4fce-8c06-a69561ff8a6d","Type":"ContainerDied","Data":"db15accdeb46a9039be894301153ff2ab7709ec808de959d20f2a048678d55cc"} Oct 06 10:46:44 crc kubenswrapper[4824]: I1006 10:46:44.146083 4824 scope.go:117] "RemoveContainer" containerID="defbb9e5d617c326a5dda4c68f7311eec28a02ec245ca3527fbdf00979129a82" Oct 06 10:46:45 crc kubenswrapper[4824]: I1006 10:46:45.170631 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" event={"ID":"e1140c8c-93f1-4fce-8c06-a69561ff8a6d","Type":"ContainerStarted","Data":"bbb34b884429016efb41d11fc1bb14bdd15d27ec1a282254f5cbaae186b1d1ef"} Oct 06 10:47:17 crc kubenswrapper[4824]: I1006 10:47:17.021444 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-kzxk8"] Oct 06 10:47:17 crc kubenswrapper[4824]: I1006 10:47:17.026291 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kzxk8" Oct 06 10:47:17 crc kubenswrapper[4824]: I1006 10:47:17.036123 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-kzxk8"] Oct 06 10:47:17 crc kubenswrapper[4824]: I1006 10:47:17.225429 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/202562de-e495-4708-8608-a857b9135578-utilities\") pod \"redhat-operators-kzxk8\" (UID: \"202562de-e495-4708-8608-a857b9135578\") " pod="openshift-marketplace/redhat-operators-kzxk8" Oct 06 10:47:17 crc kubenswrapper[4824]: I1006 10:47:17.225513 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/202562de-e495-4708-8608-a857b9135578-catalog-content\") pod \"redhat-operators-kzxk8\" (UID: \"202562de-e495-4708-8608-a857b9135578\") " pod="openshift-marketplace/redhat-operators-kzxk8" Oct 06 10:47:17 crc kubenswrapper[4824]: I1006 10:47:17.225900 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f88f7\" (UniqueName: \"kubernetes.io/projected/202562de-e495-4708-8608-a857b9135578-kube-api-access-f88f7\") pod \"redhat-operators-kzxk8\" (UID: \"202562de-e495-4708-8608-a857b9135578\") " pod="openshift-marketplace/redhat-operators-kzxk8" Oct 06 10:47:17 crc kubenswrapper[4824]: I1006 10:47:17.328549 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/202562de-e495-4708-8608-a857b9135578-catalog-content\") pod \"redhat-operators-kzxk8\" (UID: \"202562de-e495-4708-8608-a857b9135578\") " pod="openshift-marketplace/redhat-operators-kzxk8" Oct 06 10:47:17 crc kubenswrapper[4824]: I1006 10:47:17.328701 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f88f7\" (UniqueName: \"kubernetes.io/projected/202562de-e495-4708-8608-a857b9135578-kube-api-access-f88f7\") pod \"redhat-operators-kzxk8\" (UID: \"202562de-e495-4708-8608-a857b9135578\") " pod="openshift-marketplace/redhat-operators-kzxk8" Oct 06 10:47:17 crc kubenswrapper[4824]: I1006 10:47:17.328863 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/202562de-e495-4708-8608-a857b9135578-utilities\") pod \"redhat-operators-kzxk8\" (UID: \"202562de-e495-4708-8608-a857b9135578\") " pod="openshift-marketplace/redhat-operators-kzxk8" Oct 06 10:47:17 crc kubenswrapper[4824]: I1006 10:47:17.329136 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/202562de-e495-4708-8608-a857b9135578-catalog-content\") pod \"redhat-operators-kzxk8\" (UID: \"202562de-e495-4708-8608-a857b9135578\") " pod="openshift-marketplace/redhat-operators-kzxk8" Oct 06 10:47:17 crc kubenswrapper[4824]: I1006 10:47:17.329496 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/202562de-e495-4708-8608-a857b9135578-utilities\") pod \"redhat-operators-kzxk8\" (UID: \"202562de-e495-4708-8608-a857b9135578\") " pod="openshift-marketplace/redhat-operators-kzxk8" Oct 06 10:47:17 crc kubenswrapper[4824]: I1006 10:47:17.363211 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f88f7\" (UniqueName: \"kubernetes.io/projected/202562de-e495-4708-8608-a857b9135578-kube-api-access-f88f7\") pod \"redhat-operators-kzxk8\" (UID: \"202562de-e495-4708-8608-a857b9135578\") " pod="openshift-marketplace/redhat-operators-kzxk8" Oct 06 10:47:17 crc kubenswrapper[4824]: I1006 10:47:17.659153 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kzxk8" Oct 06 10:47:18 crc kubenswrapper[4824]: I1006 10:47:18.205301 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-kzxk8"] Oct 06 10:47:18 crc kubenswrapper[4824]: I1006 10:47:18.623116 4824 generic.go:334] "Generic (PLEG): container finished" podID="202562de-e495-4708-8608-a857b9135578" containerID="179d16a7ce7922c7c3ac36786ee43cacaa7c53bb09b6e94cd8818ade0c86e582" exitCode=0 Oct 06 10:47:18 crc kubenswrapper[4824]: I1006 10:47:18.623254 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kzxk8" event={"ID":"202562de-e495-4708-8608-a857b9135578","Type":"ContainerDied","Data":"179d16a7ce7922c7c3ac36786ee43cacaa7c53bb09b6e94cd8818ade0c86e582"} Oct 06 10:47:18 crc kubenswrapper[4824]: I1006 10:47:18.623706 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kzxk8" event={"ID":"202562de-e495-4708-8608-a857b9135578","Type":"ContainerStarted","Data":"3bdde4389d02870ec3cc2a6bda437213f8e99bc800c5c0885a7450879285ebc5"} Oct 06 10:47:20 crc kubenswrapper[4824]: I1006 10:47:20.650565 4824 generic.go:334] "Generic (PLEG): container finished" podID="202562de-e495-4708-8608-a857b9135578" containerID="1a3c71d6e055511bd0a7546796ad3e486a2d26993c5912c2166d71299f38aeb2" exitCode=0 Oct 06 10:47:20 crc kubenswrapper[4824]: I1006 10:47:20.650627 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kzxk8" event={"ID":"202562de-e495-4708-8608-a857b9135578","Type":"ContainerDied","Data":"1a3c71d6e055511bd0a7546796ad3e486a2d26993c5912c2166d71299f38aeb2"} Oct 06 10:47:22 crc kubenswrapper[4824]: I1006 10:47:22.675090 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kzxk8" event={"ID":"202562de-e495-4708-8608-a857b9135578","Type":"ContainerStarted","Data":"eb026b7d15f27aa72183b067e6f6a8e4673a8d6812206ffa47ee1575307fd91b"} Oct 06 10:47:22 crc kubenswrapper[4824]: I1006 10:47:22.702540 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-kzxk8" podStartSLOduration=3.833257288 podStartE2EDuration="6.702512353s" podCreationTimestamp="2025-10-06 10:47:16 +0000 UTC" firstStartedPulling="2025-10-06 10:47:18.626354589 +0000 UTC m=+2947.990777450" lastFinishedPulling="2025-10-06 10:47:21.495609654 +0000 UTC m=+2950.860032515" observedRunningTime="2025-10-06 10:47:22.700676867 +0000 UTC m=+2952.065099738" watchObservedRunningTime="2025-10-06 10:47:22.702512353 +0000 UTC m=+2952.066935224" Oct 06 10:47:27 crc kubenswrapper[4824]: I1006 10:47:27.660202 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-kzxk8" Oct 06 10:47:27 crc kubenswrapper[4824]: I1006 10:47:27.661388 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-kzxk8" Oct 06 10:47:27 crc kubenswrapper[4824]: I1006 10:47:27.725593 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-kzxk8" Oct 06 10:47:27 crc kubenswrapper[4824]: I1006 10:47:27.844113 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-kzxk8" Oct 06 10:47:27 crc kubenswrapper[4824]: I1006 10:47:27.970457 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-kzxk8"] Oct 06 10:47:29 crc kubenswrapper[4824]: I1006 10:47:29.772163 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-kzxk8" podUID="202562de-e495-4708-8608-a857b9135578" containerName="registry-server" containerID="cri-o://eb026b7d15f27aa72183b067e6f6a8e4673a8d6812206ffa47ee1575307fd91b" gracePeriod=2 Oct 06 10:47:30 crc kubenswrapper[4824]: I1006 10:47:30.329100 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kzxk8" Oct 06 10:47:30 crc kubenswrapper[4824]: I1006 10:47:30.513534 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/202562de-e495-4708-8608-a857b9135578-utilities\") pod \"202562de-e495-4708-8608-a857b9135578\" (UID: \"202562de-e495-4708-8608-a857b9135578\") " Oct 06 10:47:30 crc kubenswrapper[4824]: I1006 10:47:30.514118 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/202562de-e495-4708-8608-a857b9135578-catalog-content\") pod \"202562de-e495-4708-8608-a857b9135578\" (UID: \"202562de-e495-4708-8608-a857b9135578\") " Oct 06 10:47:30 crc kubenswrapper[4824]: I1006 10:47:30.516409 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f88f7\" (UniqueName: \"kubernetes.io/projected/202562de-e495-4708-8608-a857b9135578-kube-api-access-f88f7\") pod \"202562de-e495-4708-8608-a857b9135578\" (UID: \"202562de-e495-4708-8608-a857b9135578\") " Oct 06 10:47:30 crc kubenswrapper[4824]: I1006 10:47:30.515605 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/202562de-e495-4708-8608-a857b9135578-utilities" (OuterVolumeSpecName: "utilities") pod "202562de-e495-4708-8608-a857b9135578" (UID: "202562de-e495-4708-8608-a857b9135578"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:47:30 crc kubenswrapper[4824]: I1006 10:47:30.517333 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/202562de-e495-4708-8608-a857b9135578-utilities\") on node \"crc\" DevicePath \"\"" Oct 06 10:47:30 crc kubenswrapper[4824]: I1006 10:47:30.525740 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/202562de-e495-4708-8608-a857b9135578-kube-api-access-f88f7" (OuterVolumeSpecName: "kube-api-access-f88f7") pod "202562de-e495-4708-8608-a857b9135578" (UID: "202562de-e495-4708-8608-a857b9135578"). InnerVolumeSpecName "kube-api-access-f88f7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:47:30 crc kubenswrapper[4824]: I1006 10:47:30.607648 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/202562de-e495-4708-8608-a857b9135578-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "202562de-e495-4708-8608-a857b9135578" (UID: "202562de-e495-4708-8608-a857b9135578"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:47:30 crc kubenswrapper[4824]: I1006 10:47:30.619702 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/202562de-e495-4708-8608-a857b9135578-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 06 10:47:30 crc kubenswrapper[4824]: I1006 10:47:30.619767 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f88f7\" (UniqueName: \"kubernetes.io/projected/202562de-e495-4708-8608-a857b9135578-kube-api-access-f88f7\") on node \"crc\" DevicePath \"\"" Oct 06 10:47:30 crc kubenswrapper[4824]: I1006 10:47:30.790274 4824 generic.go:334] "Generic (PLEG): container finished" podID="202562de-e495-4708-8608-a857b9135578" containerID="eb026b7d15f27aa72183b067e6f6a8e4673a8d6812206ffa47ee1575307fd91b" exitCode=0 Oct 06 10:47:30 crc kubenswrapper[4824]: I1006 10:47:30.790337 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kzxk8" event={"ID":"202562de-e495-4708-8608-a857b9135578","Type":"ContainerDied","Data":"eb026b7d15f27aa72183b067e6f6a8e4673a8d6812206ffa47ee1575307fd91b"} Oct 06 10:47:30 crc kubenswrapper[4824]: I1006 10:47:30.790379 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kzxk8" event={"ID":"202562de-e495-4708-8608-a857b9135578","Type":"ContainerDied","Data":"3bdde4389d02870ec3cc2a6bda437213f8e99bc800c5c0885a7450879285ebc5"} Oct 06 10:47:30 crc kubenswrapper[4824]: I1006 10:47:30.790407 4824 scope.go:117] "RemoveContainer" containerID="eb026b7d15f27aa72183b067e6f6a8e4673a8d6812206ffa47ee1575307fd91b" Oct 06 10:47:30 crc kubenswrapper[4824]: I1006 10:47:30.790477 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kzxk8" Oct 06 10:47:30 crc kubenswrapper[4824]: I1006 10:47:30.838245 4824 scope.go:117] "RemoveContainer" containerID="1a3c71d6e055511bd0a7546796ad3e486a2d26993c5912c2166d71299f38aeb2" Oct 06 10:47:30 crc kubenswrapper[4824]: I1006 10:47:30.838824 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-kzxk8"] Oct 06 10:47:30 crc kubenswrapper[4824]: I1006 10:47:30.849472 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-kzxk8"] Oct 06 10:47:30 crc kubenswrapper[4824]: I1006 10:47:30.867047 4824 scope.go:117] "RemoveContainer" containerID="179d16a7ce7922c7c3ac36786ee43cacaa7c53bb09b6e94cd8818ade0c86e582" Oct 06 10:47:30 crc kubenswrapper[4824]: I1006 10:47:30.933265 4824 scope.go:117] "RemoveContainer" containerID="eb026b7d15f27aa72183b067e6f6a8e4673a8d6812206ffa47ee1575307fd91b" Oct 06 10:47:30 crc kubenswrapper[4824]: E1006 10:47:30.933964 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eb026b7d15f27aa72183b067e6f6a8e4673a8d6812206ffa47ee1575307fd91b\": container with ID starting with eb026b7d15f27aa72183b067e6f6a8e4673a8d6812206ffa47ee1575307fd91b not found: ID does not exist" containerID="eb026b7d15f27aa72183b067e6f6a8e4673a8d6812206ffa47ee1575307fd91b" Oct 06 10:47:30 crc kubenswrapper[4824]: I1006 10:47:30.934030 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eb026b7d15f27aa72183b067e6f6a8e4673a8d6812206ffa47ee1575307fd91b"} err="failed to get container status \"eb026b7d15f27aa72183b067e6f6a8e4673a8d6812206ffa47ee1575307fd91b\": rpc error: code = NotFound desc = could not find container \"eb026b7d15f27aa72183b067e6f6a8e4673a8d6812206ffa47ee1575307fd91b\": container with ID starting with eb026b7d15f27aa72183b067e6f6a8e4673a8d6812206ffa47ee1575307fd91b not found: ID does not exist" Oct 06 10:47:30 crc kubenswrapper[4824]: I1006 10:47:30.934059 4824 scope.go:117] "RemoveContainer" containerID="1a3c71d6e055511bd0a7546796ad3e486a2d26993c5912c2166d71299f38aeb2" Oct 06 10:47:30 crc kubenswrapper[4824]: E1006 10:47:30.934538 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1a3c71d6e055511bd0a7546796ad3e486a2d26993c5912c2166d71299f38aeb2\": container with ID starting with 1a3c71d6e055511bd0a7546796ad3e486a2d26993c5912c2166d71299f38aeb2 not found: ID does not exist" containerID="1a3c71d6e055511bd0a7546796ad3e486a2d26993c5912c2166d71299f38aeb2" Oct 06 10:47:30 crc kubenswrapper[4824]: I1006 10:47:30.934617 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1a3c71d6e055511bd0a7546796ad3e486a2d26993c5912c2166d71299f38aeb2"} err="failed to get container status \"1a3c71d6e055511bd0a7546796ad3e486a2d26993c5912c2166d71299f38aeb2\": rpc error: code = NotFound desc = could not find container \"1a3c71d6e055511bd0a7546796ad3e486a2d26993c5912c2166d71299f38aeb2\": container with ID starting with 1a3c71d6e055511bd0a7546796ad3e486a2d26993c5912c2166d71299f38aeb2 not found: ID does not exist" Oct 06 10:47:30 crc kubenswrapper[4824]: I1006 10:47:30.934656 4824 scope.go:117] "RemoveContainer" containerID="179d16a7ce7922c7c3ac36786ee43cacaa7c53bb09b6e94cd8818ade0c86e582" Oct 06 10:47:30 crc kubenswrapper[4824]: E1006 10:47:30.935042 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"179d16a7ce7922c7c3ac36786ee43cacaa7c53bb09b6e94cd8818ade0c86e582\": container with ID starting with 179d16a7ce7922c7c3ac36786ee43cacaa7c53bb09b6e94cd8818ade0c86e582 not found: ID does not exist" containerID="179d16a7ce7922c7c3ac36786ee43cacaa7c53bb09b6e94cd8818ade0c86e582" Oct 06 10:47:30 crc kubenswrapper[4824]: I1006 10:47:30.935081 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"179d16a7ce7922c7c3ac36786ee43cacaa7c53bb09b6e94cd8818ade0c86e582"} err="failed to get container status \"179d16a7ce7922c7c3ac36786ee43cacaa7c53bb09b6e94cd8818ade0c86e582\": rpc error: code = NotFound desc = could not find container \"179d16a7ce7922c7c3ac36786ee43cacaa7c53bb09b6e94cd8818ade0c86e582\": container with ID starting with 179d16a7ce7922c7c3ac36786ee43cacaa7c53bb09b6e94cd8818ade0c86e582 not found: ID does not exist" Oct 06 10:47:31 crc kubenswrapper[4824]: I1006 10:47:31.292714 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="202562de-e495-4708-8608-a857b9135578" path="/var/lib/kubelet/pods/202562de-e495-4708-8608-a857b9135578/volumes" Oct 06 10:49:13 crc kubenswrapper[4824]: I1006 10:49:13.916005 4824 patch_prober.go:28] interesting pod/machine-config-daemon-khgzw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 10:49:13 crc kubenswrapper[4824]: I1006 10:49:13.916852 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 10:49:26 crc kubenswrapper[4824]: I1006 10:49:26.114194 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-987hf"] Oct 06 10:49:26 crc kubenswrapper[4824]: E1006 10:49:26.117816 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="202562de-e495-4708-8608-a857b9135578" containerName="extract-content" Oct 06 10:49:26 crc kubenswrapper[4824]: I1006 10:49:26.118004 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="202562de-e495-4708-8608-a857b9135578" containerName="extract-content" Oct 06 10:49:26 crc kubenswrapper[4824]: E1006 10:49:26.118098 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="202562de-e495-4708-8608-a857b9135578" containerName="registry-server" Oct 06 10:49:26 crc kubenswrapper[4824]: I1006 10:49:26.118178 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="202562de-e495-4708-8608-a857b9135578" containerName="registry-server" Oct 06 10:49:26 crc kubenswrapper[4824]: E1006 10:49:26.118301 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="202562de-e495-4708-8608-a857b9135578" containerName="extract-utilities" Oct 06 10:49:26 crc kubenswrapper[4824]: I1006 10:49:26.118385 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="202562de-e495-4708-8608-a857b9135578" containerName="extract-utilities" Oct 06 10:49:26 crc kubenswrapper[4824]: I1006 10:49:26.119442 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="202562de-e495-4708-8608-a857b9135578" containerName="registry-server" Oct 06 10:49:26 crc kubenswrapper[4824]: I1006 10:49:26.123259 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-987hf" Oct 06 10:49:26 crc kubenswrapper[4824]: I1006 10:49:26.143079 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a9ee86f1-e121-411e-829f-de4d675605e7-catalog-content\") pod \"certified-operators-987hf\" (UID: \"a9ee86f1-e121-411e-829f-de4d675605e7\") " pod="openshift-marketplace/certified-operators-987hf" Oct 06 10:49:26 crc kubenswrapper[4824]: I1006 10:49:26.143191 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sgfl7\" (UniqueName: \"kubernetes.io/projected/a9ee86f1-e121-411e-829f-de4d675605e7-kube-api-access-sgfl7\") pod \"certified-operators-987hf\" (UID: \"a9ee86f1-e121-411e-829f-de4d675605e7\") " pod="openshift-marketplace/certified-operators-987hf" Oct 06 10:49:26 crc kubenswrapper[4824]: I1006 10:49:26.143470 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a9ee86f1-e121-411e-829f-de4d675605e7-utilities\") pod \"certified-operators-987hf\" (UID: \"a9ee86f1-e121-411e-829f-de4d675605e7\") " pod="openshift-marketplace/certified-operators-987hf" Oct 06 10:49:26 crc kubenswrapper[4824]: I1006 10:49:26.147030 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-987hf"] Oct 06 10:49:26 crc kubenswrapper[4824]: I1006 10:49:26.245625 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a9ee86f1-e121-411e-829f-de4d675605e7-utilities\") pod \"certified-operators-987hf\" (UID: \"a9ee86f1-e121-411e-829f-de4d675605e7\") " pod="openshift-marketplace/certified-operators-987hf" Oct 06 10:49:26 crc kubenswrapper[4824]: I1006 10:49:26.245753 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a9ee86f1-e121-411e-829f-de4d675605e7-catalog-content\") pod \"certified-operators-987hf\" (UID: \"a9ee86f1-e121-411e-829f-de4d675605e7\") " pod="openshift-marketplace/certified-operators-987hf" Oct 06 10:49:26 crc kubenswrapper[4824]: I1006 10:49:26.245799 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sgfl7\" (UniqueName: \"kubernetes.io/projected/a9ee86f1-e121-411e-829f-de4d675605e7-kube-api-access-sgfl7\") pod \"certified-operators-987hf\" (UID: \"a9ee86f1-e121-411e-829f-de4d675605e7\") " pod="openshift-marketplace/certified-operators-987hf" Oct 06 10:49:26 crc kubenswrapper[4824]: I1006 10:49:26.246406 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a9ee86f1-e121-411e-829f-de4d675605e7-catalog-content\") pod \"certified-operators-987hf\" (UID: \"a9ee86f1-e121-411e-829f-de4d675605e7\") " pod="openshift-marketplace/certified-operators-987hf" Oct 06 10:49:26 crc kubenswrapper[4824]: I1006 10:49:26.246636 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a9ee86f1-e121-411e-829f-de4d675605e7-utilities\") pod \"certified-operators-987hf\" (UID: \"a9ee86f1-e121-411e-829f-de4d675605e7\") " pod="openshift-marketplace/certified-operators-987hf" Oct 06 10:49:26 crc kubenswrapper[4824]: I1006 10:49:26.267066 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sgfl7\" (UniqueName: \"kubernetes.io/projected/a9ee86f1-e121-411e-829f-de4d675605e7-kube-api-access-sgfl7\") pod \"certified-operators-987hf\" (UID: \"a9ee86f1-e121-411e-829f-de4d675605e7\") " pod="openshift-marketplace/certified-operators-987hf" Oct 06 10:49:26 crc kubenswrapper[4824]: I1006 10:49:26.449469 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-987hf" Oct 06 10:49:27 crc kubenswrapper[4824]: I1006 10:49:27.019073 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-987hf"] Oct 06 10:49:27 crc kubenswrapper[4824]: I1006 10:49:27.147805 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-987hf" event={"ID":"a9ee86f1-e121-411e-829f-de4d675605e7","Type":"ContainerStarted","Data":"da291761ef54f9cec4703e3e3dc4fc768d74700b982deacd12209c061bbdab9a"} Oct 06 10:49:28 crc kubenswrapper[4824]: I1006 10:49:28.159320 4824 generic.go:334] "Generic (PLEG): container finished" podID="a9ee86f1-e121-411e-829f-de4d675605e7" containerID="34aba73e6d6f8254cceb6a27586c5b7ef850f2ee567168f278b3a90e237d8a81" exitCode=0 Oct 06 10:49:28 crc kubenswrapper[4824]: I1006 10:49:28.159392 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-987hf" event={"ID":"a9ee86f1-e121-411e-829f-de4d675605e7","Type":"ContainerDied","Data":"34aba73e6d6f8254cceb6a27586c5b7ef850f2ee567168f278b3a90e237d8a81"} Oct 06 10:49:28 crc kubenswrapper[4824]: I1006 10:49:28.162800 4824 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 06 10:49:29 crc kubenswrapper[4824]: I1006 10:49:29.173795 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-987hf" event={"ID":"a9ee86f1-e121-411e-829f-de4d675605e7","Type":"ContainerStarted","Data":"55bbdf9ca487f8a3d001e8f27250cf7e542ce00c5a46298d64166ad38655eef5"} Oct 06 10:49:30 crc kubenswrapper[4824]: I1006 10:49:30.191504 4824 generic.go:334] "Generic (PLEG): container finished" podID="a9ee86f1-e121-411e-829f-de4d675605e7" containerID="55bbdf9ca487f8a3d001e8f27250cf7e542ce00c5a46298d64166ad38655eef5" exitCode=0 Oct 06 10:49:30 crc kubenswrapper[4824]: I1006 10:49:30.191617 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-987hf" event={"ID":"a9ee86f1-e121-411e-829f-de4d675605e7","Type":"ContainerDied","Data":"55bbdf9ca487f8a3d001e8f27250cf7e542ce00c5a46298d64166ad38655eef5"} Oct 06 10:49:31 crc kubenswrapper[4824]: I1006 10:49:31.231861 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-987hf" event={"ID":"a9ee86f1-e121-411e-829f-de4d675605e7","Type":"ContainerStarted","Data":"46762f7b90e285817eeefafe1900265f0045a05b266cce8ae421243852c48fb4"} Oct 06 10:49:31 crc kubenswrapper[4824]: I1006 10:49:31.257472 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-987hf" podStartSLOduration=2.460192828 podStartE2EDuration="5.257445017s" podCreationTimestamp="2025-10-06 10:49:26 +0000 UTC" firstStartedPulling="2025-10-06 10:49:28.162524234 +0000 UTC m=+3077.526947105" lastFinishedPulling="2025-10-06 10:49:30.959776423 +0000 UTC m=+3080.324199294" observedRunningTime="2025-10-06 10:49:31.255682114 +0000 UTC m=+3080.620104975" watchObservedRunningTime="2025-10-06 10:49:31.257445017 +0000 UTC m=+3080.621867878" Oct 06 10:49:36 crc kubenswrapper[4824]: I1006 10:49:36.451364 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-987hf" Oct 06 10:49:36 crc kubenswrapper[4824]: I1006 10:49:36.452272 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-987hf" Oct 06 10:49:36 crc kubenswrapper[4824]: I1006 10:49:36.511766 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-987hf" Oct 06 10:49:37 crc kubenswrapper[4824]: I1006 10:49:37.378751 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-987hf" Oct 06 10:49:37 crc kubenswrapper[4824]: I1006 10:49:37.440658 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-987hf"] Oct 06 10:49:39 crc kubenswrapper[4824]: I1006 10:49:39.308462 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-987hf" podUID="a9ee86f1-e121-411e-829f-de4d675605e7" containerName="registry-server" containerID="cri-o://46762f7b90e285817eeefafe1900265f0045a05b266cce8ae421243852c48fb4" gracePeriod=2 Oct 06 10:49:39 crc kubenswrapper[4824]: I1006 10:49:39.948492 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-987hf" Oct 06 10:49:39 crc kubenswrapper[4824]: I1006 10:49:39.962911 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sgfl7\" (UniqueName: \"kubernetes.io/projected/a9ee86f1-e121-411e-829f-de4d675605e7-kube-api-access-sgfl7\") pod \"a9ee86f1-e121-411e-829f-de4d675605e7\" (UID: \"a9ee86f1-e121-411e-829f-de4d675605e7\") " Oct 06 10:49:39 crc kubenswrapper[4824]: I1006 10:49:39.963155 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a9ee86f1-e121-411e-829f-de4d675605e7-catalog-content\") pod \"a9ee86f1-e121-411e-829f-de4d675605e7\" (UID: \"a9ee86f1-e121-411e-829f-de4d675605e7\") " Oct 06 10:49:39 crc kubenswrapper[4824]: I1006 10:49:39.963276 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a9ee86f1-e121-411e-829f-de4d675605e7-utilities\") pod \"a9ee86f1-e121-411e-829f-de4d675605e7\" (UID: \"a9ee86f1-e121-411e-829f-de4d675605e7\") " Oct 06 10:49:39 crc kubenswrapper[4824]: I1006 10:49:39.964711 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a9ee86f1-e121-411e-829f-de4d675605e7-utilities" (OuterVolumeSpecName: "utilities") pod "a9ee86f1-e121-411e-829f-de4d675605e7" (UID: "a9ee86f1-e121-411e-829f-de4d675605e7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:49:39 crc kubenswrapper[4824]: I1006 10:49:39.977648 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a9ee86f1-e121-411e-829f-de4d675605e7-kube-api-access-sgfl7" (OuterVolumeSpecName: "kube-api-access-sgfl7") pod "a9ee86f1-e121-411e-829f-de4d675605e7" (UID: "a9ee86f1-e121-411e-829f-de4d675605e7"). InnerVolumeSpecName "kube-api-access-sgfl7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:49:40 crc kubenswrapper[4824]: I1006 10:49:40.035585 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a9ee86f1-e121-411e-829f-de4d675605e7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a9ee86f1-e121-411e-829f-de4d675605e7" (UID: "a9ee86f1-e121-411e-829f-de4d675605e7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:49:40 crc kubenswrapper[4824]: I1006 10:49:40.067104 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sgfl7\" (UniqueName: \"kubernetes.io/projected/a9ee86f1-e121-411e-829f-de4d675605e7-kube-api-access-sgfl7\") on node \"crc\" DevicePath \"\"" Oct 06 10:49:40 crc kubenswrapper[4824]: I1006 10:49:40.067157 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a9ee86f1-e121-411e-829f-de4d675605e7-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 06 10:49:40 crc kubenswrapper[4824]: I1006 10:49:40.067176 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a9ee86f1-e121-411e-829f-de4d675605e7-utilities\") on node \"crc\" DevicePath \"\"" Oct 06 10:49:40 crc kubenswrapper[4824]: I1006 10:49:40.320942 4824 generic.go:334] "Generic (PLEG): container finished" podID="a9ee86f1-e121-411e-829f-de4d675605e7" containerID="46762f7b90e285817eeefafe1900265f0045a05b266cce8ae421243852c48fb4" exitCode=0 Oct 06 10:49:40 crc kubenswrapper[4824]: I1006 10:49:40.321011 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-987hf" event={"ID":"a9ee86f1-e121-411e-829f-de4d675605e7","Type":"ContainerDied","Data":"46762f7b90e285817eeefafe1900265f0045a05b266cce8ae421243852c48fb4"} Oct 06 10:49:40 crc kubenswrapper[4824]: I1006 10:49:40.321041 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-987hf" event={"ID":"a9ee86f1-e121-411e-829f-de4d675605e7","Type":"ContainerDied","Data":"da291761ef54f9cec4703e3e3dc4fc768d74700b982deacd12209c061bbdab9a"} Oct 06 10:49:40 crc kubenswrapper[4824]: I1006 10:49:40.321060 4824 scope.go:117] "RemoveContainer" containerID="46762f7b90e285817eeefafe1900265f0045a05b266cce8ae421243852c48fb4" Oct 06 10:49:40 crc kubenswrapper[4824]: I1006 10:49:40.321178 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-987hf" Oct 06 10:49:40 crc kubenswrapper[4824]: I1006 10:49:40.346606 4824 scope.go:117] "RemoveContainer" containerID="55bbdf9ca487f8a3d001e8f27250cf7e542ce00c5a46298d64166ad38655eef5" Oct 06 10:49:40 crc kubenswrapper[4824]: I1006 10:49:40.386416 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-987hf"] Oct 06 10:49:40 crc kubenswrapper[4824]: I1006 10:49:40.394328 4824 scope.go:117] "RemoveContainer" containerID="34aba73e6d6f8254cceb6a27586c5b7ef850f2ee567168f278b3a90e237d8a81" Oct 06 10:49:40 crc kubenswrapper[4824]: I1006 10:49:40.406539 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-987hf"] Oct 06 10:49:40 crc kubenswrapper[4824]: I1006 10:49:40.445474 4824 scope.go:117] "RemoveContainer" containerID="46762f7b90e285817eeefafe1900265f0045a05b266cce8ae421243852c48fb4" Oct 06 10:49:40 crc kubenswrapper[4824]: E1006 10:49:40.446294 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"46762f7b90e285817eeefafe1900265f0045a05b266cce8ae421243852c48fb4\": container with ID starting with 46762f7b90e285817eeefafe1900265f0045a05b266cce8ae421243852c48fb4 not found: ID does not exist" containerID="46762f7b90e285817eeefafe1900265f0045a05b266cce8ae421243852c48fb4" Oct 06 10:49:40 crc kubenswrapper[4824]: I1006 10:49:40.446339 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"46762f7b90e285817eeefafe1900265f0045a05b266cce8ae421243852c48fb4"} err="failed to get container status \"46762f7b90e285817eeefafe1900265f0045a05b266cce8ae421243852c48fb4\": rpc error: code = NotFound desc = could not find container \"46762f7b90e285817eeefafe1900265f0045a05b266cce8ae421243852c48fb4\": container with ID starting with 46762f7b90e285817eeefafe1900265f0045a05b266cce8ae421243852c48fb4 not found: ID does not exist" Oct 06 10:49:40 crc kubenswrapper[4824]: I1006 10:49:40.446366 4824 scope.go:117] "RemoveContainer" containerID="55bbdf9ca487f8a3d001e8f27250cf7e542ce00c5a46298d64166ad38655eef5" Oct 06 10:49:40 crc kubenswrapper[4824]: E1006 10:49:40.447136 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"55bbdf9ca487f8a3d001e8f27250cf7e542ce00c5a46298d64166ad38655eef5\": container with ID starting with 55bbdf9ca487f8a3d001e8f27250cf7e542ce00c5a46298d64166ad38655eef5 not found: ID does not exist" containerID="55bbdf9ca487f8a3d001e8f27250cf7e542ce00c5a46298d64166ad38655eef5" Oct 06 10:49:40 crc kubenswrapper[4824]: I1006 10:49:40.447249 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"55bbdf9ca487f8a3d001e8f27250cf7e542ce00c5a46298d64166ad38655eef5"} err="failed to get container status \"55bbdf9ca487f8a3d001e8f27250cf7e542ce00c5a46298d64166ad38655eef5\": rpc error: code = NotFound desc = could not find container \"55bbdf9ca487f8a3d001e8f27250cf7e542ce00c5a46298d64166ad38655eef5\": container with ID starting with 55bbdf9ca487f8a3d001e8f27250cf7e542ce00c5a46298d64166ad38655eef5 not found: ID does not exist" Oct 06 10:49:40 crc kubenswrapper[4824]: I1006 10:49:40.447318 4824 scope.go:117] "RemoveContainer" containerID="34aba73e6d6f8254cceb6a27586c5b7ef850f2ee567168f278b3a90e237d8a81" Oct 06 10:49:40 crc kubenswrapper[4824]: E1006 10:49:40.447839 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"34aba73e6d6f8254cceb6a27586c5b7ef850f2ee567168f278b3a90e237d8a81\": container with ID starting with 34aba73e6d6f8254cceb6a27586c5b7ef850f2ee567168f278b3a90e237d8a81 not found: ID does not exist" containerID="34aba73e6d6f8254cceb6a27586c5b7ef850f2ee567168f278b3a90e237d8a81" Oct 06 10:49:40 crc kubenswrapper[4824]: I1006 10:49:40.447877 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"34aba73e6d6f8254cceb6a27586c5b7ef850f2ee567168f278b3a90e237d8a81"} err="failed to get container status \"34aba73e6d6f8254cceb6a27586c5b7ef850f2ee567168f278b3a90e237d8a81\": rpc error: code = NotFound desc = could not find container \"34aba73e6d6f8254cceb6a27586c5b7ef850f2ee567168f278b3a90e237d8a81\": container with ID starting with 34aba73e6d6f8254cceb6a27586c5b7ef850f2ee567168f278b3a90e237d8a81 not found: ID does not exist" Oct 06 10:49:41 crc kubenswrapper[4824]: I1006 10:49:41.295323 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a9ee86f1-e121-411e-829f-de4d675605e7" path="/var/lib/kubelet/pods/a9ee86f1-e121-411e-829f-de4d675605e7/volumes" Oct 06 10:49:43 crc kubenswrapper[4824]: I1006 10:49:43.915566 4824 patch_prober.go:28] interesting pod/machine-config-daemon-khgzw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 10:49:43 crc kubenswrapper[4824]: I1006 10:49:43.916060 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 10:50:13 crc kubenswrapper[4824]: I1006 10:50:13.916012 4824 patch_prober.go:28] interesting pod/machine-config-daemon-khgzw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 10:50:13 crc kubenswrapper[4824]: I1006 10:50:13.916673 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 10:50:13 crc kubenswrapper[4824]: I1006 10:50:13.916742 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" Oct 06 10:50:13 crc kubenswrapper[4824]: I1006 10:50:13.917798 4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"bbb34b884429016efb41d11fc1bb14bdd15d27ec1a282254f5cbaae186b1d1ef"} pod="openshift-machine-config-operator/machine-config-daemon-khgzw" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 06 10:50:13 crc kubenswrapper[4824]: I1006 10:50:13.917892 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" containerName="machine-config-daemon" containerID="cri-o://bbb34b884429016efb41d11fc1bb14bdd15d27ec1a282254f5cbaae186b1d1ef" gracePeriod=600 Oct 06 10:50:14 crc kubenswrapper[4824]: E1006 10:50:14.054839 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-khgzw_openshift-machine-config-operator(e1140c8c-93f1-4fce-8c06-a69561ff8a6d)\"" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" Oct 06 10:50:14 crc kubenswrapper[4824]: I1006 10:50:14.689329 4824 generic.go:334] "Generic (PLEG): container finished" podID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" containerID="bbb34b884429016efb41d11fc1bb14bdd15d27ec1a282254f5cbaae186b1d1ef" exitCode=0 Oct 06 10:50:14 crc kubenswrapper[4824]: I1006 10:50:14.689449 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" event={"ID":"e1140c8c-93f1-4fce-8c06-a69561ff8a6d","Type":"ContainerDied","Data":"bbb34b884429016efb41d11fc1bb14bdd15d27ec1a282254f5cbaae186b1d1ef"} Oct 06 10:50:14 crc kubenswrapper[4824]: I1006 10:50:14.689542 4824 scope.go:117] "RemoveContainer" containerID="db15accdeb46a9039be894301153ff2ab7709ec808de959d20f2a048678d55cc" Oct 06 10:50:14 crc kubenswrapper[4824]: I1006 10:50:14.690347 4824 scope.go:117] "RemoveContainer" containerID="bbb34b884429016efb41d11fc1bb14bdd15d27ec1a282254f5cbaae186b1d1ef" Oct 06 10:50:14 crc kubenswrapper[4824]: E1006 10:50:14.690631 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-khgzw_openshift-machine-config-operator(e1140c8c-93f1-4fce-8c06-a69561ff8a6d)\"" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" Oct 06 10:50:26 crc kubenswrapper[4824]: I1006 10:50:26.274500 4824 scope.go:117] "RemoveContainer" containerID="bbb34b884429016efb41d11fc1bb14bdd15d27ec1a282254f5cbaae186b1d1ef" Oct 06 10:50:26 crc kubenswrapper[4824]: E1006 10:50:26.275276 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-khgzw_openshift-machine-config-operator(e1140c8c-93f1-4fce-8c06-a69561ff8a6d)\"" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" Oct 06 10:50:40 crc kubenswrapper[4824]: I1006 10:50:40.274628 4824 scope.go:117] "RemoveContainer" containerID="bbb34b884429016efb41d11fc1bb14bdd15d27ec1a282254f5cbaae186b1d1ef" Oct 06 10:50:40 crc kubenswrapper[4824]: E1006 10:50:40.275780 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-khgzw_openshift-machine-config-operator(e1140c8c-93f1-4fce-8c06-a69561ff8a6d)\"" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" Oct 06 10:50:53 crc kubenswrapper[4824]: I1006 10:50:53.274918 4824 scope.go:117] "RemoveContainer" containerID="bbb34b884429016efb41d11fc1bb14bdd15d27ec1a282254f5cbaae186b1d1ef" Oct 06 10:50:53 crc kubenswrapper[4824]: E1006 10:50:53.275897 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-khgzw_openshift-machine-config-operator(e1140c8c-93f1-4fce-8c06-a69561ff8a6d)\"" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" Oct 06 10:51:08 crc kubenswrapper[4824]: I1006 10:51:08.273555 4824 scope.go:117] "RemoveContainer" containerID="bbb34b884429016efb41d11fc1bb14bdd15d27ec1a282254f5cbaae186b1d1ef" Oct 06 10:51:08 crc kubenswrapper[4824]: E1006 10:51:08.274441 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-khgzw_openshift-machine-config-operator(e1140c8c-93f1-4fce-8c06-a69561ff8a6d)\"" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" Oct 06 10:51:23 crc kubenswrapper[4824]: I1006 10:51:23.275506 4824 scope.go:117] "RemoveContainer" containerID="bbb34b884429016efb41d11fc1bb14bdd15d27ec1a282254f5cbaae186b1d1ef" Oct 06 10:51:23 crc kubenswrapper[4824]: E1006 10:51:23.276495 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-khgzw_openshift-machine-config-operator(e1140c8c-93f1-4fce-8c06-a69561ff8a6d)\"" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" Oct 06 10:51:38 crc kubenswrapper[4824]: I1006 10:51:38.274752 4824 scope.go:117] "RemoveContainer" containerID="bbb34b884429016efb41d11fc1bb14bdd15d27ec1a282254f5cbaae186b1d1ef" Oct 06 10:51:38 crc kubenswrapper[4824]: E1006 10:51:38.276556 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-khgzw_openshift-machine-config-operator(e1140c8c-93f1-4fce-8c06-a69561ff8a6d)\"" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" Oct 06 10:51:52 crc kubenswrapper[4824]: I1006 10:51:52.275067 4824 scope.go:117] "RemoveContainer" containerID="bbb34b884429016efb41d11fc1bb14bdd15d27ec1a282254f5cbaae186b1d1ef" Oct 06 10:51:52 crc kubenswrapper[4824]: E1006 10:51:52.276072 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-khgzw_openshift-machine-config-operator(e1140c8c-93f1-4fce-8c06-a69561ff8a6d)\"" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" Oct 06 10:52:07 crc kubenswrapper[4824]: I1006 10:52:07.275429 4824 scope.go:117] "RemoveContainer" containerID="bbb34b884429016efb41d11fc1bb14bdd15d27ec1a282254f5cbaae186b1d1ef" Oct 06 10:52:07 crc kubenswrapper[4824]: E1006 10:52:07.276298 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-khgzw_openshift-machine-config-operator(e1140c8c-93f1-4fce-8c06-a69561ff8a6d)\"" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" Oct 06 10:52:19 crc kubenswrapper[4824]: I1006 10:52:19.274610 4824 scope.go:117] "RemoveContainer" containerID="bbb34b884429016efb41d11fc1bb14bdd15d27ec1a282254f5cbaae186b1d1ef" Oct 06 10:52:19 crc kubenswrapper[4824]: E1006 10:52:19.275278 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-khgzw_openshift-machine-config-operator(e1140c8c-93f1-4fce-8c06-a69561ff8a6d)\"" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" Oct 06 10:52:31 crc kubenswrapper[4824]: I1006 10:52:31.283023 4824 scope.go:117] "RemoveContainer" containerID="bbb34b884429016efb41d11fc1bb14bdd15d27ec1a282254f5cbaae186b1d1ef" Oct 06 10:52:31 crc kubenswrapper[4824]: E1006 10:52:31.284090 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-khgzw_openshift-machine-config-operator(e1140c8c-93f1-4fce-8c06-a69561ff8a6d)\"" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" Oct 06 10:52:46 crc kubenswrapper[4824]: I1006 10:52:46.274942 4824 scope.go:117] "RemoveContainer" containerID="bbb34b884429016efb41d11fc1bb14bdd15d27ec1a282254f5cbaae186b1d1ef" Oct 06 10:52:46 crc kubenswrapper[4824]: E1006 10:52:46.277217 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-khgzw_openshift-machine-config-operator(e1140c8c-93f1-4fce-8c06-a69561ff8a6d)\"" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" Oct 06 10:53:01 crc kubenswrapper[4824]: I1006 10:53:01.284515 4824 scope.go:117] "RemoveContainer" containerID="bbb34b884429016efb41d11fc1bb14bdd15d27ec1a282254f5cbaae186b1d1ef" Oct 06 10:53:01 crc kubenswrapper[4824]: E1006 10:53:01.285635 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-khgzw_openshift-machine-config-operator(e1140c8c-93f1-4fce-8c06-a69561ff8a6d)\"" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" Oct 06 10:53:12 crc kubenswrapper[4824]: I1006 10:53:12.274626 4824 scope.go:117] "RemoveContainer" containerID="bbb34b884429016efb41d11fc1bb14bdd15d27ec1a282254f5cbaae186b1d1ef" Oct 06 10:53:12 crc kubenswrapper[4824]: E1006 10:53:12.275505 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-khgzw_openshift-machine-config-operator(e1140c8c-93f1-4fce-8c06-a69561ff8a6d)\"" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" Oct 06 10:53:23 crc kubenswrapper[4824]: I1006 10:53:23.274858 4824 scope.go:117] "RemoveContainer" containerID="bbb34b884429016efb41d11fc1bb14bdd15d27ec1a282254f5cbaae186b1d1ef" Oct 06 10:53:23 crc kubenswrapper[4824]: E1006 10:53:23.276155 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-khgzw_openshift-machine-config-operator(e1140c8c-93f1-4fce-8c06-a69561ff8a6d)\"" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" Oct 06 10:53:35 crc kubenswrapper[4824]: I1006 10:53:35.274317 4824 scope.go:117] "RemoveContainer" containerID="bbb34b884429016efb41d11fc1bb14bdd15d27ec1a282254f5cbaae186b1d1ef" Oct 06 10:53:35 crc kubenswrapper[4824]: E1006 10:53:35.275311 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-khgzw_openshift-machine-config-operator(e1140c8c-93f1-4fce-8c06-a69561ff8a6d)\"" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" Oct 06 10:53:39 crc kubenswrapper[4824]: I1006 10:53:39.923393 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-5l7gb"] Oct 06 10:53:39 crc kubenswrapper[4824]: E1006 10:53:39.925428 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9ee86f1-e121-411e-829f-de4d675605e7" containerName="registry-server" Oct 06 10:53:39 crc kubenswrapper[4824]: I1006 10:53:39.925445 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9ee86f1-e121-411e-829f-de4d675605e7" containerName="registry-server" Oct 06 10:53:39 crc kubenswrapper[4824]: E1006 10:53:39.925465 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9ee86f1-e121-411e-829f-de4d675605e7" containerName="extract-content" Oct 06 10:53:39 crc kubenswrapper[4824]: I1006 10:53:39.925473 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9ee86f1-e121-411e-829f-de4d675605e7" containerName="extract-content" Oct 06 10:53:39 crc kubenswrapper[4824]: E1006 10:53:39.925504 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9ee86f1-e121-411e-829f-de4d675605e7" containerName="extract-utilities" Oct 06 10:53:39 crc kubenswrapper[4824]: I1006 10:53:39.925513 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9ee86f1-e121-411e-829f-de4d675605e7" containerName="extract-utilities" Oct 06 10:53:39 crc kubenswrapper[4824]: I1006 10:53:39.925745 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="a9ee86f1-e121-411e-829f-de4d675605e7" containerName="registry-server" Oct 06 10:53:39 crc kubenswrapper[4824]: I1006 10:53:39.927090 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5l7gb" Oct 06 10:53:39 crc kubenswrapper[4824]: I1006 10:53:39.958930 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-5l7gb"] Oct 06 10:53:40 crc kubenswrapper[4824]: I1006 10:53:40.059620 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qj4gk\" (UniqueName: \"kubernetes.io/projected/3c2c1933-ef36-4f64-965d-7e686a71bcb3-kube-api-access-qj4gk\") pod \"community-operators-5l7gb\" (UID: \"3c2c1933-ef36-4f64-965d-7e686a71bcb3\") " pod="openshift-marketplace/community-operators-5l7gb" Oct 06 10:53:40 crc kubenswrapper[4824]: I1006 10:53:40.059736 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c2c1933-ef36-4f64-965d-7e686a71bcb3-utilities\") pod \"community-operators-5l7gb\" (UID: \"3c2c1933-ef36-4f64-965d-7e686a71bcb3\") " pod="openshift-marketplace/community-operators-5l7gb" Oct 06 10:53:40 crc kubenswrapper[4824]: I1006 10:53:40.059929 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c2c1933-ef36-4f64-965d-7e686a71bcb3-catalog-content\") pod \"community-operators-5l7gb\" (UID: \"3c2c1933-ef36-4f64-965d-7e686a71bcb3\") " pod="openshift-marketplace/community-operators-5l7gb" Oct 06 10:53:40 crc kubenswrapper[4824]: I1006 10:53:40.161626 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qj4gk\" (UniqueName: \"kubernetes.io/projected/3c2c1933-ef36-4f64-965d-7e686a71bcb3-kube-api-access-qj4gk\") pod \"community-operators-5l7gb\" (UID: \"3c2c1933-ef36-4f64-965d-7e686a71bcb3\") " pod="openshift-marketplace/community-operators-5l7gb" Oct 06 10:53:40 crc kubenswrapper[4824]: I1006 10:53:40.162013 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c2c1933-ef36-4f64-965d-7e686a71bcb3-utilities\") pod \"community-operators-5l7gb\" (UID: \"3c2c1933-ef36-4f64-965d-7e686a71bcb3\") " pod="openshift-marketplace/community-operators-5l7gb" Oct 06 10:53:40 crc kubenswrapper[4824]: I1006 10:53:40.162090 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c2c1933-ef36-4f64-965d-7e686a71bcb3-catalog-content\") pod \"community-operators-5l7gb\" (UID: \"3c2c1933-ef36-4f64-965d-7e686a71bcb3\") " pod="openshift-marketplace/community-operators-5l7gb" Oct 06 10:53:40 crc kubenswrapper[4824]: I1006 10:53:40.162764 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c2c1933-ef36-4f64-965d-7e686a71bcb3-catalog-content\") pod \"community-operators-5l7gb\" (UID: \"3c2c1933-ef36-4f64-965d-7e686a71bcb3\") " pod="openshift-marketplace/community-operators-5l7gb" Oct 06 10:53:40 crc kubenswrapper[4824]: I1006 10:53:40.162748 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c2c1933-ef36-4f64-965d-7e686a71bcb3-utilities\") pod \"community-operators-5l7gb\" (UID: \"3c2c1933-ef36-4f64-965d-7e686a71bcb3\") " pod="openshift-marketplace/community-operators-5l7gb" Oct 06 10:53:40 crc kubenswrapper[4824]: I1006 10:53:40.194670 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qj4gk\" (UniqueName: \"kubernetes.io/projected/3c2c1933-ef36-4f64-965d-7e686a71bcb3-kube-api-access-qj4gk\") pod \"community-operators-5l7gb\" (UID: \"3c2c1933-ef36-4f64-965d-7e686a71bcb3\") " pod="openshift-marketplace/community-operators-5l7gb" Oct 06 10:53:40 crc kubenswrapper[4824]: I1006 10:53:40.262772 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5l7gb" Oct 06 10:53:40 crc kubenswrapper[4824]: I1006 10:53:40.819842 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-5l7gb"] Oct 06 10:53:41 crc kubenswrapper[4824]: I1006 10:53:41.068568 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5l7gb" event={"ID":"3c2c1933-ef36-4f64-965d-7e686a71bcb3","Type":"ContainerStarted","Data":"ed1cc1020968d9a8ad995ed3666cc543dc8bd4f428b1cc28e7e4cd27574001ba"} Oct 06 10:53:41 crc kubenswrapper[4824]: I1006 10:53:41.069226 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5l7gb" event={"ID":"3c2c1933-ef36-4f64-965d-7e686a71bcb3","Type":"ContainerStarted","Data":"e9fd4c2952c7642facec43516bce4ff54416b96a0a9cf5597ecb35166ee98c1d"} Oct 06 10:53:42 crc kubenswrapper[4824]: I1006 10:53:42.084022 4824 generic.go:334] "Generic (PLEG): container finished" podID="3c2c1933-ef36-4f64-965d-7e686a71bcb3" containerID="ed1cc1020968d9a8ad995ed3666cc543dc8bd4f428b1cc28e7e4cd27574001ba" exitCode=0 Oct 06 10:53:42 crc kubenswrapper[4824]: I1006 10:53:42.084201 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5l7gb" event={"ID":"3c2c1933-ef36-4f64-965d-7e686a71bcb3","Type":"ContainerDied","Data":"ed1cc1020968d9a8ad995ed3666cc543dc8bd4f428b1cc28e7e4cd27574001ba"} Oct 06 10:53:43 crc kubenswrapper[4824]: I1006 10:53:43.096844 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5l7gb" event={"ID":"3c2c1933-ef36-4f64-965d-7e686a71bcb3","Type":"ContainerStarted","Data":"a2c0ad51368b863dd68c318a5476402a2958b522ec2faa362a86e8970c253b61"} Oct 06 10:53:44 crc kubenswrapper[4824]: I1006 10:53:44.108904 4824 generic.go:334] "Generic (PLEG): container finished" podID="3c2c1933-ef36-4f64-965d-7e686a71bcb3" containerID="a2c0ad51368b863dd68c318a5476402a2958b522ec2faa362a86e8970c253b61" exitCode=0 Oct 06 10:53:44 crc kubenswrapper[4824]: I1006 10:53:44.109466 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5l7gb" event={"ID":"3c2c1933-ef36-4f64-965d-7e686a71bcb3","Type":"ContainerDied","Data":"a2c0ad51368b863dd68c318a5476402a2958b522ec2faa362a86e8970c253b61"} Oct 06 10:53:45 crc kubenswrapper[4824]: I1006 10:53:45.121869 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5l7gb" event={"ID":"3c2c1933-ef36-4f64-965d-7e686a71bcb3","Type":"ContainerStarted","Data":"b879208494b04ca1fc995e8371057812bb6480e81b65f37bd85434674e399526"} Oct 06 10:53:45 crc kubenswrapper[4824]: I1006 10:53:45.153646 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-5l7gb" podStartSLOduration=3.771526409 podStartE2EDuration="6.153608973s" podCreationTimestamp="2025-10-06 10:53:39 +0000 UTC" firstStartedPulling="2025-10-06 10:53:42.087426919 +0000 UTC m=+3331.451849820" lastFinishedPulling="2025-10-06 10:53:44.469509523 +0000 UTC m=+3333.833932384" observedRunningTime="2025-10-06 10:53:45.148113857 +0000 UTC m=+3334.512536718" watchObservedRunningTime="2025-10-06 10:53:45.153608973 +0000 UTC m=+3334.518031834" Oct 06 10:53:46 crc kubenswrapper[4824]: I1006 10:53:46.275547 4824 scope.go:117] "RemoveContainer" containerID="bbb34b884429016efb41d11fc1bb14bdd15d27ec1a282254f5cbaae186b1d1ef" Oct 06 10:53:46 crc kubenswrapper[4824]: E1006 10:53:46.276348 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-khgzw_openshift-machine-config-operator(e1140c8c-93f1-4fce-8c06-a69561ff8a6d)\"" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" Oct 06 10:53:50 crc kubenswrapper[4824]: I1006 10:53:50.264391 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-5l7gb" Oct 06 10:53:50 crc kubenswrapper[4824]: I1006 10:53:50.265143 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-5l7gb" Oct 06 10:53:50 crc kubenswrapper[4824]: I1006 10:53:50.322936 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-5l7gb" Oct 06 10:53:51 crc kubenswrapper[4824]: I1006 10:53:51.252101 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-5l7gb" Oct 06 10:53:51 crc kubenswrapper[4824]: I1006 10:53:51.914622 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-5l7gb"] Oct 06 10:53:53 crc kubenswrapper[4824]: I1006 10:53:53.213872 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-5l7gb" podUID="3c2c1933-ef36-4f64-965d-7e686a71bcb3" containerName="registry-server" containerID="cri-o://b879208494b04ca1fc995e8371057812bb6480e81b65f37bd85434674e399526" gracePeriod=2 Oct 06 10:53:53 crc kubenswrapper[4824]: E1006 10:53:53.416310 4824 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3c2c1933_ef36_4f64_965d_7e686a71bcb3.slice/crio-conmon-b879208494b04ca1fc995e8371057812bb6480e81b65f37bd85434674e399526.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3c2c1933_ef36_4f64_965d_7e686a71bcb3.slice/crio-b879208494b04ca1fc995e8371057812bb6480e81b65f37bd85434674e399526.scope\": RecentStats: unable to find data in memory cache]" Oct 06 10:53:53 crc kubenswrapper[4824]: I1006 10:53:53.788839 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5l7gb" Oct 06 10:53:53 crc kubenswrapper[4824]: I1006 10:53:53.883633 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c2c1933-ef36-4f64-965d-7e686a71bcb3-utilities\") pod \"3c2c1933-ef36-4f64-965d-7e686a71bcb3\" (UID: \"3c2c1933-ef36-4f64-965d-7e686a71bcb3\") " Oct 06 10:53:53 crc kubenswrapper[4824]: I1006 10:53:53.883854 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qj4gk\" (UniqueName: \"kubernetes.io/projected/3c2c1933-ef36-4f64-965d-7e686a71bcb3-kube-api-access-qj4gk\") pod \"3c2c1933-ef36-4f64-965d-7e686a71bcb3\" (UID: \"3c2c1933-ef36-4f64-965d-7e686a71bcb3\") " Oct 06 10:53:53 crc kubenswrapper[4824]: I1006 10:53:53.883895 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c2c1933-ef36-4f64-965d-7e686a71bcb3-catalog-content\") pod \"3c2c1933-ef36-4f64-965d-7e686a71bcb3\" (UID: \"3c2c1933-ef36-4f64-965d-7e686a71bcb3\") " Oct 06 10:53:53 crc kubenswrapper[4824]: I1006 10:53:53.888452 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3c2c1933-ef36-4f64-965d-7e686a71bcb3-utilities" (OuterVolumeSpecName: "utilities") pod "3c2c1933-ef36-4f64-965d-7e686a71bcb3" (UID: "3c2c1933-ef36-4f64-965d-7e686a71bcb3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:53:53 crc kubenswrapper[4824]: I1006 10:53:53.896920 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3c2c1933-ef36-4f64-965d-7e686a71bcb3-kube-api-access-qj4gk" (OuterVolumeSpecName: "kube-api-access-qj4gk") pod "3c2c1933-ef36-4f64-965d-7e686a71bcb3" (UID: "3c2c1933-ef36-4f64-965d-7e686a71bcb3"). InnerVolumeSpecName "kube-api-access-qj4gk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:53:53 crc kubenswrapper[4824]: I1006 10:53:53.939785 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3c2c1933-ef36-4f64-965d-7e686a71bcb3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3c2c1933-ef36-4f64-965d-7e686a71bcb3" (UID: "3c2c1933-ef36-4f64-965d-7e686a71bcb3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:53:53 crc kubenswrapper[4824]: I1006 10:53:53.987421 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qj4gk\" (UniqueName: \"kubernetes.io/projected/3c2c1933-ef36-4f64-965d-7e686a71bcb3-kube-api-access-qj4gk\") on node \"crc\" DevicePath \"\"" Oct 06 10:53:53 crc kubenswrapper[4824]: I1006 10:53:53.987781 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c2c1933-ef36-4f64-965d-7e686a71bcb3-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 06 10:53:53 crc kubenswrapper[4824]: I1006 10:53:53.987924 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c2c1933-ef36-4f64-965d-7e686a71bcb3-utilities\") on node \"crc\" DevicePath \"\"" Oct 06 10:53:54 crc kubenswrapper[4824]: I1006 10:53:54.237962 4824 generic.go:334] "Generic (PLEG): container finished" podID="3c2c1933-ef36-4f64-965d-7e686a71bcb3" containerID="b879208494b04ca1fc995e8371057812bb6480e81b65f37bd85434674e399526" exitCode=0 Oct 06 10:53:54 crc kubenswrapper[4824]: I1006 10:53:54.238149 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5l7gb" event={"ID":"3c2c1933-ef36-4f64-965d-7e686a71bcb3","Type":"ContainerDied","Data":"b879208494b04ca1fc995e8371057812bb6480e81b65f37bd85434674e399526"} Oct 06 10:53:54 crc kubenswrapper[4824]: I1006 10:53:54.238224 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5l7gb" Oct 06 10:53:54 crc kubenswrapper[4824]: I1006 10:53:54.238267 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5l7gb" event={"ID":"3c2c1933-ef36-4f64-965d-7e686a71bcb3","Type":"ContainerDied","Data":"e9fd4c2952c7642facec43516bce4ff54416b96a0a9cf5597ecb35166ee98c1d"} Oct 06 10:53:54 crc kubenswrapper[4824]: I1006 10:53:54.238365 4824 scope.go:117] "RemoveContainer" containerID="b879208494b04ca1fc995e8371057812bb6480e81b65f37bd85434674e399526" Oct 06 10:53:54 crc kubenswrapper[4824]: I1006 10:53:54.286269 4824 scope.go:117] "RemoveContainer" containerID="a2c0ad51368b863dd68c318a5476402a2958b522ec2faa362a86e8970c253b61" Oct 06 10:53:54 crc kubenswrapper[4824]: I1006 10:53:54.295625 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-5l7gb"] Oct 06 10:53:54 crc kubenswrapper[4824]: I1006 10:53:54.303724 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-5l7gb"] Oct 06 10:53:54 crc kubenswrapper[4824]: I1006 10:53:54.327852 4824 scope.go:117] "RemoveContainer" containerID="ed1cc1020968d9a8ad995ed3666cc543dc8bd4f428b1cc28e7e4cd27574001ba" Oct 06 10:53:54 crc kubenswrapper[4824]: I1006 10:53:54.367999 4824 scope.go:117] "RemoveContainer" containerID="b879208494b04ca1fc995e8371057812bb6480e81b65f37bd85434674e399526" Oct 06 10:53:54 crc kubenswrapper[4824]: E1006 10:53:54.368469 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b879208494b04ca1fc995e8371057812bb6480e81b65f37bd85434674e399526\": container with ID starting with b879208494b04ca1fc995e8371057812bb6480e81b65f37bd85434674e399526 not found: ID does not exist" containerID="b879208494b04ca1fc995e8371057812bb6480e81b65f37bd85434674e399526" Oct 06 10:53:54 crc kubenswrapper[4824]: I1006 10:53:54.368513 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b879208494b04ca1fc995e8371057812bb6480e81b65f37bd85434674e399526"} err="failed to get container status \"b879208494b04ca1fc995e8371057812bb6480e81b65f37bd85434674e399526\": rpc error: code = NotFound desc = could not find container \"b879208494b04ca1fc995e8371057812bb6480e81b65f37bd85434674e399526\": container with ID starting with b879208494b04ca1fc995e8371057812bb6480e81b65f37bd85434674e399526 not found: ID does not exist" Oct 06 10:53:54 crc kubenswrapper[4824]: I1006 10:53:54.368560 4824 scope.go:117] "RemoveContainer" containerID="a2c0ad51368b863dd68c318a5476402a2958b522ec2faa362a86e8970c253b61" Oct 06 10:53:54 crc kubenswrapper[4824]: E1006 10:53:54.369107 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a2c0ad51368b863dd68c318a5476402a2958b522ec2faa362a86e8970c253b61\": container with ID starting with a2c0ad51368b863dd68c318a5476402a2958b522ec2faa362a86e8970c253b61 not found: ID does not exist" containerID="a2c0ad51368b863dd68c318a5476402a2958b522ec2faa362a86e8970c253b61" Oct 06 10:53:54 crc kubenswrapper[4824]: I1006 10:53:54.369189 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a2c0ad51368b863dd68c318a5476402a2958b522ec2faa362a86e8970c253b61"} err="failed to get container status \"a2c0ad51368b863dd68c318a5476402a2958b522ec2faa362a86e8970c253b61\": rpc error: code = NotFound desc = could not find container \"a2c0ad51368b863dd68c318a5476402a2958b522ec2faa362a86e8970c253b61\": container with ID starting with a2c0ad51368b863dd68c318a5476402a2958b522ec2faa362a86e8970c253b61 not found: ID does not exist" Oct 06 10:53:54 crc kubenswrapper[4824]: I1006 10:53:54.369251 4824 scope.go:117] "RemoveContainer" containerID="ed1cc1020968d9a8ad995ed3666cc543dc8bd4f428b1cc28e7e4cd27574001ba" Oct 06 10:53:54 crc kubenswrapper[4824]: E1006 10:53:54.369698 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ed1cc1020968d9a8ad995ed3666cc543dc8bd4f428b1cc28e7e4cd27574001ba\": container with ID starting with ed1cc1020968d9a8ad995ed3666cc543dc8bd4f428b1cc28e7e4cd27574001ba not found: ID does not exist" containerID="ed1cc1020968d9a8ad995ed3666cc543dc8bd4f428b1cc28e7e4cd27574001ba" Oct 06 10:53:54 crc kubenswrapper[4824]: I1006 10:53:54.369734 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ed1cc1020968d9a8ad995ed3666cc543dc8bd4f428b1cc28e7e4cd27574001ba"} err="failed to get container status \"ed1cc1020968d9a8ad995ed3666cc543dc8bd4f428b1cc28e7e4cd27574001ba\": rpc error: code = NotFound desc = could not find container \"ed1cc1020968d9a8ad995ed3666cc543dc8bd4f428b1cc28e7e4cd27574001ba\": container with ID starting with ed1cc1020968d9a8ad995ed3666cc543dc8bd4f428b1cc28e7e4cd27574001ba not found: ID does not exist" Oct 06 10:53:55 crc kubenswrapper[4824]: I1006 10:53:55.286459 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3c2c1933-ef36-4f64-965d-7e686a71bcb3" path="/var/lib/kubelet/pods/3c2c1933-ef36-4f64-965d-7e686a71bcb3/volumes" Oct 06 10:53:59 crc kubenswrapper[4824]: I1006 10:53:59.274585 4824 scope.go:117] "RemoveContainer" containerID="bbb34b884429016efb41d11fc1bb14bdd15d27ec1a282254f5cbaae186b1d1ef" Oct 06 10:53:59 crc kubenswrapper[4824]: E1006 10:53:59.275603 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-khgzw_openshift-machine-config-operator(e1140c8c-93f1-4fce-8c06-a69561ff8a6d)\"" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" Oct 06 10:54:12 crc kubenswrapper[4824]: I1006 10:54:12.275167 4824 scope.go:117] "RemoveContainer" containerID="bbb34b884429016efb41d11fc1bb14bdd15d27ec1a282254f5cbaae186b1d1ef" Oct 06 10:54:12 crc kubenswrapper[4824]: E1006 10:54:12.276216 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-khgzw_openshift-machine-config-operator(e1140c8c-93f1-4fce-8c06-a69561ff8a6d)\"" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" Oct 06 10:54:25 crc kubenswrapper[4824]: I1006 10:54:25.274468 4824 scope.go:117] "RemoveContainer" containerID="bbb34b884429016efb41d11fc1bb14bdd15d27ec1a282254f5cbaae186b1d1ef" Oct 06 10:54:25 crc kubenswrapper[4824]: E1006 10:54:25.275214 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-khgzw_openshift-machine-config-operator(e1140c8c-93f1-4fce-8c06-a69561ff8a6d)\"" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" Oct 06 10:54:40 crc kubenswrapper[4824]: I1006 10:54:40.274161 4824 scope.go:117] "RemoveContainer" containerID="bbb34b884429016efb41d11fc1bb14bdd15d27ec1a282254f5cbaae186b1d1ef" Oct 06 10:54:40 crc kubenswrapper[4824]: E1006 10:54:40.275484 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-khgzw_openshift-machine-config-operator(e1140c8c-93f1-4fce-8c06-a69561ff8a6d)\"" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" Oct 06 10:54:55 crc kubenswrapper[4824]: I1006 10:54:55.275865 4824 scope.go:117] "RemoveContainer" containerID="bbb34b884429016efb41d11fc1bb14bdd15d27ec1a282254f5cbaae186b1d1ef" Oct 06 10:54:55 crc kubenswrapper[4824]: E1006 10:54:55.277593 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-khgzw_openshift-machine-config-operator(e1140c8c-93f1-4fce-8c06-a69561ff8a6d)\"" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" Oct 06 10:55:07 crc kubenswrapper[4824]: I1006 10:55:07.277237 4824 scope.go:117] "RemoveContainer" containerID="bbb34b884429016efb41d11fc1bb14bdd15d27ec1a282254f5cbaae186b1d1ef" Oct 06 10:55:07 crc kubenswrapper[4824]: E1006 10:55:07.279205 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-khgzw_openshift-machine-config-operator(e1140c8c-93f1-4fce-8c06-a69561ff8a6d)\"" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" Oct 06 10:55:19 crc kubenswrapper[4824]: I1006 10:55:19.274928 4824 scope.go:117] "RemoveContainer" containerID="bbb34b884429016efb41d11fc1bb14bdd15d27ec1a282254f5cbaae186b1d1ef" Oct 06 10:55:20 crc kubenswrapper[4824]: I1006 10:55:20.175215 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" event={"ID":"e1140c8c-93f1-4fce-8c06-a69561ff8a6d","Type":"ContainerStarted","Data":"d960e8d823dfd2a405439bf2eac79b76e3ee46509993c3394eac4ba53ab10143"} Oct 06 10:57:43 crc kubenswrapper[4824]: I1006 10:57:43.915770 4824 patch_prober.go:28] interesting pod/machine-config-daemon-khgzw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 10:57:43 crc kubenswrapper[4824]: I1006 10:57:43.917413 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 10:57:45 crc kubenswrapper[4824]: I1006 10:57:45.942903 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-d8s5f"] Oct 06 10:57:45 crc kubenswrapper[4824]: E1006 10:57:45.944116 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c2c1933-ef36-4f64-965d-7e686a71bcb3" containerName="registry-server" Oct 06 10:57:45 crc kubenswrapper[4824]: I1006 10:57:45.944152 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c2c1933-ef36-4f64-965d-7e686a71bcb3" containerName="registry-server" Oct 06 10:57:45 crc kubenswrapper[4824]: E1006 10:57:45.944198 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c2c1933-ef36-4f64-965d-7e686a71bcb3" containerName="extract-utilities" Oct 06 10:57:45 crc kubenswrapper[4824]: I1006 10:57:45.944217 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c2c1933-ef36-4f64-965d-7e686a71bcb3" containerName="extract-utilities" Oct 06 10:57:45 crc kubenswrapper[4824]: E1006 10:57:45.944246 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c2c1933-ef36-4f64-965d-7e686a71bcb3" containerName="extract-content" Oct 06 10:57:45 crc kubenswrapper[4824]: I1006 10:57:45.944292 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c2c1933-ef36-4f64-965d-7e686a71bcb3" containerName="extract-content" Oct 06 10:57:45 crc kubenswrapper[4824]: I1006 10:57:45.944664 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c2c1933-ef36-4f64-965d-7e686a71bcb3" containerName="registry-server" Oct 06 10:57:45 crc kubenswrapper[4824]: I1006 10:57:45.947381 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-d8s5f" Oct 06 10:57:45 crc kubenswrapper[4824]: I1006 10:57:45.957082 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-d8s5f"] Oct 06 10:57:46 crc kubenswrapper[4824]: I1006 10:57:46.022302 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bcb79b86-65c0-44c1-9ce2-588ad66206a6-catalog-content\") pod \"redhat-operators-d8s5f\" (UID: \"bcb79b86-65c0-44c1-9ce2-588ad66206a6\") " pod="openshift-marketplace/redhat-operators-d8s5f" Oct 06 10:57:46 crc kubenswrapper[4824]: I1006 10:57:46.022402 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bcb79b86-65c0-44c1-9ce2-588ad66206a6-utilities\") pod \"redhat-operators-d8s5f\" (UID: \"bcb79b86-65c0-44c1-9ce2-588ad66206a6\") " pod="openshift-marketplace/redhat-operators-d8s5f" Oct 06 10:57:46 crc kubenswrapper[4824]: I1006 10:57:46.022645 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vqgw8\" (UniqueName: \"kubernetes.io/projected/bcb79b86-65c0-44c1-9ce2-588ad66206a6-kube-api-access-vqgw8\") pod \"redhat-operators-d8s5f\" (UID: \"bcb79b86-65c0-44c1-9ce2-588ad66206a6\") " pod="openshift-marketplace/redhat-operators-d8s5f" Oct 06 10:57:46 crc kubenswrapper[4824]: I1006 10:57:46.125128 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bcb79b86-65c0-44c1-9ce2-588ad66206a6-catalog-content\") pod \"redhat-operators-d8s5f\" (UID: \"bcb79b86-65c0-44c1-9ce2-588ad66206a6\") " pod="openshift-marketplace/redhat-operators-d8s5f" Oct 06 10:57:46 crc kubenswrapper[4824]: I1006 10:57:46.125651 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bcb79b86-65c0-44c1-9ce2-588ad66206a6-utilities\") pod \"redhat-operators-d8s5f\" (UID: \"bcb79b86-65c0-44c1-9ce2-588ad66206a6\") " pod="openshift-marketplace/redhat-operators-d8s5f" Oct 06 10:57:46 crc kubenswrapper[4824]: I1006 10:57:46.125691 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vqgw8\" (UniqueName: \"kubernetes.io/projected/bcb79b86-65c0-44c1-9ce2-588ad66206a6-kube-api-access-vqgw8\") pod \"redhat-operators-d8s5f\" (UID: \"bcb79b86-65c0-44c1-9ce2-588ad66206a6\") " pod="openshift-marketplace/redhat-operators-d8s5f" Oct 06 10:57:46 crc kubenswrapper[4824]: I1006 10:57:46.125821 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bcb79b86-65c0-44c1-9ce2-588ad66206a6-catalog-content\") pod \"redhat-operators-d8s5f\" (UID: \"bcb79b86-65c0-44c1-9ce2-588ad66206a6\") " pod="openshift-marketplace/redhat-operators-d8s5f" Oct 06 10:57:46 crc kubenswrapper[4824]: I1006 10:57:46.125913 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bcb79b86-65c0-44c1-9ce2-588ad66206a6-utilities\") pod \"redhat-operators-d8s5f\" (UID: \"bcb79b86-65c0-44c1-9ce2-588ad66206a6\") " pod="openshift-marketplace/redhat-operators-d8s5f" Oct 06 10:57:46 crc kubenswrapper[4824]: I1006 10:57:46.148034 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vqgw8\" (UniqueName: \"kubernetes.io/projected/bcb79b86-65c0-44c1-9ce2-588ad66206a6-kube-api-access-vqgw8\") pod \"redhat-operators-d8s5f\" (UID: \"bcb79b86-65c0-44c1-9ce2-588ad66206a6\") " pod="openshift-marketplace/redhat-operators-d8s5f" Oct 06 10:57:46 crc kubenswrapper[4824]: I1006 10:57:46.291425 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-d8s5f" Oct 06 10:57:46 crc kubenswrapper[4824]: I1006 10:57:46.794630 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-d8s5f"] Oct 06 10:57:47 crc kubenswrapper[4824]: I1006 10:57:47.793564 4824 generic.go:334] "Generic (PLEG): container finished" podID="bcb79b86-65c0-44c1-9ce2-588ad66206a6" containerID="a91442ea8509109802e0fe1fedba31b78a7032ebbf9f895bfe9fa45c116aa4b2" exitCode=0 Oct 06 10:57:47 crc kubenswrapper[4824]: I1006 10:57:47.793650 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d8s5f" event={"ID":"bcb79b86-65c0-44c1-9ce2-588ad66206a6","Type":"ContainerDied","Data":"a91442ea8509109802e0fe1fedba31b78a7032ebbf9f895bfe9fa45c116aa4b2"} Oct 06 10:57:47 crc kubenswrapper[4824]: I1006 10:57:47.793960 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d8s5f" event={"ID":"bcb79b86-65c0-44c1-9ce2-588ad66206a6","Type":"ContainerStarted","Data":"3579207cd69b6f7baf1e789a4f31e36a0bb99d6896634712b677103f45576cd9"} Oct 06 10:57:47 crc kubenswrapper[4824]: I1006 10:57:47.798063 4824 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 06 10:57:54 crc kubenswrapper[4824]: I1006 10:57:54.885296 4824 generic.go:334] "Generic (PLEG): container finished" podID="7fd677dd-afa0-45a9-8198-ea78e02f4fb7" containerID="97e83913e74175b39ebaa5d81693fba451d2731ce38d61aae9ded9fa6eef80d9" exitCode=0 Oct 06 10:57:54 crc kubenswrapper[4824]: I1006 10:57:54.885397 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"7fd677dd-afa0-45a9-8198-ea78e02f4fb7","Type":"ContainerDied","Data":"97e83913e74175b39ebaa5d81693fba451d2731ce38d61aae9ded9fa6eef80d9"} Oct 06 10:57:55 crc kubenswrapper[4824]: I1006 10:57:55.897463 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d8s5f" event={"ID":"bcb79b86-65c0-44c1-9ce2-588ad66206a6","Type":"ContainerStarted","Data":"d5ee73cf46eef985237808074599c16ec84ce705bc1e742e08cbcf793641d88a"} Oct 06 10:57:56 crc kubenswrapper[4824]: I1006 10:57:56.406731 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Oct 06 10:57:56 crc kubenswrapper[4824]: I1006 10:57:56.448415 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/7fd677dd-afa0-45a9-8198-ea78e02f4fb7-test-operator-ephemeral-workdir\") pod \"7fd677dd-afa0-45a9-8198-ea78e02f4fb7\" (UID: \"7fd677dd-afa0-45a9-8198-ea78e02f4fb7\") " Oct 06 10:57:56 crc kubenswrapper[4824]: I1006 10:57:56.448583 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/7fd677dd-afa0-45a9-8198-ea78e02f4fb7-test-operator-ephemeral-temporary\") pod \"7fd677dd-afa0-45a9-8198-ea78e02f4fb7\" (UID: \"7fd677dd-afa0-45a9-8198-ea78e02f4fb7\") " Oct 06 10:57:56 crc kubenswrapper[4824]: I1006 10:57:56.448694 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/7fd677dd-afa0-45a9-8198-ea78e02f4fb7-ca-certs\") pod \"7fd677dd-afa0-45a9-8198-ea78e02f4fb7\" (UID: \"7fd677dd-afa0-45a9-8198-ea78e02f4fb7\") " Oct 06 10:57:56 crc kubenswrapper[4824]: I1006 10:57:56.448869 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"7fd677dd-afa0-45a9-8198-ea78e02f4fb7\" (UID: \"7fd677dd-afa0-45a9-8198-ea78e02f4fb7\") " Oct 06 10:57:56 crc kubenswrapper[4824]: I1006 10:57:56.449016 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8l7lg\" (UniqueName: \"kubernetes.io/projected/7fd677dd-afa0-45a9-8198-ea78e02f4fb7-kube-api-access-8l7lg\") pod \"7fd677dd-afa0-45a9-8198-ea78e02f4fb7\" (UID: \"7fd677dd-afa0-45a9-8198-ea78e02f4fb7\") " Oct 06 10:57:56 crc kubenswrapper[4824]: I1006 10:57:56.449082 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7fd677dd-afa0-45a9-8198-ea78e02f4fb7-config-data\") pod \"7fd677dd-afa0-45a9-8198-ea78e02f4fb7\" (UID: \"7fd677dd-afa0-45a9-8198-ea78e02f4fb7\") " Oct 06 10:57:56 crc kubenswrapper[4824]: I1006 10:57:56.450126 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/7fd677dd-afa0-45a9-8198-ea78e02f4fb7-openstack-config\") pod \"7fd677dd-afa0-45a9-8198-ea78e02f4fb7\" (UID: \"7fd677dd-afa0-45a9-8198-ea78e02f4fb7\") " Oct 06 10:57:56 crc kubenswrapper[4824]: I1006 10:57:56.450218 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7fd677dd-afa0-45a9-8198-ea78e02f4fb7-ssh-key\") pod \"7fd677dd-afa0-45a9-8198-ea78e02f4fb7\" (UID: \"7fd677dd-afa0-45a9-8198-ea78e02f4fb7\") " Oct 06 10:57:56 crc kubenswrapper[4824]: I1006 10:57:56.450249 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/7fd677dd-afa0-45a9-8198-ea78e02f4fb7-openstack-config-secret\") pod \"7fd677dd-afa0-45a9-8198-ea78e02f4fb7\" (UID: \"7fd677dd-afa0-45a9-8198-ea78e02f4fb7\") " Oct 06 10:57:56 crc kubenswrapper[4824]: I1006 10:57:56.452285 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7fd677dd-afa0-45a9-8198-ea78e02f4fb7-config-data" (OuterVolumeSpecName: "config-data") pod "7fd677dd-afa0-45a9-8198-ea78e02f4fb7" (UID: "7fd677dd-afa0-45a9-8198-ea78e02f4fb7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:57:56 crc kubenswrapper[4824]: I1006 10:57:56.452326 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7fd677dd-afa0-45a9-8198-ea78e02f4fb7-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "7fd677dd-afa0-45a9-8198-ea78e02f4fb7" (UID: "7fd677dd-afa0-45a9-8198-ea78e02f4fb7"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:57:56 crc kubenswrapper[4824]: I1006 10:57:56.458456 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7fd677dd-afa0-45a9-8198-ea78e02f4fb7-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "7fd677dd-afa0-45a9-8198-ea78e02f4fb7" (UID: "7fd677dd-afa0-45a9-8198-ea78e02f4fb7"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:57:56 crc kubenswrapper[4824]: I1006 10:57:56.458943 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "test-operator-logs") pod "7fd677dd-afa0-45a9-8198-ea78e02f4fb7" (UID: "7fd677dd-afa0-45a9-8198-ea78e02f4fb7"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 06 10:57:56 crc kubenswrapper[4824]: I1006 10:57:56.459038 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7fd677dd-afa0-45a9-8198-ea78e02f4fb7-kube-api-access-8l7lg" (OuterVolumeSpecName: "kube-api-access-8l7lg") pod "7fd677dd-afa0-45a9-8198-ea78e02f4fb7" (UID: "7fd677dd-afa0-45a9-8198-ea78e02f4fb7"). InnerVolumeSpecName "kube-api-access-8l7lg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:57:56 crc kubenswrapper[4824]: I1006 10:57:56.491513 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7fd677dd-afa0-45a9-8198-ea78e02f4fb7-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "7fd677dd-afa0-45a9-8198-ea78e02f4fb7" (UID: "7fd677dd-afa0-45a9-8198-ea78e02f4fb7"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:57:56 crc kubenswrapper[4824]: I1006 10:57:56.493378 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7fd677dd-afa0-45a9-8198-ea78e02f4fb7-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "7fd677dd-afa0-45a9-8198-ea78e02f4fb7" (UID: "7fd677dd-afa0-45a9-8198-ea78e02f4fb7"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:57:56 crc kubenswrapper[4824]: I1006 10:57:56.494898 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7fd677dd-afa0-45a9-8198-ea78e02f4fb7-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "7fd677dd-afa0-45a9-8198-ea78e02f4fb7" (UID: "7fd677dd-afa0-45a9-8198-ea78e02f4fb7"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:57:56 crc kubenswrapper[4824]: I1006 10:57:56.513800 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7fd677dd-afa0-45a9-8198-ea78e02f4fb7-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "7fd677dd-afa0-45a9-8198-ea78e02f4fb7" (UID: "7fd677dd-afa0-45a9-8198-ea78e02f4fb7"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:57:56 crc kubenswrapper[4824]: I1006 10:57:56.553902 4824 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7fd677dd-afa0-45a9-8198-ea78e02f4fb7-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 06 10:57:56 crc kubenswrapper[4824]: I1006 10:57:56.553949 4824 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/7fd677dd-afa0-45a9-8198-ea78e02f4fb7-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Oct 06 10:57:56 crc kubenswrapper[4824]: I1006 10:57:56.553965 4824 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/7fd677dd-afa0-45a9-8198-ea78e02f4fb7-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Oct 06 10:57:56 crc kubenswrapper[4824]: I1006 10:57:56.554030 4824 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/7fd677dd-afa0-45a9-8198-ea78e02f4fb7-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Oct 06 10:57:56 crc kubenswrapper[4824]: I1006 10:57:56.554046 4824 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/7fd677dd-afa0-45a9-8198-ea78e02f4fb7-ca-certs\") on node \"crc\" DevicePath \"\"" Oct 06 10:57:56 crc kubenswrapper[4824]: I1006 10:57:56.554092 4824 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Oct 06 10:57:56 crc kubenswrapper[4824]: I1006 10:57:56.554104 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8l7lg\" (UniqueName: \"kubernetes.io/projected/7fd677dd-afa0-45a9-8198-ea78e02f4fb7-kube-api-access-8l7lg\") on node \"crc\" DevicePath \"\"" Oct 06 10:57:56 crc kubenswrapper[4824]: I1006 10:57:56.554116 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7fd677dd-afa0-45a9-8198-ea78e02f4fb7-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 10:57:56 crc kubenswrapper[4824]: I1006 10:57:56.554125 4824 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/7fd677dd-afa0-45a9-8198-ea78e02f4fb7-openstack-config\") on node \"crc\" DevicePath \"\"" Oct 06 10:57:56 crc kubenswrapper[4824]: I1006 10:57:56.577518 4824 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Oct 06 10:57:56 crc kubenswrapper[4824]: I1006 10:57:56.656141 4824 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Oct 06 10:57:56 crc kubenswrapper[4824]: I1006 10:57:56.909044 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Oct 06 10:57:56 crc kubenswrapper[4824]: I1006 10:57:56.909034 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"7fd677dd-afa0-45a9-8198-ea78e02f4fb7","Type":"ContainerDied","Data":"3ac1dd564fcdc44b871406cb8ec2680f9e87759df1fc6c97e05f085c2b189d5d"} Oct 06 10:57:56 crc kubenswrapper[4824]: I1006 10:57:56.909122 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3ac1dd564fcdc44b871406cb8ec2680f9e87759df1fc6c97e05f085c2b189d5d" Oct 06 10:57:57 crc kubenswrapper[4824]: I1006 10:57:57.937784 4824 generic.go:334] "Generic (PLEG): container finished" podID="bcb79b86-65c0-44c1-9ce2-588ad66206a6" containerID="d5ee73cf46eef985237808074599c16ec84ce705bc1e742e08cbcf793641d88a" exitCode=0 Oct 06 10:57:57 crc kubenswrapper[4824]: I1006 10:57:57.937864 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d8s5f" event={"ID":"bcb79b86-65c0-44c1-9ce2-588ad66206a6","Type":"ContainerDied","Data":"d5ee73cf46eef985237808074599c16ec84ce705bc1e742e08cbcf793641d88a"} Oct 06 10:57:58 crc kubenswrapper[4824]: I1006 10:57:58.950543 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d8s5f" event={"ID":"bcb79b86-65c0-44c1-9ce2-588ad66206a6","Type":"ContainerStarted","Data":"3dbfa2d1cceff33c9796d627ecc7549b61179530e8740ee99dc94982bf5c2b01"} Oct 06 10:57:58 crc kubenswrapper[4824]: I1006 10:57:58.975327 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-d8s5f" podStartSLOduration=3.392537606 podStartE2EDuration="13.975294315s" podCreationTimestamp="2025-10-06 10:57:45 +0000 UTC" firstStartedPulling="2025-10-06 10:57:47.796805793 +0000 UTC m=+3577.161228664" lastFinishedPulling="2025-10-06 10:57:58.379562502 +0000 UTC m=+3587.743985373" observedRunningTime="2025-10-06 10:57:58.969066321 +0000 UTC m=+3588.333489192" watchObservedRunningTime="2025-10-06 10:57:58.975294315 +0000 UTC m=+3588.339717186" Oct 06 10:58:05 crc kubenswrapper[4824]: I1006 10:58:05.045647 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Oct 06 10:58:05 crc kubenswrapper[4824]: E1006 10:58:05.046616 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fd677dd-afa0-45a9-8198-ea78e02f4fb7" containerName="tempest-tests-tempest-tests-runner" Oct 06 10:58:05 crc kubenswrapper[4824]: I1006 10:58:05.046629 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fd677dd-afa0-45a9-8198-ea78e02f4fb7" containerName="tempest-tests-tempest-tests-runner" Oct 06 10:58:05 crc kubenswrapper[4824]: I1006 10:58:05.046842 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="7fd677dd-afa0-45a9-8198-ea78e02f4fb7" containerName="tempest-tests-tempest-tests-runner" Oct 06 10:58:05 crc kubenswrapper[4824]: I1006 10:58:05.047588 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 06 10:58:05 crc kubenswrapper[4824]: I1006 10:58:05.050477 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-5bmsf" Oct 06 10:58:05 crc kubenswrapper[4824]: I1006 10:58:05.059536 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Oct 06 10:58:05 crc kubenswrapper[4824]: I1006 10:58:05.146119 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"fbc3bf37-42b1-4a02-9922-2f8bad6f7ad5\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 06 10:58:05 crc kubenswrapper[4824]: I1006 10:58:05.146277 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m4dpv\" (UniqueName: \"kubernetes.io/projected/fbc3bf37-42b1-4a02-9922-2f8bad6f7ad5-kube-api-access-m4dpv\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"fbc3bf37-42b1-4a02-9922-2f8bad6f7ad5\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 06 10:58:05 crc kubenswrapper[4824]: I1006 10:58:05.249112 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m4dpv\" (UniqueName: \"kubernetes.io/projected/fbc3bf37-42b1-4a02-9922-2f8bad6f7ad5-kube-api-access-m4dpv\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"fbc3bf37-42b1-4a02-9922-2f8bad6f7ad5\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 06 10:58:05 crc kubenswrapper[4824]: I1006 10:58:05.249266 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"fbc3bf37-42b1-4a02-9922-2f8bad6f7ad5\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 06 10:58:05 crc kubenswrapper[4824]: I1006 10:58:05.249808 4824 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"fbc3bf37-42b1-4a02-9922-2f8bad6f7ad5\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 06 10:58:05 crc kubenswrapper[4824]: I1006 10:58:05.268411 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m4dpv\" (UniqueName: \"kubernetes.io/projected/fbc3bf37-42b1-4a02-9922-2f8bad6f7ad5-kube-api-access-m4dpv\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"fbc3bf37-42b1-4a02-9922-2f8bad6f7ad5\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 06 10:58:05 crc kubenswrapper[4824]: I1006 10:58:05.282315 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"fbc3bf37-42b1-4a02-9922-2f8bad6f7ad5\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 06 10:58:05 crc kubenswrapper[4824]: I1006 10:58:05.378945 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 06 10:58:05 crc kubenswrapper[4824]: I1006 10:58:05.661088 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Oct 06 10:58:06 crc kubenswrapper[4824]: I1006 10:58:06.032630 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"fbc3bf37-42b1-4a02-9922-2f8bad6f7ad5","Type":"ContainerStarted","Data":"91a7a4e9137cf763753293da6de9f7d662440d9948ac7ec31c864a87751bc640"} Oct 06 10:58:06 crc kubenswrapper[4824]: I1006 10:58:06.291878 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-d8s5f" Oct 06 10:58:06 crc kubenswrapper[4824]: I1006 10:58:06.291959 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-d8s5f" Oct 06 10:58:06 crc kubenswrapper[4824]: I1006 10:58:06.361872 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-d8s5f" Oct 06 10:58:07 crc kubenswrapper[4824]: I1006 10:58:07.048626 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"fbc3bf37-42b1-4a02-9922-2f8bad6f7ad5","Type":"ContainerStarted","Data":"86c8b6dd4e31177233112927fe6aefa937149afcc1c6be5b01a8adaec2035f0b"} Oct 06 10:58:07 crc kubenswrapper[4824]: I1006 10:58:07.095607 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" podStartSLOduration=1.052666516 podStartE2EDuration="2.095569302s" podCreationTimestamp="2025-10-06 10:58:05 +0000 UTC" firstStartedPulling="2025-10-06 10:58:05.67857518 +0000 UTC m=+3595.042998051" lastFinishedPulling="2025-10-06 10:58:06.721477966 +0000 UTC m=+3596.085900837" observedRunningTime="2025-10-06 10:58:07.07535422 +0000 UTC m=+3596.439777111" watchObservedRunningTime="2025-10-06 10:58:07.095569302 +0000 UTC m=+3596.459992183" Oct 06 10:58:07 crc kubenswrapper[4824]: I1006 10:58:07.128725 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-d8s5f" Oct 06 10:58:07 crc kubenswrapper[4824]: I1006 10:58:07.246278 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-d8s5f"] Oct 06 10:58:07 crc kubenswrapper[4824]: I1006 10:58:07.322873 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-txrmb"] Oct 06 10:58:07 crc kubenswrapper[4824]: I1006 10:58:07.323265 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-txrmb" podUID="5fe63731-833a-45c3-b4f0-73ddc05383e1" containerName="registry-server" containerID="cri-o://363a47a824f4f4cf7625979b822234ca2e0ecdcc86a4192924e3b14ce374e3e1" gracePeriod=2 Oct 06 10:58:07 crc kubenswrapper[4824]: I1006 10:58:07.914274 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-txrmb" Oct 06 10:58:08 crc kubenswrapper[4824]: I1006 10:58:08.012747 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5fe63731-833a-45c3-b4f0-73ddc05383e1-utilities\") pod \"5fe63731-833a-45c3-b4f0-73ddc05383e1\" (UID: \"5fe63731-833a-45c3-b4f0-73ddc05383e1\") " Oct 06 10:58:08 crc kubenswrapper[4824]: I1006 10:58:08.012929 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5fe63731-833a-45c3-b4f0-73ddc05383e1-catalog-content\") pod \"5fe63731-833a-45c3-b4f0-73ddc05383e1\" (UID: \"5fe63731-833a-45c3-b4f0-73ddc05383e1\") " Oct 06 10:58:08 crc kubenswrapper[4824]: I1006 10:58:08.013397 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w2lkd\" (UniqueName: \"kubernetes.io/projected/5fe63731-833a-45c3-b4f0-73ddc05383e1-kube-api-access-w2lkd\") pod \"5fe63731-833a-45c3-b4f0-73ddc05383e1\" (UID: \"5fe63731-833a-45c3-b4f0-73ddc05383e1\") " Oct 06 10:58:08 crc kubenswrapper[4824]: I1006 10:58:08.014469 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5fe63731-833a-45c3-b4f0-73ddc05383e1-utilities" (OuterVolumeSpecName: "utilities") pod "5fe63731-833a-45c3-b4f0-73ddc05383e1" (UID: "5fe63731-833a-45c3-b4f0-73ddc05383e1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:58:08 crc kubenswrapper[4824]: I1006 10:58:08.023819 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe63731-833a-45c3-b4f0-73ddc05383e1-kube-api-access-w2lkd" (OuterVolumeSpecName: "kube-api-access-w2lkd") pod "5fe63731-833a-45c3-b4f0-73ddc05383e1" (UID: "5fe63731-833a-45c3-b4f0-73ddc05383e1"). InnerVolumeSpecName "kube-api-access-w2lkd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:58:08 crc kubenswrapper[4824]: I1006 10:58:08.061538 4824 generic.go:334] "Generic (PLEG): container finished" podID="5fe63731-833a-45c3-b4f0-73ddc05383e1" containerID="363a47a824f4f4cf7625979b822234ca2e0ecdcc86a4192924e3b14ce374e3e1" exitCode=0 Oct 06 10:58:08 crc kubenswrapper[4824]: I1006 10:58:08.061600 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-txrmb" event={"ID":"5fe63731-833a-45c3-b4f0-73ddc05383e1","Type":"ContainerDied","Data":"363a47a824f4f4cf7625979b822234ca2e0ecdcc86a4192924e3b14ce374e3e1"} Oct 06 10:58:08 crc kubenswrapper[4824]: I1006 10:58:08.061660 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-txrmb" Oct 06 10:58:08 crc kubenswrapper[4824]: I1006 10:58:08.061686 4824 scope.go:117] "RemoveContainer" containerID="363a47a824f4f4cf7625979b822234ca2e0ecdcc86a4192924e3b14ce374e3e1" Oct 06 10:58:08 crc kubenswrapper[4824]: I1006 10:58:08.061671 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-txrmb" event={"ID":"5fe63731-833a-45c3-b4f0-73ddc05383e1","Type":"ContainerDied","Data":"93099c5b8c0ba07ad990ae7d135501f077a0493d7559f9bdc215e4e5e4172787"} Oct 06 10:58:08 crc kubenswrapper[4824]: I1006 10:58:08.096644 4824 scope.go:117] "RemoveContainer" containerID="4608802294a820026e0deae051a3356124536fdfe9af562cd07233f5f8546bc7" Oct 06 10:58:08 crc kubenswrapper[4824]: I1006 10:58:08.115889 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w2lkd\" (UniqueName: \"kubernetes.io/projected/5fe63731-833a-45c3-b4f0-73ddc05383e1-kube-api-access-w2lkd\") on node \"crc\" DevicePath \"\"" Oct 06 10:58:08 crc kubenswrapper[4824]: I1006 10:58:08.115923 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5fe63731-833a-45c3-b4f0-73ddc05383e1-utilities\") on node \"crc\" DevicePath \"\"" Oct 06 10:58:08 crc kubenswrapper[4824]: I1006 10:58:08.124272 4824 scope.go:117] "RemoveContainer" containerID="a5fad09e3dcaedd867384b2a523e9663c29845ea78983ea35cfcd6dce8583e2e" Oct 06 10:58:08 crc kubenswrapper[4824]: I1006 10:58:08.143102 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5fe63731-833a-45c3-b4f0-73ddc05383e1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5fe63731-833a-45c3-b4f0-73ddc05383e1" (UID: "5fe63731-833a-45c3-b4f0-73ddc05383e1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:58:08 crc kubenswrapper[4824]: I1006 10:58:08.169267 4824 scope.go:117] "RemoveContainer" containerID="363a47a824f4f4cf7625979b822234ca2e0ecdcc86a4192924e3b14ce374e3e1" Oct 06 10:58:08 crc kubenswrapper[4824]: E1006 10:58:08.169889 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"363a47a824f4f4cf7625979b822234ca2e0ecdcc86a4192924e3b14ce374e3e1\": container with ID starting with 363a47a824f4f4cf7625979b822234ca2e0ecdcc86a4192924e3b14ce374e3e1 not found: ID does not exist" containerID="363a47a824f4f4cf7625979b822234ca2e0ecdcc86a4192924e3b14ce374e3e1" Oct 06 10:58:08 crc kubenswrapper[4824]: I1006 10:58:08.169956 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"363a47a824f4f4cf7625979b822234ca2e0ecdcc86a4192924e3b14ce374e3e1"} err="failed to get container status \"363a47a824f4f4cf7625979b822234ca2e0ecdcc86a4192924e3b14ce374e3e1\": rpc error: code = NotFound desc = could not find container \"363a47a824f4f4cf7625979b822234ca2e0ecdcc86a4192924e3b14ce374e3e1\": container with ID starting with 363a47a824f4f4cf7625979b822234ca2e0ecdcc86a4192924e3b14ce374e3e1 not found: ID does not exist" Oct 06 10:58:08 crc kubenswrapper[4824]: I1006 10:58:08.170008 4824 scope.go:117] "RemoveContainer" containerID="4608802294a820026e0deae051a3356124536fdfe9af562cd07233f5f8546bc7" Oct 06 10:58:08 crc kubenswrapper[4824]: E1006 10:58:08.170634 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4608802294a820026e0deae051a3356124536fdfe9af562cd07233f5f8546bc7\": container with ID starting with 4608802294a820026e0deae051a3356124536fdfe9af562cd07233f5f8546bc7 not found: ID does not exist" containerID="4608802294a820026e0deae051a3356124536fdfe9af562cd07233f5f8546bc7" Oct 06 10:58:08 crc kubenswrapper[4824]: I1006 10:58:08.170687 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4608802294a820026e0deae051a3356124536fdfe9af562cd07233f5f8546bc7"} err="failed to get container status \"4608802294a820026e0deae051a3356124536fdfe9af562cd07233f5f8546bc7\": rpc error: code = NotFound desc = could not find container \"4608802294a820026e0deae051a3356124536fdfe9af562cd07233f5f8546bc7\": container with ID starting with 4608802294a820026e0deae051a3356124536fdfe9af562cd07233f5f8546bc7 not found: ID does not exist" Oct 06 10:58:08 crc kubenswrapper[4824]: I1006 10:58:08.170726 4824 scope.go:117] "RemoveContainer" containerID="a5fad09e3dcaedd867384b2a523e9663c29845ea78983ea35cfcd6dce8583e2e" Oct 06 10:58:08 crc kubenswrapper[4824]: E1006 10:58:08.171022 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a5fad09e3dcaedd867384b2a523e9663c29845ea78983ea35cfcd6dce8583e2e\": container with ID starting with a5fad09e3dcaedd867384b2a523e9663c29845ea78983ea35cfcd6dce8583e2e not found: ID does not exist" containerID="a5fad09e3dcaedd867384b2a523e9663c29845ea78983ea35cfcd6dce8583e2e" Oct 06 10:58:08 crc kubenswrapper[4824]: I1006 10:58:08.171056 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a5fad09e3dcaedd867384b2a523e9663c29845ea78983ea35cfcd6dce8583e2e"} err="failed to get container status \"a5fad09e3dcaedd867384b2a523e9663c29845ea78983ea35cfcd6dce8583e2e\": rpc error: code = NotFound desc = could not find container \"a5fad09e3dcaedd867384b2a523e9663c29845ea78983ea35cfcd6dce8583e2e\": container with ID starting with a5fad09e3dcaedd867384b2a523e9663c29845ea78983ea35cfcd6dce8583e2e not found: ID does not exist" Oct 06 10:58:08 crc kubenswrapper[4824]: I1006 10:58:08.218042 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5fe63731-833a-45c3-b4f0-73ddc05383e1-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 06 10:58:08 crc kubenswrapper[4824]: I1006 10:58:08.400463 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-txrmb"] Oct 06 10:58:08 crc kubenswrapper[4824]: I1006 10:58:08.411163 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-txrmb"] Oct 06 10:58:09 crc kubenswrapper[4824]: I1006 10:58:09.289762 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe63731-833a-45c3-b4f0-73ddc05383e1" path="/var/lib/kubelet/pods/5fe63731-833a-45c3-b4f0-73ddc05383e1/volumes" Oct 06 10:58:10 crc kubenswrapper[4824]: I1006 10:58:10.424223 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-kqn5z"] Oct 06 10:58:10 crc kubenswrapper[4824]: E1006 10:58:10.425248 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5fe63731-833a-45c3-b4f0-73ddc05383e1" containerName="extract-utilities" Oct 06 10:58:10 crc kubenswrapper[4824]: I1006 10:58:10.425270 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="5fe63731-833a-45c3-b4f0-73ddc05383e1" containerName="extract-utilities" Oct 06 10:58:10 crc kubenswrapper[4824]: E1006 10:58:10.425290 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5fe63731-833a-45c3-b4f0-73ddc05383e1" containerName="extract-content" Oct 06 10:58:10 crc kubenswrapper[4824]: I1006 10:58:10.425302 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="5fe63731-833a-45c3-b4f0-73ddc05383e1" containerName="extract-content" Oct 06 10:58:10 crc kubenswrapper[4824]: E1006 10:58:10.425357 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5fe63731-833a-45c3-b4f0-73ddc05383e1" containerName="registry-server" Oct 06 10:58:10 crc kubenswrapper[4824]: I1006 10:58:10.425369 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="5fe63731-833a-45c3-b4f0-73ddc05383e1" containerName="registry-server" Oct 06 10:58:10 crc kubenswrapper[4824]: I1006 10:58:10.425690 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="5fe63731-833a-45c3-b4f0-73ddc05383e1" containerName="registry-server" Oct 06 10:58:10 crc kubenswrapper[4824]: I1006 10:58:10.427936 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kqn5z" Oct 06 10:58:10 crc kubenswrapper[4824]: I1006 10:58:10.438312 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-kqn5z"] Oct 06 10:58:10 crc kubenswrapper[4824]: I1006 10:58:10.468535 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0d8759a1-8ed7-4455-b12e-8cab44ad09fa-catalog-content\") pod \"redhat-marketplace-kqn5z\" (UID: \"0d8759a1-8ed7-4455-b12e-8cab44ad09fa\") " pod="openshift-marketplace/redhat-marketplace-kqn5z" Oct 06 10:58:10 crc kubenswrapper[4824]: I1006 10:58:10.468611 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f89hn\" (UniqueName: \"kubernetes.io/projected/0d8759a1-8ed7-4455-b12e-8cab44ad09fa-kube-api-access-f89hn\") pod \"redhat-marketplace-kqn5z\" (UID: \"0d8759a1-8ed7-4455-b12e-8cab44ad09fa\") " pod="openshift-marketplace/redhat-marketplace-kqn5z" Oct 06 10:58:10 crc kubenswrapper[4824]: I1006 10:58:10.469062 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0d8759a1-8ed7-4455-b12e-8cab44ad09fa-utilities\") pod \"redhat-marketplace-kqn5z\" (UID: \"0d8759a1-8ed7-4455-b12e-8cab44ad09fa\") " pod="openshift-marketplace/redhat-marketplace-kqn5z" Oct 06 10:58:10 crc kubenswrapper[4824]: I1006 10:58:10.571803 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0d8759a1-8ed7-4455-b12e-8cab44ad09fa-catalog-content\") pod \"redhat-marketplace-kqn5z\" (UID: \"0d8759a1-8ed7-4455-b12e-8cab44ad09fa\") " pod="openshift-marketplace/redhat-marketplace-kqn5z" Oct 06 10:58:10 crc kubenswrapper[4824]: I1006 10:58:10.571892 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f89hn\" (UniqueName: \"kubernetes.io/projected/0d8759a1-8ed7-4455-b12e-8cab44ad09fa-kube-api-access-f89hn\") pod \"redhat-marketplace-kqn5z\" (UID: \"0d8759a1-8ed7-4455-b12e-8cab44ad09fa\") " pod="openshift-marketplace/redhat-marketplace-kqn5z" Oct 06 10:58:10 crc kubenswrapper[4824]: I1006 10:58:10.572059 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0d8759a1-8ed7-4455-b12e-8cab44ad09fa-utilities\") pod \"redhat-marketplace-kqn5z\" (UID: \"0d8759a1-8ed7-4455-b12e-8cab44ad09fa\") " pod="openshift-marketplace/redhat-marketplace-kqn5z" Oct 06 10:58:10 crc kubenswrapper[4824]: I1006 10:58:10.572510 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0d8759a1-8ed7-4455-b12e-8cab44ad09fa-catalog-content\") pod \"redhat-marketplace-kqn5z\" (UID: \"0d8759a1-8ed7-4455-b12e-8cab44ad09fa\") " pod="openshift-marketplace/redhat-marketplace-kqn5z" Oct 06 10:58:10 crc kubenswrapper[4824]: I1006 10:58:10.572653 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0d8759a1-8ed7-4455-b12e-8cab44ad09fa-utilities\") pod \"redhat-marketplace-kqn5z\" (UID: \"0d8759a1-8ed7-4455-b12e-8cab44ad09fa\") " pod="openshift-marketplace/redhat-marketplace-kqn5z" Oct 06 10:58:10 crc kubenswrapper[4824]: I1006 10:58:10.613129 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f89hn\" (UniqueName: \"kubernetes.io/projected/0d8759a1-8ed7-4455-b12e-8cab44ad09fa-kube-api-access-f89hn\") pod \"redhat-marketplace-kqn5z\" (UID: \"0d8759a1-8ed7-4455-b12e-8cab44ad09fa\") " pod="openshift-marketplace/redhat-marketplace-kqn5z" Oct 06 10:58:10 crc kubenswrapper[4824]: I1006 10:58:10.771931 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kqn5z" Oct 06 10:58:11 crc kubenswrapper[4824]: I1006 10:58:11.331613 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-kqn5z"] Oct 06 10:58:12 crc kubenswrapper[4824]: I1006 10:58:12.124276 4824 generic.go:334] "Generic (PLEG): container finished" podID="0d8759a1-8ed7-4455-b12e-8cab44ad09fa" containerID="f21a5ac877603b6b037f5c357fa67863b76a43ae1431d13a0530a682bf57aa8f" exitCode=0 Oct 06 10:58:12 crc kubenswrapper[4824]: I1006 10:58:12.124644 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kqn5z" event={"ID":"0d8759a1-8ed7-4455-b12e-8cab44ad09fa","Type":"ContainerDied","Data":"f21a5ac877603b6b037f5c357fa67863b76a43ae1431d13a0530a682bf57aa8f"} Oct 06 10:58:12 crc kubenswrapper[4824]: I1006 10:58:12.125067 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kqn5z" event={"ID":"0d8759a1-8ed7-4455-b12e-8cab44ad09fa","Type":"ContainerStarted","Data":"838d2a82302062ce619c3278f42737168aa1099eee16b771f7026edb7386758a"} Oct 06 10:58:13 crc kubenswrapper[4824]: I1006 10:58:13.139322 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kqn5z" event={"ID":"0d8759a1-8ed7-4455-b12e-8cab44ad09fa","Type":"ContainerStarted","Data":"abe91595c8c7be3608ff710549ec89d5517cfd0ff68f45c91ee200ba7a923206"} Oct 06 10:58:13 crc kubenswrapper[4824]: I1006 10:58:13.915776 4824 patch_prober.go:28] interesting pod/machine-config-daemon-khgzw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 10:58:13 crc kubenswrapper[4824]: I1006 10:58:13.915856 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 10:58:14 crc kubenswrapper[4824]: I1006 10:58:14.153580 4824 generic.go:334] "Generic (PLEG): container finished" podID="0d8759a1-8ed7-4455-b12e-8cab44ad09fa" containerID="abe91595c8c7be3608ff710549ec89d5517cfd0ff68f45c91ee200ba7a923206" exitCode=0 Oct 06 10:58:14 crc kubenswrapper[4824]: I1006 10:58:14.153638 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kqn5z" event={"ID":"0d8759a1-8ed7-4455-b12e-8cab44ad09fa","Type":"ContainerDied","Data":"abe91595c8c7be3608ff710549ec89d5517cfd0ff68f45c91ee200ba7a923206"} Oct 06 10:58:15 crc kubenswrapper[4824]: I1006 10:58:15.171119 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kqn5z" event={"ID":"0d8759a1-8ed7-4455-b12e-8cab44ad09fa","Type":"ContainerStarted","Data":"995b84b3ef1b2e2e41a7b3f7fd935f8c8aa01273dfaa3c0615e58ac6f56532b9"} Oct 06 10:58:15 crc kubenswrapper[4824]: I1006 10:58:15.202113 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-kqn5z" podStartSLOduration=2.663997957 podStartE2EDuration="5.202094637s" podCreationTimestamp="2025-10-06 10:58:10 +0000 UTC" firstStartedPulling="2025-10-06 10:58:12.127184147 +0000 UTC m=+3601.491607038" lastFinishedPulling="2025-10-06 10:58:14.665280857 +0000 UTC m=+3604.029703718" observedRunningTime="2025-10-06 10:58:15.201045751 +0000 UTC m=+3604.565468622" watchObservedRunningTime="2025-10-06 10:58:15.202094637 +0000 UTC m=+3604.566517498" Oct 06 10:58:20 crc kubenswrapper[4824]: I1006 10:58:20.772629 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-kqn5z" Oct 06 10:58:20 crc kubenswrapper[4824]: I1006 10:58:20.776413 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-kqn5z" Oct 06 10:58:20 crc kubenswrapper[4824]: I1006 10:58:20.859124 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-kqn5z" Oct 06 10:58:21 crc kubenswrapper[4824]: I1006 10:58:21.314877 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-kqn5z" Oct 06 10:58:21 crc kubenswrapper[4824]: I1006 10:58:21.821113 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-kqn5z"] Oct 06 10:58:23 crc kubenswrapper[4824]: I1006 10:58:23.270547 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-kqn5z" podUID="0d8759a1-8ed7-4455-b12e-8cab44ad09fa" containerName="registry-server" containerID="cri-o://995b84b3ef1b2e2e41a7b3f7fd935f8c8aa01273dfaa3c0615e58ac6f56532b9" gracePeriod=2 Oct 06 10:58:23 crc kubenswrapper[4824]: I1006 10:58:23.778013 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kqn5z" Oct 06 10:58:23 crc kubenswrapper[4824]: I1006 10:58:23.894741 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0d8759a1-8ed7-4455-b12e-8cab44ad09fa-catalog-content\") pod \"0d8759a1-8ed7-4455-b12e-8cab44ad09fa\" (UID: \"0d8759a1-8ed7-4455-b12e-8cab44ad09fa\") " Oct 06 10:58:23 crc kubenswrapper[4824]: I1006 10:58:23.895018 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f89hn\" (UniqueName: \"kubernetes.io/projected/0d8759a1-8ed7-4455-b12e-8cab44ad09fa-kube-api-access-f89hn\") pod \"0d8759a1-8ed7-4455-b12e-8cab44ad09fa\" (UID: \"0d8759a1-8ed7-4455-b12e-8cab44ad09fa\") " Oct 06 10:58:23 crc kubenswrapper[4824]: I1006 10:58:23.895052 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0d8759a1-8ed7-4455-b12e-8cab44ad09fa-utilities\") pod \"0d8759a1-8ed7-4455-b12e-8cab44ad09fa\" (UID: \"0d8759a1-8ed7-4455-b12e-8cab44ad09fa\") " Oct 06 10:58:23 crc kubenswrapper[4824]: I1006 10:58:23.896155 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0d8759a1-8ed7-4455-b12e-8cab44ad09fa-utilities" (OuterVolumeSpecName: "utilities") pod "0d8759a1-8ed7-4455-b12e-8cab44ad09fa" (UID: "0d8759a1-8ed7-4455-b12e-8cab44ad09fa"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:58:23 crc kubenswrapper[4824]: I1006 10:58:23.902421 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0d8759a1-8ed7-4455-b12e-8cab44ad09fa-kube-api-access-f89hn" (OuterVolumeSpecName: "kube-api-access-f89hn") pod "0d8759a1-8ed7-4455-b12e-8cab44ad09fa" (UID: "0d8759a1-8ed7-4455-b12e-8cab44ad09fa"). InnerVolumeSpecName "kube-api-access-f89hn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:58:23 crc kubenswrapper[4824]: I1006 10:58:23.915298 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0d8759a1-8ed7-4455-b12e-8cab44ad09fa-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0d8759a1-8ed7-4455-b12e-8cab44ad09fa" (UID: "0d8759a1-8ed7-4455-b12e-8cab44ad09fa"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:58:23 crc kubenswrapper[4824]: I1006 10:58:23.997770 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f89hn\" (UniqueName: \"kubernetes.io/projected/0d8759a1-8ed7-4455-b12e-8cab44ad09fa-kube-api-access-f89hn\") on node \"crc\" DevicePath \"\"" Oct 06 10:58:23 crc kubenswrapper[4824]: I1006 10:58:23.997813 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0d8759a1-8ed7-4455-b12e-8cab44ad09fa-utilities\") on node \"crc\" DevicePath \"\"" Oct 06 10:58:23 crc kubenswrapper[4824]: I1006 10:58:23.997825 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0d8759a1-8ed7-4455-b12e-8cab44ad09fa-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 06 10:58:24 crc kubenswrapper[4824]: I1006 10:58:24.283038 4824 generic.go:334] "Generic (PLEG): container finished" podID="0d8759a1-8ed7-4455-b12e-8cab44ad09fa" containerID="995b84b3ef1b2e2e41a7b3f7fd935f8c8aa01273dfaa3c0615e58ac6f56532b9" exitCode=0 Oct 06 10:58:24 crc kubenswrapper[4824]: I1006 10:58:24.283101 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kqn5z" event={"ID":"0d8759a1-8ed7-4455-b12e-8cab44ad09fa","Type":"ContainerDied","Data":"995b84b3ef1b2e2e41a7b3f7fd935f8c8aa01273dfaa3c0615e58ac6f56532b9"} Oct 06 10:58:24 crc kubenswrapper[4824]: I1006 10:58:24.283156 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kqn5z" event={"ID":"0d8759a1-8ed7-4455-b12e-8cab44ad09fa","Type":"ContainerDied","Data":"838d2a82302062ce619c3278f42737168aa1099eee16b771f7026edb7386758a"} Oct 06 10:58:24 crc kubenswrapper[4824]: I1006 10:58:24.283153 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kqn5z" Oct 06 10:58:24 crc kubenswrapper[4824]: I1006 10:58:24.283174 4824 scope.go:117] "RemoveContainer" containerID="995b84b3ef1b2e2e41a7b3f7fd935f8c8aa01273dfaa3c0615e58ac6f56532b9" Oct 06 10:58:24 crc kubenswrapper[4824]: I1006 10:58:24.336079 4824 scope.go:117] "RemoveContainer" containerID="abe91595c8c7be3608ff710549ec89d5517cfd0ff68f45c91ee200ba7a923206" Oct 06 10:58:24 crc kubenswrapper[4824]: I1006 10:58:24.351506 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-kqn5z"] Oct 06 10:58:24 crc kubenswrapper[4824]: I1006 10:58:24.359693 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-kqn5z"] Oct 06 10:58:24 crc kubenswrapper[4824]: I1006 10:58:24.369837 4824 scope.go:117] "RemoveContainer" containerID="f21a5ac877603b6b037f5c357fa67863b76a43ae1431d13a0530a682bf57aa8f" Oct 06 10:58:24 crc kubenswrapper[4824]: I1006 10:58:24.402194 4824 scope.go:117] "RemoveContainer" containerID="995b84b3ef1b2e2e41a7b3f7fd935f8c8aa01273dfaa3c0615e58ac6f56532b9" Oct 06 10:58:24 crc kubenswrapper[4824]: E1006 10:58:24.402679 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"995b84b3ef1b2e2e41a7b3f7fd935f8c8aa01273dfaa3c0615e58ac6f56532b9\": container with ID starting with 995b84b3ef1b2e2e41a7b3f7fd935f8c8aa01273dfaa3c0615e58ac6f56532b9 not found: ID does not exist" containerID="995b84b3ef1b2e2e41a7b3f7fd935f8c8aa01273dfaa3c0615e58ac6f56532b9" Oct 06 10:58:24 crc kubenswrapper[4824]: I1006 10:58:24.402714 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"995b84b3ef1b2e2e41a7b3f7fd935f8c8aa01273dfaa3c0615e58ac6f56532b9"} err="failed to get container status \"995b84b3ef1b2e2e41a7b3f7fd935f8c8aa01273dfaa3c0615e58ac6f56532b9\": rpc error: code = NotFound desc = could not find container \"995b84b3ef1b2e2e41a7b3f7fd935f8c8aa01273dfaa3c0615e58ac6f56532b9\": container with ID starting with 995b84b3ef1b2e2e41a7b3f7fd935f8c8aa01273dfaa3c0615e58ac6f56532b9 not found: ID does not exist" Oct 06 10:58:24 crc kubenswrapper[4824]: I1006 10:58:24.402738 4824 scope.go:117] "RemoveContainer" containerID="abe91595c8c7be3608ff710549ec89d5517cfd0ff68f45c91ee200ba7a923206" Oct 06 10:58:24 crc kubenswrapper[4824]: E1006 10:58:24.402949 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"abe91595c8c7be3608ff710549ec89d5517cfd0ff68f45c91ee200ba7a923206\": container with ID starting with abe91595c8c7be3608ff710549ec89d5517cfd0ff68f45c91ee200ba7a923206 not found: ID does not exist" containerID="abe91595c8c7be3608ff710549ec89d5517cfd0ff68f45c91ee200ba7a923206" Oct 06 10:58:24 crc kubenswrapper[4824]: I1006 10:58:24.402992 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"abe91595c8c7be3608ff710549ec89d5517cfd0ff68f45c91ee200ba7a923206"} err="failed to get container status \"abe91595c8c7be3608ff710549ec89d5517cfd0ff68f45c91ee200ba7a923206\": rpc error: code = NotFound desc = could not find container \"abe91595c8c7be3608ff710549ec89d5517cfd0ff68f45c91ee200ba7a923206\": container with ID starting with abe91595c8c7be3608ff710549ec89d5517cfd0ff68f45c91ee200ba7a923206 not found: ID does not exist" Oct 06 10:58:24 crc kubenswrapper[4824]: I1006 10:58:24.403014 4824 scope.go:117] "RemoveContainer" containerID="f21a5ac877603b6b037f5c357fa67863b76a43ae1431d13a0530a682bf57aa8f" Oct 06 10:58:24 crc kubenswrapper[4824]: E1006 10:58:24.403569 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f21a5ac877603b6b037f5c357fa67863b76a43ae1431d13a0530a682bf57aa8f\": container with ID starting with f21a5ac877603b6b037f5c357fa67863b76a43ae1431d13a0530a682bf57aa8f not found: ID does not exist" containerID="f21a5ac877603b6b037f5c357fa67863b76a43ae1431d13a0530a682bf57aa8f" Oct 06 10:58:24 crc kubenswrapper[4824]: I1006 10:58:24.403595 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f21a5ac877603b6b037f5c357fa67863b76a43ae1431d13a0530a682bf57aa8f"} err="failed to get container status \"f21a5ac877603b6b037f5c357fa67863b76a43ae1431d13a0530a682bf57aa8f\": rpc error: code = NotFound desc = could not find container \"f21a5ac877603b6b037f5c357fa67863b76a43ae1431d13a0530a682bf57aa8f\": container with ID starting with f21a5ac877603b6b037f5c357fa67863b76a43ae1431d13a0530a682bf57aa8f not found: ID does not exist" Oct 06 10:58:24 crc kubenswrapper[4824]: I1006 10:58:24.720713 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-v7b5k/must-gather-kg2w2"] Oct 06 10:58:24 crc kubenswrapper[4824]: E1006 10:58:24.722369 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d8759a1-8ed7-4455-b12e-8cab44ad09fa" containerName="extract-utilities" Oct 06 10:58:24 crc kubenswrapper[4824]: I1006 10:58:24.722393 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d8759a1-8ed7-4455-b12e-8cab44ad09fa" containerName="extract-utilities" Oct 06 10:58:24 crc kubenswrapper[4824]: E1006 10:58:24.722429 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d8759a1-8ed7-4455-b12e-8cab44ad09fa" containerName="extract-content" Oct 06 10:58:24 crc kubenswrapper[4824]: I1006 10:58:24.722438 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d8759a1-8ed7-4455-b12e-8cab44ad09fa" containerName="extract-content" Oct 06 10:58:24 crc kubenswrapper[4824]: E1006 10:58:24.722515 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d8759a1-8ed7-4455-b12e-8cab44ad09fa" containerName="registry-server" Oct 06 10:58:24 crc kubenswrapper[4824]: I1006 10:58:24.722525 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d8759a1-8ed7-4455-b12e-8cab44ad09fa" containerName="registry-server" Oct 06 10:58:24 crc kubenswrapper[4824]: I1006 10:58:24.723093 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d8759a1-8ed7-4455-b12e-8cab44ad09fa" containerName="registry-server" Oct 06 10:58:24 crc kubenswrapper[4824]: I1006 10:58:24.732228 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-v7b5k/must-gather-kg2w2" Oct 06 10:58:24 crc kubenswrapper[4824]: I1006 10:58:24.737707 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-v7b5k"/"kube-root-ca.crt" Oct 06 10:58:24 crc kubenswrapper[4824]: I1006 10:58:24.740259 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-v7b5k"/"openshift-service-ca.crt" Oct 06 10:58:24 crc kubenswrapper[4824]: I1006 10:58:24.740655 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-v7b5k"/"default-dockercfg-nqdd8" Oct 06 10:58:24 crc kubenswrapper[4824]: I1006 10:58:24.769467 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-v7b5k/must-gather-kg2w2"] Oct 06 10:58:24 crc kubenswrapper[4824]: I1006 10:58:24.860641 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4gf44\" (UniqueName: \"kubernetes.io/projected/2a18f37a-4010-4742-8811-a2fe992fe4f1-kube-api-access-4gf44\") pod \"must-gather-kg2w2\" (UID: \"2a18f37a-4010-4742-8811-a2fe992fe4f1\") " pod="openshift-must-gather-v7b5k/must-gather-kg2w2" Oct 06 10:58:24 crc kubenswrapper[4824]: I1006 10:58:24.860729 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/2a18f37a-4010-4742-8811-a2fe992fe4f1-must-gather-output\") pod \"must-gather-kg2w2\" (UID: \"2a18f37a-4010-4742-8811-a2fe992fe4f1\") " pod="openshift-must-gather-v7b5k/must-gather-kg2w2" Oct 06 10:58:24 crc kubenswrapper[4824]: I1006 10:58:24.962195 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4gf44\" (UniqueName: \"kubernetes.io/projected/2a18f37a-4010-4742-8811-a2fe992fe4f1-kube-api-access-4gf44\") pod \"must-gather-kg2w2\" (UID: \"2a18f37a-4010-4742-8811-a2fe992fe4f1\") " pod="openshift-must-gather-v7b5k/must-gather-kg2w2" Oct 06 10:58:24 crc kubenswrapper[4824]: I1006 10:58:24.962335 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/2a18f37a-4010-4742-8811-a2fe992fe4f1-must-gather-output\") pod \"must-gather-kg2w2\" (UID: \"2a18f37a-4010-4742-8811-a2fe992fe4f1\") " pod="openshift-must-gather-v7b5k/must-gather-kg2w2" Oct 06 10:58:24 crc kubenswrapper[4824]: I1006 10:58:24.962787 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/2a18f37a-4010-4742-8811-a2fe992fe4f1-must-gather-output\") pod \"must-gather-kg2w2\" (UID: \"2a18f37a-4010-4742-8811-a2fe992fe4f1\") " pod="openshift-must-gather-v7b5k/must-gather-kg2w2" Oct 06 10:58:24 crc kubenswrapper[4824]: I1006 10:58:24.979725 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4gf44\" (UniqueName: \"kubernetes.io/projected/2a18f37a-4010-4742-8811-a2fe992fe4f1-kube-api-access-4gf44\") pod \"must-gather-kg2w2\" (UID: \"2a18f37a-4010-4742-8811-a2fe992fe4f1\") " pod="openshift-must-gather-v7b5k/must-gather-kg2w2" Oct 06 10:58:25 crc kubenswrapper[4824]: I1006 10:58:25.057457 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-v7b5k/must-gather-kg2w2" Oct 06 10:58:25 crc kubenswrapper[4824]: I1006 10:58:25.295671 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0d8759a1-8ed7-4455-b12e-8cab44ad09fa" path="/var/lib/kubelet/pods/0d8759a1-8ed7-4455-b12e-8cab44ad09fa/volumes" Oct 06 10:58:25 crc kubenswrapper[4824]: I1006 10:58:25.552288 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-v7b5k/must-gather-kg2w2"] Oct 06 10:58:25 crc kubenswrapper[4824]: W1006 10:58:25.562695 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2a18f37a_4010_4742_8811_a2fe992fe4f1.slice/crio-0e74a4361d20b73c4b0d793129aaa7a67d80c78d32d72f10da33046430c3d7e2 WatchSource:0}: Error finding container 0e74a4361d20b73c4b0d793129aaa7a67d80c78d32d72f10da33046430c3d7e2: Status 404 returned error can't find the container with id 0e74a4361d20b73c4b0d793129aaa7a67d80c78d32d72f10da33046430c3d7e2 Oct 06 10:58:26 crc kubenswrapper[4824]: I1006 10:58:26.314755 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-v7b5k/must-gather-kg2w2" event={"ID":"2a18f37a-4010-4742-8811-a2fe992fe4f1","Type":"ContainerStarted","Data":"0e74a4361d20b73c4b0d793129aaa7a67d80c78d32d72f10da33046430c3d7e2"} Oct 06 10:58:30 crc kubenswrapper[4824]: I1006 10:58:30.364851 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-v7b5k/must-gather-kg2w2" event={"ID":"2a18f37a-4010-4742-8811-a2fe992fe4f1","Type":"ContainerStarted","Data":"e1bed53b164b7508a7a0ac7d00bc63f9c58c23644ccfa0d031e8357cff0078dc"} Oct 06 10:58:30 crc kubenswrapper[4824]: I1006 10:58:30.365431 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-v7b5k/must-gather-kg2w2" event={"ID":"2a18f37a-4010-4742-8811-a2fe992fe4f1","Type":"ContainerStarted","Data":"a97e4ecc6d64ffef527b64c876826b3236ed56d7ed64f580f35f4e350b2b7331"} Oct 06 10:58:30 crc kubenswrapper[4824]: I1006 10:58:30.392435 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-v7b5k/must-gather-kg2w2" podStartSLOduration=2.567387553 podStartE2EDuration="6.392401433s" podCreationTimestamp="2025-10-06 10:58:24 +0000 UTC" firstStartedPulling="2025-10-06 10:58:25.564126961 +0000 UTC m=+3614.928549812" lastFinishedPulling="2025-10-06 10:58:29.389140841 +0000 UTC m=+3618.753563692" observedRunningTime="2025-10-06 10:58:30.385780648 +0000 UTC m=+3619.750203549" watchObservedRunningTime="2025-10-06 10:58:30.392401433 +0000 UTC m=+3619.756824324" Oct 06 10:58:34 crc kubenswrapper[4824]: I1006 10:58:34.345767 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-v7b5k/crc-debug-vbchb"] Oct 06 10:58:34 crc kubenswrapper[4824]: I1006 10:58:34.347389 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-v7b5k/crc-debug-vbchb" Oct 06 10:58:34 crc kubenswrapper[4824]: I1006 10:58:34.469735 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/fa6ffdef-9f65-4e2c-ae14-24ffdb8ceff9-host\") pod \"crc-debug-vbchb\" (UID: \"fa6ffdef-9f65-4e2c-ae14-24ffdb8ceff9\") " pod="openshift-must-gather-v7b5k/crc-debug-vbchb" Oct 06 10:58:34 crc kubenswrapper[4824]: I1006 10:58:34.470045 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rtfvs\" (UniqueName: \"kubernetes.io/projected/fa6ffdef-9f65-4e2c-ae14-24ffdb8ceff9-kube-api-access-rtfvs\") pod \"crc-debug-vbchb\" (UID: \"fa6ffdef-9f65-4e2c-ae14-24ffdb8ceff9\") " pod="openshift-must-gather-v7b5k/crc-debug-vbchb" Oct 06 10:58:34 crc kubenswrapper[4824]: I1006 10:58:34.572464 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/fa6ffdef-9f65-4e2c-ae14-24ffdb8ceff9-host\") pod \"crc-debug-vbchb\" (UID: \"fa6ffdef-9f65-4e2c-ae14-24ffdb8ceff9\") " pod="openshift-must-gather-v7b5k/crc-debug-vbchb" Oct 06 10:58:34 crc kubenswrapper[4824]: I1006 10:58:34.572600 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rtfvs\" (UniqueName: \"kubernetes.io/projected/fa6ffdef-9f65-4e2c-ae14-24ffdb8ceff9-kube-api-access-rtfvs\") pod \"crc-debug-vbchb\" (UID: \"fa6ffdef-9f65-4e2c-ae14-24ffdb8ceff9\") " pod="openshift-must-gather-v7b5k/crc-debug-vbchb" Oct 06 10:58:34 crc kubenswrapper[4824]: I1006 10:58:34.573246 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/fa6ffdef-9f65-4e2c-ae14-24ffdb8ceff9-host\") pod \"crc-debug-vbchb\" (UID: \"fa6ffdef-9f65-4e2c-ae14-24ffdb8ceff9\") " pod="openshift-must-gather-v7b5k/crc-debug-vbchb" Oct 06 10:58:34 crc kubenswrapper[4824]: I1006 10:58:34.598162 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rtfvs\" (UniqueName: \"kubernetes.io/projected/fa6ffdef-9f65-4e2c-ae14-24ffdb8ceff9-kube-api-access-rtfvs\") pod \"crc-debug-vbchb\" (UID: \"fa6ffdef-9f65-4e2c-ae14-24ffdb8ceff9\") " pod="openshift-must-gather-v7b5k/crc-debug-vbchb" Oct 06 10:58:34 crc kubenswrapper[4824]: I1006 10:58:34.679174 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-v7b5k/crc-debug-vbchb" Oct 06 10:58:35 crc kubenswrapper[4824]: I1006 10:58:35.427263 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-v7b5k/crc-debug-vbchb" event={"ID":"fa6ffdef-9f65-4e2c-ae14-24ffdb8ceff9","Type":"ContainerStarted","Data":"8e88bcf55188658933bd0c65c684146db3e8c654d242763c6d30b5ef8d6abff1"} Oct 06 10:58:43 crc kubenswrapper[4824]: I1006 10:58:43.915498 4824 patch_prober.go:28] interesting pod/machine-config-daemon-khgzw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 10:58:43 crc kubenswrapper[4824]: I1006 10:58:43.916140 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 10:58:43 crc kubenswrapper[4824]: I1006 10:58:43.916187 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" Oct 06 10:58:43 crc kubenswrapper[4824]: I1006 10:58:43.917118 4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d960e8d823dfd2a405439bf2eac79b76e3ee46509993c3394eac4ba53ab10143"} pod="openshift-machine-config-operator/machine-config-daemon-khgzw" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 06 10:58:43 crc kubenswrapper[4824]: I1006 10:58:43.917180 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" containerName="machine-config-daemon" containerID="cri-o://d960e8d823dfd2a405439bf2eac79b76e3ee46509993c3394eac4ba53ab10143" gracePeriod=600 Oct 06 10:58:44 crc kubenswrapper[4824]: I1006 10:58:44.530250 4824 generic.go:334] "Generic (PLEG): container finished" podID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" containerID="d960e8d823dfd2a405439bf2eac79b76e3ee46509993c3394eac4ba53ab10143" exitCode=0 Oct 06 10:58:44 crc kubenswrapper[4824]: I1006 10:58:44.530310 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" event={"ID":"e1140c8c-93f1-4fce-8c06-a69561ff8a6d","Type":"ContainerDied","Data":"d960e8d823dfd2a405439bf2eac79b76e3ee46509993c3394eac4ba53ab10143"} Oct 06 10:58:44 crc kubenswrapper[4824]: I1006 10:58:44.530384 4824 scope.go:117] "RemoveContainer" containerID="bbb34b884429016efb41d11fc1bb14bdd15d27ec1a282254f5cbaae186b1d1ef" Oct 06 10:58:46 crc kubenswrapper[4824]: I1006 10:58:46.550344 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" event={"ID":"e1140c8c-93f1-4fce-8c06-a69561ff8a6d","Type":"ContainerStarted","Data":"5a47a5d2e5b1ba0ad6ad09ca0bd7f5568851664618ce5ad15f9b3204fecd76c5"} Oct 06 10:58:46 crc kubenswrapper[4824]: I1006 10:58:46.553360 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-v7b5k/crc-debug-vbchb" event={"ID":"fa6ffdef-9f65-4e2c-ae14-24ffdb8ceff9","Type":"ContainerStarted","Data":"22fdcfb9df44916a40b6ce06b4258b25582a99185978d15eb47c665ed73431a1"} Oct 06 10:58:46 crc kubenswrapper[4824]: I1006 10:58:46.584183 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-v7b5k/crc-debug-vbchb" podStartSLOduration=1.185147141 podStartE2EDuration="12.584160573s" podCreationTimestamp="2025-10-06 10:58:34 +0000 UTC" firstStartedPulling="2025-10-06 10:58:34.72754614 +0000 UTC m=+3624.091969001" lastFinishedPulling="2025-10-06 10:58:46.126559572 +0000 UTC m=+3635.490982433" observedRunningTime="2025-10-06 10:58:46.579730073 +0000 UTC m=+3635.944152934" watchObservedRunningTime="2025-10-06 10:58:46.584160573 +0000 UTC m=+3635.948583434" Oct 06 10:59:42 crc kubenswrapper[4824]: I1006 10:59:42.187339 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-78c986c6fb-bzcht_b8a8eafe-1a7a-4eca-b61c-2d744512a54b/barbican-api-log/0.log" Oct 06 10:59:42 crc kubenswrapper[4824]: I1006 10:59:42.202376 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-78c986c6fb-bzcht_b8a8eafe-1a7a-4eca-b61c-2d744512a54b/barbican-api/0.log" Oct 06 10:59:42 crc kubenswrapper[4824]: I1006 10:59:42.475865 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-67979769c4-7fg5z_5e844107-f8f2-48e8-9731-84eff0f8c752/barbican-keystone-listener/0.log" Oct 06 10:59:42 crc kubenswrapper[4824]: I1006 10:59:42.492259 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-67979769c4-7fg5z_5e844107-f8f2-48e8-9731-84eff0f8c752/barbican-keystone-listener-log/0.log" Oct 06 10:59:42 crc kubenswrapper[4824]: I1006 10:59:42.689300 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-659dd56d6c-qpxtd_fb46f73f-9d47-4189-990a-051fed587851/barbican-worker/0.log" Oct 06 10:59:42 crc kubenswrapper[4824]: I1006 10:59:42.742498 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-659dd56d6c-qpxtd_fb46f73f-9d47-4189-990a-051fed587851/barbican-worker-log/0.log" Oct 06 10:59:42 crc kubenswrapper[4824]: I1006 10:59:42.990803 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-mlckl_c995dce1-6fa5-456b-b984-b397dcc9f9dc/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Oct 06 10:59:43 crc kubenswrapper[4824]: I1006 10:59:43.237815 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_369bbc94-27b7-4016-9a43-1e20b28f4323/ceilometer-central-agent/0.log" Oct 06 10:59:43 crc kubenswrapper[4824]: I1006 10:59:43.263224 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_369bbc94-27b7-4016-9a43-1e20b28f4323/ceilometer-notification-agent/0.log" Oct 06 10:59:43 crc kubenswrapper[4824]: I1006 10:59:43.308528 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_369bbc94-27b7-4016-9a43-1e20b28f4323/proxy-httpd/0.log" Oct 06 10:59:43 crc kubenswrapper[4824]: I1006 10:59:43.463603 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_369bbc94-27b7-4016-9a43-1e20b28f4323/sg-core/0.log" Oct 06 10:59:43 crc kubenswrapper[4824]: I1006 10:59:43.635263 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_13a74e55-383d-4105-9c9c-aafb63b75c39/cinder-api/0.log" Oct 06 10:59:43 crc kubenswrapper[4824]: I1006 10:59:43.702905 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_13a74e55-383d-4105-9c9c-aafb63b75c39/cinder-api-log/0.log" Oct 06 10:59:43 crc kubenswrapper[4824]: I1006 10:59:43.911903 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_18194590-7963-41e8-9253-dcf486d2cea4/cinder-scheduler/0.log" Oct 06 10:59:44 crc kubenswrapper[4824]: I1006 10:59:44.020580 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_18194590-7963-41e8-9253-dcf486d2cea4/probe/0.log" Oct 06 10:59:44 crc kubenswrapper[4824]: I1006 10:59:44.221296 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-xtlm8_1961267a-e5c0-469e-9f0e-9d4edbc8e64c/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Oct 06 10:59:44 crc kubenswrapper[4824]: I1006 10:59:44.432166 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-nm76g_87320828-94da-481c-b903-e7d478e3df65/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 06 10:59:44 crc kubenswrapper[4824]: I1006 10:59:44.533439 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-shvtw_3a56b28d-58ae-4cd4-a06a-b942a9365de7/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 06 10:59:44 crc kubenswrapper[4824]: I1006 10:59:44.715904 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-cb6ffcf87-xt4l8_7648fd80-2fcf-4ddd-bb32-1dd33c86a0fb/init/0.log" Oct 06 10:59:44 crc kubenswrapper[4824]: I1006 10:59:44.982697 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-cb6ffcf87-xt4l8_7648fd80-2fcf-4ddd-bb32-1dd33c86a0fb/init/0.log" Oct 06 10:59:45 crc kubenswrapper[4824]: I1006 10:59:45.006234 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-cb6ffcf87-xt4l8_7648fd80-2fcf-4ddd-bb32-1dd33c86a0fb/dnsmasq-dns/0.log" Oct 06 10:59:45 crc kubenswrapper[4824]: I1006 10:59:45.129843 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-kcvhh_8ea3d60c-80c7-4163-9f81-c0ec20e758c9/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Oct 06 10:59:45 crc kubenswrapper[4824]: I1006 10:59:45.310350 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_4b875c3d-d6b9-47c6-ae58-4ba7b78fb338/glance-log/0.log" Oct 06 10:59:45 crc kubenswrapper[4824]: I1006 10:59:45.325782 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_4b875c3d-d6b9-47c6-ae58-4ba7b78fb338/glance-httpd/0.log" Oct 06 10:59:45 crc kubenswrapper[4824]: I1006 10:59:45.509134 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_7d31f37b-142f-4f46-962a-e806e3d1269c/glance-httpd/0.log" Oct 06 10:59:45 crc kubenswrapper[4824]: I1006 10:59:45.595223 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_7d31f37b-142f-4f46-962a-e806e3d1269c/glance-log/0.log" Oct 06 10:59:45 crc kubenswrapper[4824]: I1006 10:59:45.822593 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-68fdc5dd48-gfq4d_194bef17-98ac-413f-a91e-0abc9835f2f6/horizon/0.log" Oct 06 10:59:45 crc kubenswrapper[4824]: I1006 10:59:45.916168 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-jqgjd_54e35b51-aeb1-4e2f-9be0-11230cb89fbe/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Oct 06 10:59:46 crc kubenswrapper[4824]: I1006 10:59:46.055179 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-68fdc5dd48-gfq4d_194bef17-98ac-413f-a91e-0abc9835f2f6/horizon-log/0.log" Oct 06 10:59:46 crc kubenswrapper[4824]: I1006 10:59:46.132494 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-k82nk_7db64f14-df4b-4519-81c5-f4e03e053925/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 06 10:59:46 crc kubenswrapper[4824]: I1006 10:59:46.405405 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_58776586-f2ad-4b0e-bf7e-303324ffe7f2/kube-state-metrics/0.log" Oct 06 10:59:46 crc kubenswrapper[4824]: I1006 10:59:46.510031 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-fbf8cb574-lrj2t_1015a252-b07f-4398-9423-41daf44c33b6/keystone-api/0.log" Oct 06 10:59:46 crc kubenswrapper[4824]: I1006 10:59:46.617838 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-8s745_4555aa06-f941-492e-9a1c-89a7c1a5c50f/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Oct 06 10:59:46 crc kubenswrapper[4824]: I1006 10:59:46.990084 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-c7c794649-hcxzr_2919256c-ce36-4092-898d-024f10692bef/neutron-api/0.log" Oct 06 10:59:47 crc kubenswrapper[4824]: I1006 10:59:47.127607 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-c7c794649-hcxzr_2919256c-ce36-4092-898d-024f10692bef/neutron-httpd/0.log" Oct 06 10:59:47 crc kubenswrapper[4824]: I1006 10:59:47.311109 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-kcljj_7e9c21f0-458e-493c-8dee-3ca1b591c60b/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Oct 06 10:59:48 crc kubenswrapper[4824]: I1006 10:59:48.003793 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_20fa260b-d7c3-4f58-82e3-070c45d494f4/nova-api-log/0.log" Oct 06 10:59:48 crc kubenswrapper[4824]: I1006 10:59:48.148098 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_2e76a912-576d-41a3-86b0-f7fc6a6814a2/nova-cell0-conductor-conductor/0.log" Oct 06 10:59:48 crc kubenswrapper[4824]: I1006 10:59:48.185692 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_20fa260b-d7c3-4f58-82e3-070c45d494f4/nova-api-api/0.log" Oct 06 10:59:48 crc kubenswrapper[4824]: I1006 10:59:48.510703 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_e57703dd-1748-4d3d-8f2f-e4edc0901a76/nova-cell1-conductor-conductor/0.log" Oct 06 10:59:48 crc kubenswrapper[4824]: I1006 10:59:48.696819 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_db4434eb-50c9-4004-8ff3-752cffaa01b4/nova-cell1-novncproxy-novncproxy/0.log" Oct 06 10:59:48 crc kubenswrapper[4824]: I1006 10:59:48.868969 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-59jnp_910d1ffe-8746-4412-bba0-74bf7ff0e1ea/nova-edpm-deployment-openstack-edpm-ipam/0.log" Oct 06 10:59:49 crc kubenswrapper[4824]: I1006 10:59:49.142054 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_754abe5c-3e1b-4a4a-b5a7-619326c06846/nova-metadata-log/0.log" Oct 06 10:59:49 crc kubenswrapper[4824]: I1006 10:59:49.649774 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_1166f1ef-43be-42e7-9d5f-84252fccfb69/nova-scheduler-scheduler/0.log" Oct 06 10:59:49 crc kubenswrapper[4824]: I1006 10:59:49.855376 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_1510d553-3505-4bd2-9666-c95cecc43d01/mysql-bootstrap/0.log" Oct 06 10:59:50 crc kubenswrapper[4824]: I1006 10:59:50.092752 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_1510d553-3505-4bd2-9666-c95cecc43d01/mysql-bootstrap/0.log" Oct 06 10:59:50 crc kubenswrapper[4824]: I1006 10:59:50.144087 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_1510d553-3505-4bd2-9666-c95cecc43d01/galera/0.log" Oct 06 10:59:50 crc kubenswrapper[4824]: I1006 10:59:50.401487 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_58f6ad3c-7a20-4be6-8ffb-cadff1c6adef/mysql-bootstrap/0.log" Oct 06 10:59:50 crc kubenswrapper[4824]: I1006 10:59:50.651604 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_58f6ad3c-7a20-4be6-8ffb-cadff1c6adef/mysql-bootstrap/0.log" Oct 06 10:59:50 crc kubenswrapper[4824]: I1006 10:59:50.674719 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_58f6ad3c-7a20-4be6-8ffb-cadff1c6adef/galera/0.log" Oct 06 10:59:50 crc kubenswrapper[4824]: I1006 10:59:50.730915 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_754abe5c-3e1b-4a4a-b5a7-619326c06846/nova-metadata-metadata/0.log" Oct 06 10:59:50 crc kubenswrapper[4824]: I1006 10:59:50.946365 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_6cdeb7d5-11b7-4e30-92fd-5c88b658ea92/openstackclient/0.log" Oct 06 10:59:51 crc kubenswrapper[4824]: I1006 10:59:51.180711 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-48qsh_03d76a8a-df87-4a48-8fb9-e6a502b37ae8/ovn-controller/0.log" Oct 06 10:59:51 crc kubenswrapper[4824]: I1006 10:59:51.240104 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-77kzd_419276d8-4ecf-4133-a98e-6e9373438668/openstack-network-exporter/0.log" Oct 06 10:59:51 crc kubenswrapper[4824]: I1006 10:59:51.507630 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-v22g5_59148e21-e79a-466f-bed5-24671942c24f/ovsdb-server-init/0.log" Oct 06 10:59:51 crc kubenswrapper[4824]: I1006 10:59:51.709075 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-v22g5_59148e21-e79a-466f-bed5-24671942c24f/ovsdb-server-init/0.log" Oct 06 10:59:51 crc kubenswrapper[4824]: I1006 10:59:51.800430 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-v22g5_59148e21-e79a-466f-bed5-24671942c24f/ovs-vswitchd/0.log" Oct 06 10:59:51 crc kubenswrapper[4824]: I1006 10:59:51.842992 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-v22g5_59148e21-e79a-466f-bed5-24671942c24f/ovsdb-server/0.log" Oct 06 10:59:52 crc kubenswrapper[4824]: I1006 10:59:52.107616 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-fn42c_97deb005-9b94-45a5-8d7f-84147ae5807c/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Oct 06 10:59:52 crc kubenswrapper[4824]: I1006 10:59:52.269709 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_43002b2d-db5d-488e-9791-56388700cf9f/openstack-network-exporter/0.log" Oct 06 10:59:52 crc kubenswrapper[4824]: I1006 10:59:52.363025 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_43002b2d-db5d-488e-9791-56388700cf9f/ovn-northd/0.log" Oct 06 10:59:52 crc kubenswrapper[4824]: I1006 10:59:52.561903 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_a39bac9c-ff87-4d4a-ad96-fa5618ca4e40/openstack-network-exporter/0.log" Oct 06 10:59:52 crc kubenswrapper[4824]: I1006 10:59:52.635246 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_a39bac9c-ff87-4d4a-ad96-fa5618ca4e40/ovsdbserver-nb/0.log" Oct 06 10:59:52 crc kubenswrapper[4824]: I1006 10:59:52.801678 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_b6303c51-6db3-4ab6-aeab-edf39e88fdec/openstack-network-exporter/0.log" Oct 06 10:59:52 crc kubenswrapper[4824]: I1006 10:59:52.889403 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_b6303c51-6db3-4ab6-aeab-edf39e88fdec/ovsdbserver-sb/0.log" Oct 06 10:59:53 crc kubenswrapper[4824]: I1006 10:59:53.116190 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-5f446cffd4-v2pxw_1d3b99b4-b383-4c33-ae51-b61c611a5b3d/placement-api/0.log" Oct 06 10:59:53 crc kubenswrapper[4824]: I1006 10:59:53.283900 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-5f446cffd4-v2pxw_1d3b99b4-b383-4c33-ae51-b61c611a5b3d/placement-log/0.log" Oct 06 10:59:53 crc kubenswrapper[4824]: I1006 10:59:53.391414 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_0295abaf-72dd-46d7-9125-543921a6ef61/setup-container/0.log" Oct 06 10:59:53 crc kubenswrapper[4824]: I1006 10:59:53.680747 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_0295abaf-72dd-46d7-9125-543921a6ef61/rabbitmq/0.log" Oct 06 10:59:53 crc kubenswrapper[4824]: I1006 10:59:53.767161 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_0295abaf-72dd-46d7-9125-543921a6ef61/setup-container/0.log" Oct 06 10:59:53 crc kubenswrapper[4824]: I1006 10:59:53.965284 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_9b7a49d4-5482-4da9-aad2-4aabf9584534/setup-container/0.log" Oct 06 10:59:54 crc kubenswrapper[4824]: I1006 10:59:54.110121 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_9b7a49d4-5482-4da9-aad2-4aabf9584534/setup-container/0.log" Oct 06 10:59:54 crc kubenswrapper[4824]: I1006 10:59:54.207345 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_9b7a49d4-5482-4da9-aad2-4aabf9584534/rabbitmq/0.log" Oct 06 10:59:54 crc kubenswrapper[4824]: I1006 10:59:54.347030 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-97h9g_073c5011-b49a-4900-b00b-488164295d4d/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 06 10:59:54 crc kubenswrapper[4824]: I1006 10:59:54.627043 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-pqwkw_a043cdf1-f56a-415d-8914-dc689aca5a9f/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Oct 06 10:59:54 crc kubenswrapper[4824]: I1006 10:59:54.981882 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-9kpm9_4e90bc2c-77ea-40b7-87b1-82cd30dbb082/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Oct 06 10:59:55 crc kubenswrapper[4824]: I1006 10:59:55.033543 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-vwg8r_626b5d8f-93f2-4bbc-add6-a530ddf9a6fa/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 06 10:59:55 crc kubenswrapper[4824]: I1006 10:59:55.323086 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-d24cq_7dcfc0e7-18d6-4a54-9033-acf435eda511/ssh-known-hosts-edpm-deployment/0.log" Oct 06 10:59:55 crc kubenswrapper[4824]: I1006 10:59:55.586996 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-78d766d697-s7wzp_925ed890-0e01-4422-a2f8-4871bf90087d/proxy-server/0.log" Oct 06 10:59:55 crc kubenswrapper[4824]: I1006 10:59:55.741815 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-78d766d697-s7wzp_925ed890-0e01-4422-a2f8-4871bf90087d/proxy-httpd/0.log" Oct 06 10:59:55 crc kubenswrapper[4824]: I1006 10:59:55.825264 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-z6wvs_7f85d924-3fee-4cec-a9df-5a0e41567719/swift-ring-rebalance/0.log" Oct 06 10:59:56 crc kubenswrapper[4824]: I1006 10:59:56.016392 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6cdeba93-f1c9-487d-b776-5c2f1bb353d9/account-auditor/0.log" Oct 06 10:59:56 crc kubenswrapper[4824]: I1006 10:59:56.173789 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6cdeba93-f1c9-487d-b776-5c2f1bb353d9/account-reaper/0.log" Oct 06 10:59:56 crc kubenswrapper[4824]: I1006 10:59:56.249432 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6cdeba93-f1c9-487d-b776-5c2f1bb353d9/account-server/0.log" Oct 06 10:59:56 crc kubenswrapper[4824]: I1006 10:59:56.268953 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6cdeba93-f1c9-487d-b776-5c2f1bb353d9/account-replicator/0.log" Oct 06 10:59:56 crc kubenswrapper[4824]: I1006 10:59:56.468323 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6cdeba93-f1c9-487d-b776-5c2f1bb353d9/container-auditor/0.log" Oct 06 10:59:56 crc kubenswrapper[4824]: I1006 10:59:56.557102 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6cdeba93-f1c9-487d-b776-5c2f1bb353d9/container-server/0.log" Oct 06 10:59:56 crc kubenswrapper[4824]: I1006 10:59:56.582549 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6cdeba93-f1c9-487d-b776-5c2f1bb353d9/container-replicator/0.log" Oct 06 10:59:56 crc kubenswrapper[4824]: I1006 10:59:56.731576 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6cdeba93-f1c9-487d-b776-5c2f1bb353d9/container-updater/0.log" Oct 06 10:59:56 crc kubenswrapper[4824]: I1006 10:59:56.783083 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6cdeba93-f1c9-487d-b776-5c2f1bb353d9/object-auditor/0.log" Oct 06 10:59:56 crc kubenswrapper[4824]: I1006 10:59:56.792502 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6cdeba93-f1c9-487d-b776-5c2f1bb353d9/object-expirer/0.log" Oct 06 10:59:57 crc kubenswrapper[4824]: I1006 10:59:57.021000 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6cdeba93-f1c9-487d-b776-5c2f1bb353d9/object-server/0.log" Oct 06 10:59:57 crc kubenswrapper[4824]: I1006 10:59:57.034733 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6cdeba93-f1c9-487d-b776-5c2f1bb353d9/object-replicator/0.log" Oct 06 10:59:57 crc kubenswrapper[4824]: I1006 10:59:57.053413 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6cdeba93-f1c9-487d-b776-5c2f1bb353d9/object-updater/0.log" Oct 06 10:59:57 crc kubenswrapper[4824]: I1006 10:59:57.241785 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6cdeba93-f1c9-487d-b776-5c2f1bb353d9/rsync/0.log" Oct 06 10:59:57 crc kubenswrapper[4824]: I1006 10:59:57.299283 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6cdeba93-f1c9-487d-b776-5c2f1bb353d9/swift-recon-cron/0.log" Oct 06 10:59:57 crc kubenswrapper[4824]: I1006 10:59:57.520453 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-4br4q_3a410146-5f2b-48a0-824f-ee424f685694/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Oct 06 10:59:57 crc kubenswrapper[4824]: I1006 10:59:57.616365 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_7fd677dd-afa0-45a9-8198-ea78e02f4fb7/tempest-tests-tempest-tests-runner/0.log" Oct 06 10:59:57 crc kubenswrapper[4824]: I1006 10:59:57.856415 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_fbc3bf37-42b1-4a02-9922-2f8bad6f7ad5/test-operator-logs-container/0.log" Oct 06 10:59:58 crc kubenswrapper[4824]: I1006 10:59:58.072671 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-jm56f_9684787a-cab3-4930-9ada-f29df39d21a6/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Oct 06 11:00:00 crc kubenswrapper[4824]: I1006 11:00:00.198577 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29329140-fpf6h"] Oct 06 11:00:00 crc kubenswrapper[4824]: I1006 11:00:00.201374 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29329140-fpf6h" Oct 06 11:00:00 crc kubenswrapper[4824]: I1006 11:00:00.205167 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 06 11:00:00 crc kubenswrapper[4824]: I1006 11:00:00.205635 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 06 11:00:00 crc kubenswrapper[4824]: I1006 11:00:00.216947 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29329140-fpf6h"] Oct 06 11:00:00 crc kubenswrapper[4824]: I1006 11:00:00.341416 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4fk8p\" (UniqueName: \"kubernetes.io/projected/128dbee3-b000-44f2-ace5-b3f799a28bef-kube-api-access-4fk8p\") pod \"collect-profiles-29329140-fpf6h\" (UID: \"128dbee3-b000-44f2-ace5-b3f799a28bef\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329140-fpf6h" Oct 06 11:00:00 crc kubenswrapper[4824]: I1006 11:00:00.341519 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/128dbee3-b000-44f2-ace5-b3f799a28bef-config-volume\") pod \"collect-profiles-29329140-fpf6h\" (UID: \"128dbee3-b000-44f2-ace5-b3f799a28bef\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329140-fpf6h" Oct 06 11:00:00 crc kubenswrapper[4824]: I1006 11:00:00.341579 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/128dbee3-b000-44f2-ace5-b3f799a28bef-secret-volume\") pod \"collect-profiles-29329140-fpf6h\" (UID: \"128dbee3-b000-44f2-ace5-b3f799a28bef\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329140-fpf6h" Oct 06 11:00:00 crc kubenswrapper[4824]: I1006 11:00:00.443771 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/128dbee3-b000-44f2-ace5-b3f799a28bef-config-volume\") pod \"collect-profiles-29329140-fpf6h\" (UID: \"128dbee3-b000-44f2-ace5-b3f799a28bef\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329140-fpf6h" Oct 06 11:00:00 crc kubenswrapper[4824]: I1006 11:00:00.443884 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/128dbee3-b000-44f2-ace5-b3f799a28bef-secret-volume\") pod \"collect-profiles-29329140-fpf6h\" (UID: \"128dbee3-b000-44f2-ace5-b3f799a28bef\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329140-fpf6h" Oct 06 11:00:00 crc kubenswrapper[4824]: I1006 11:00:00.444064 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4fk8p\" (UniqueName: \"kubernetes.io/projected/128dbee3-b000-44f2-ace5-b3f799a28bef-kube-api-access-4fk8p\") pod \"collect-profiles-29329140-fpf6h\" (UID: \"128dbee3-b000-44f2-ace5-b3f799a28bef\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329140-fpf6h" Oct 06 11:00:00 crc kubenswrapper[4824]: I1006 11:00:00.445590 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/128dbee3-b000-44f2-ace5-b3f799a28bef-config-volume\") pod \"collect-profiles-29329140-fpf6h\" (UID: \"128dbee3-b000-44f2-ace5-b3f799a28bef\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329140-fpf6h" Oct 06 11:00:00 crc kubenswrapper[4824]: I1006 11:00:00.467690 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/128dbee3-b000-44f2-ace5-b3f799a28bef-secret-volume\") pod \"collect-profiles-29329140-fpf6h\" (UID: \"128dbee3-b000-44f2-ace5-b3f799a28bef\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329140-fpf6h" Oct 06 11:00:00 crc kubenswrapper[4824]: I1006 11:00:00.468720 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4fk8p\" (UniqueName: \"kubernetes.io/projected/128dbee3-b000-44f2-ace5-b3f799a28bef-kube-api-access-4fk8p\") pod \"collect-profiles-29329140-fpf6h\" (UID: \"128dbee3-b000-44f2-ace5-b3f799a28bef\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329140-fpf6h" Oct 06 11:00:00 crc kubenswrapper[4824]: I1006 11:00:00.525268 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29329140-fpf6h" Oct 06 11:00:01 crc kubenswrapper[4824]: I1006 11:00:01.070323 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29329140-fpf6h"] Oct 06 11:00:01 crc kubenswrapper[4824]: I1006 11:00:01.483937 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29329140-fpf6h" event={"ID":"128dbee3-b000-44f2-ace5-b3f799a28bef","Type":"ContainerStarted","Data":"ed2f4256267231457e6bbdeae73abaebeead60009647533dc6465a1292322e2e"} Oct 06 11:00:01 crc kubenswrapper[4824]: I1006 11:00:01.484439 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29329140-fpf6h" event={"ID":"128dbee3-b000-44f2-ace5-b3f799a28bef","Type":"ContainerStarted","Data":"07be80dad9a18fc3b5fc0b03ac6a8d384cc6cc6dee19f4470484e30057478838"} Oct 06 11:00:01 crc kubenswrapper[4824]: I1006 11:00:01.503548 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29329140-fpf6h" podStartSLOduration=1.5035204100000001 podStartE2EDuration="1.50352041s" podCreationTimestamp="2025-10-06 11:00:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 11:00:01.500032143 +0000 UTC m=+3710.864455004" watchObservedRunningTime="2025-10-06 11:00:01.50352041 +0000 UTC m=+3710.867943271" Oct 06 11:00:02 crc kubenswrapper[4824]: I1006 11:00:02.528156 4824 generic.go:334] "Generic (PLEG): container finished" podID="128dbee3-b000-44f2-ace5-b3f799a28bef" containerID="ed2f4256267231457e6bbdeae73abaebeead60009647533dc6465a1292322e2e" exitCode=0 Oct 06 11:00:02 crc kubenswrapper[4824]: I1006 11:00:02.528233 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29329140-fpf6h" event={"ID":"128dbee3-b000-44f2-ace5-b3f799a28bef","Type":"ContainerDied","Data":"ed2f4256267231457e6bbdeae73abaebeead60009647533dc6465a1292322e2e"} Oct 06 11:00:04 crc kubenswrapper[4824]: I1006 11:00:04.020520 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29329140-fpf6h" Oct 06 11:00:04 crc kubenswrapper[4824]: I1006 11:00:04.159783 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/128dbee3-b000-44f2-ace5-b3f799a28bef-secret-volume\") pod \"128dbee3-b000-44f2-ace5-b3f799a28bef\" (UID: \"128dbee3-b000-44f2-ace5-b3f799a28bef\") " Oct 06 11:00:04 crc kubenswrapper[4824]: I1006 11:00:04.159866 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/128dbee3-b000-44f2-ace5-b3f799a28bef-config-volume\") pod \"128dbee3-b000-44f2-ace5-b3f799a28bef\" (UID: \"128dbee3-b000-44f2-ace5-b3f799a28bef\") " Oct 06 11:00:04 crc kubenswrapper[4824]: I1006 11:00:04.159939 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4fk8p\" (UniqueName: \"kubernetes.io/projected/128dbee3-b000-44f2-ace5-b3f799a28bef-kube-api-access-4fk8p\") pod \"128dbee3-b000-44f2-ace5-b3f799a28bef\" (UID: \"128dbee3-b000-44f2-ace5-b3f799a28bef\") " Oct 06 11:00:04 crc kubenswrapper[4824]: I1006 11:00:04.161494 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/128dbee3-b000-44f2-ace5-b3f799a28bef-config-volume" (OuterVolumeSpecName: "config-volume") pod "128dbee3-b000-44f2-ace5-b3f799a28bef" (UID: "128dbee3-b000-44f2-ace5-b3f799a28bef"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 11:00:04 crc kubenswrapper[4824]: I1006 11:00:04.167532 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/128dbee3-b000-44f2-ace5-b3f799a28bef-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "128dbee3-b000-44f2-ace5-b3f799a28bef" (UID: "128dbee3-b000-44f2-ace5-b3f799a28bef"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 11:00:04 crc kubenswrapper[4824]: I1006 11:00:04.168279 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/128dbee3-b000-44f2-ace5-b3f799a28bef-kube-api-access-4fk8p" (OuterVolumeSpecName: "kube-api-access-4fk8p") pod "128dbee3-b000-44f2-ace5-b3f799a28bef" (UID: "128dbee3-b000-44f2-ace5-b3f799a28bef"). InnerVolumeSpecName "kube-api-access-4fk8p". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 11:00:04 crc kubenswrapper[4824]: I1006 11:00:04.261869 4824 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/128dbee3-b000-44f2-ace5-b3f799a28bef-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 06 11:00:04 crc kubenswrapper[4824]: I1006 11:00:04.262434 4824 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/128dbee3-b000-44f2-ace5-b3f799a28bef-config-volume\") on node \"crc\" DevicePath \"\"" Oct 06 11:00:04 crc kubenswrapper[4824]: I1006 11:00:04.262448 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4fk8p\" (UniqueName: \"kubernetes.io/projected/128dbee3-b000-44f2-ace5-b3f799a28bef-kube-api-access-4fk8p\") on node \"crc\" DevicePath \"\"" Oct 06 11:00:04 crc kubenswrapper[4824]: I1006 11:00:04.394042 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29329095-vgrt8"] Oct 06 11:00:04 crc kubenswrapper[4824]: I1006 11:00:04.404444 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29329095-vgrt8"] Oct 06 11:00:04 crc kubenswrapper[4824]: I1006 11:00:04.561541 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29329140-fpf6h" event={"ID":"128dbee3-b000-44f2-ace5-b3f799a28bef","Type":"ContainerDied","Data":"07be80dad9a18fc3b5fc0b03ac6a8d384cc6cc6dee19f4470484e30057478838"} Oct 06 11:00:04 crc kubenswrapper[4824]: I1006 11:00:04.561582 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="07be80dad9a18fc3b5fc0b03ac6a8d384cc6cc6dee19f4470484e30057478838" Oct 06 11:00:04 crc kubenswrapper[4824]: I1006 11:00:04.561680 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29329140-fpf6h" Oct 06 11:00:05 crc kubenswrapper[4824]: I1006 11:00:05.296969 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="29e4895c-1440-496e-a3f8-06635f103456" path="/var/lib/kubelet/pods/29e4895c-1440-496e-a3f8-06635f103456/volumes" Oct 06 11:00:06 crc kubenswrapper[4824]: I1006 11:00:06.449731 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_3a12653c-c0d0-46e3-b909-77bd2a7f6c4a/memcached/0.log" Oct 06 11:00:47 crc kubenswrapper[4824]: I1006 11:00:47.240462 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-lh2p7"] Oct 06 11:00:47 crc kubenswrapper[4824]: E1006 11:00:47.241782 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="128dbee3-b000-44f2-ace5-b3f799a28bef" containerName="collect-profiles" Oct 06 11:00:47 crc kubenswrapper[4824]: I1006 11:00:47.241798 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="128dbee3-b000-44f2-ace5-b3f799a28bef" containerName="collect-profiles" Oct 06 11:00:47 crc kubenswrapper[4824]: I1006 11:00:47.242072 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="128dbee3-b000-44f2-ace5-b3f799a28bef" containerName="collect-profiles" Oct 06 11:00:47 crc kubenswrapper[4824]: I1006 11:00:47.243665 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lh2p7" Oct 06 11:00:47 crc kubenswrapper[4824]: I1006 11:00:47.255238 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-lh2p7"] Oct 06 11:00:47 crc kubenswrapper[4824]: I1006 11:00:47.435112 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/20a76519-ef12-4100-b3ca-d9fc0bfa667d-catalog-content\") pod \"certified-operators-lh2p7\" (UID: \"20a76519-ef12-4100-b3ca-d9fc0bfa667d\") " pod="openshift-marketplace/certified-operators-lh2p7" Oct 06 11:00:47 crc kubenswrapper[4824]: I1006 11:00:47.435682 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/20a76519-ef12-4100-b3ca-d9fc0bfa667d-utilities\") pod \"certified-operators-lh2p7\" (UID: \"20a76519-ef12-4100-b3ca-d9fc0bfa667d\") " pod="openshift-marketplace/certified-operators-lh2p7" Oct 06 11:00:47 crc kubenswrapper[4824]: I1006 11:00:47.440418 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hd5pm\" (UniqueName: \"kubernetes.io/projected/20a76519-ef12-4100-b3ca-d9fc0bfa667d-kube-api-access-hd5pm\") pod \"certified-operators-lh2p7\" (UID: \"20a76519-ef12-4100-b3ca-d9fc0bfa667d\") " pod="openshift-marketplace/certified-operators-lh2p7" Oct 06 11:00:47 crc kubenswrapper[4824]: I1006 11:00:47.542846 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/20a76519-ef12-4100-b3ca-d9fc0bfa667d-utilities\") pod \"certified-operators-lh2p7\" (UID: \"20a76519-ef12-4100-b3ca-d9fc0bfa667d\") " pod="openshift-marketplace/certified-operators-lh2p7" Oct 06 11:00:47 crc kubenswrapper[4824]: I1006 11:00:47.542958 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hd5pm\" (UniqueName: \"kubernetes.io/projected/20a76519-ef12-4100-b3ca-d9fc0bfa667d-kube-api-access-hd5pm\") pod \"certified-operators-lh2p7\" (UID: \"20a76519-ef12-4100-b3ca-d9fc0bfa667d\") " pod="openshift-marketplace/certified-operators-lh2p7" Oct 06 11:00:47 crc kubenswrapper[4824]: I1006 11:00:47.543060 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/20a76519-ef12-4100-b3ca-d9fc0bfa667d-catalog-content\") pod \"certified-operators-lh2p7\" (UID: \"20a76519-ef12-4100-b3ca-d9fc0bfa667d\") " pod="openshift-marketplace/certified-operators-lh2p7" Oct 06 11:00:47 crc kubenswrapper[4824]: I1006 11:00:47.544347 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/20a76519-ef12-4100-b3ca-d9fc0bfa667d-utilities\") pod \"certified-operators-lh2p7\" (UID: \"20a76519-ef12-4100-b3ca-d9fc0bfa667d\") " pod="openshift-marketplace/certified-operators-lh2p7" Oct 06 11:00:47 crc kubenswrapper[4824]: I1006 11:00:47.544365 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/20a76519-ef12-4100-b3ca-d9fc0bfa667d-catalog-content\") pod \"certified-operators-lh2p7\" (UID: \"20a76519-ef12-4100-b3ca-d9fc0bfa667d\") " pod="openshift-marketplace/certified-operators-lh2p7" Oct 06 11:00:47 crc kubenswrapper[4824]: I1006 11:00:47.568467 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hd5pm\" (UniqueName: \"kubernetes.io/projected/20a76519-ef12-4100-b3ca-d9fc0bfa667d-kube-api-access-hd5pm\") pod \"certified-operators-lh2p7\" (UID: \"20a76519-ef12-4100-b3ca-d9fc0bfa667d\") " pod="openshift-marketplace/certified-operators-lh2p7" Oct 06 11:00:47 crc kubenswrapper[4824]: I1006 11:00:47.571964 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lh2p7" Oct 06 11:00:48 crc kubenswrapper[4824]: I1006 11:00:48.133457 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-lh2p7"] Oct 06 11:00:48 crc kubenswrapper[4824]: I1006 11:00:48.200497 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lh2p7" event={"ID":"20a76519-ef12-4100-b3ca-d9fc0bfa667d","Type":"ContainerStarted","Data":"fcc78842344bed6915e0ed3c133013ff1d27d208de033e805a454b759cd36b9e"} Oct 06 11:00:49 crc kubenswrapper[4824]: I1006 11:00:49.214147 4824 generic.go:334] "Generic (PLEG): container finished" podID="20a76519-ef12-4100-b3ca-d9fc0bfa667d" containerID="131d230d1f3fd55553a371e632c02a2ebdca9d3a889d217f3063c0edaf9e1c48" exitCode=0 Oct 06 11:00:49 crc kubenswrapper[4824]: I1006 11:00:49.214261 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lh2p7" event={"ID":"20a76519-ef12-4100-b3ca-d9fc0bfa667d","Type":"ContainerDied","Data":"131d230d1f3fd55553a371e632c02a2ebdca9d3a889d217f3063c0edaf9e1c48"} Oct 06 11:00:51 crc kubenswrapper[4824]: I1006 11:00:51.195425 4824 scope.go:117] "RemoveContainer" containerID="40f3924e148377f827f93497c59e0add28dc20bb1fc85ce545164fb23eb032a4" Oct 06 11:00:51 crc kubenswrapper[4824]: I1006 11:00:51.236428 4824 generic.go:334] "Generic (PLEG): container finished" podID="20a76519-ef12-4100-b3ca-d9fc0bfa667d" containerID="f47d0a633a6952e97f2365e41478122c32f7a2e5f97a264d31bbb088c2838eaa" exitCode=0 Oct 06 11:00:51 crc kubenswrapper[4824]: I1006 11:00:51.236478 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lh2p7" event={"ID":"20a76519-ef12-4100-b3ca-d9fc0bfa667d","Type":"ContainerDied","Data":"f47d0a633a6952e97f2365e41478122c32f7a2e5f97a264d31bbb088c2838eaa"} Oct 06 11:00:52 crc kubenswrapper[4824]: I1006 11:00:52.254142 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lh2p7" event={"ID":"20a76519-ef12-4100-b3ca-d9fc0bfa667d","Type":"ContainerStarted","Data":"0c46307182016d650173546873a7907a3d72bf56bf8c20ebadde8153139f4f65"} Oct 06 11:00:52 crc kubenswrapper[4824]: I1006 11:00:52.287332 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-lh2p7" podStartSLOduration=2.840465011 podStartE2EDuration="5.287300575s" podCreationTimestamp="2025-10-06 11:00:47 +0000 UTC" firstStartedPulling="2025-10-06 11:00:49.21665114 +0000 UTC m=+3758.581074041" lastFinishedPulling="2025-10-06 11:00:51.663486734 +0000 UTC m=+3761.027909605" observedRunningTime="2025-10-06 11:00:52.283818689 +0000 UTC m=+3761.648241550" watchObservedRunningTime="2025-10-06 11:00:52.287300575 +0000 UTC m=+3761.651723436" Oct 06 11:00:56 crc kubenswrapper[4824]: I1006 11:00:56.308812 4824 generic.go:334] "Generic (PLEG): container finished" podID="fa6ffdef-9f65-4e2c-ae14-24ffdb8ceff9" containerID="22fdcfb9df44916a40b6ce06b4258b25582a99185978d15eb47c665ed73431a1" exitCode=0 Oct 06 11:00:56 crc kubenswrapper[4824]: I1006 11:00:56.308961 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-v7b5k/crc-debug-vbchb" event={"ID":"fa6ffdef-9f65-4e2c-ae14-24ffdb8ceff9","Type":"ContainerDied","Data":"22fdcfb9df44916a40b6ce06b4258b25582a99185978d15eb47c665ed73431a1"} Oct 06 11:00:57 crc kubenswrapper[4824]: I1006 11:00:57.440654 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-v7b5k/crc-debug-vbchb" Oct 06 11:00:57 crc kubenswrapper[4824]: I1006 11:00:57.474724 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-v7b5k/crc-debug-vbchb"] Oct 06 11:00:57 crc kubenswrapper[4824]: I1006 11:00:57.481101 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-v7b5k/crc-debug-vbchb"] Oct 06 11:00:57 crc kubenswrapper[4824]: I1006 11:00:57.492687 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/fa6ffdef-9f65-4e2c-ae14-24ffdb8ceff9-host\") pod \"fa6ffdef-9f65-4e2c-ae14-24ffdb8ceff9\" (UID: \"fa6ffdef-9f65-4e2c-ae14-24ffdb8ceff9\") " Oct 06 11:00:57 crc kubenswrapper[4824]: I1006 11:00:57.492817 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fa6ffdef-9f65-4e2c-ae14-24ffdb8ceff9-host" (OuterVolumeSpecName: "host") pod "fa6ffdef-9f65-4e2c-ae14-24ffdb8ceff9" (UID: "fa6ffdef-9f65-4e2c-ae14-24ffdb8ceff9"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 06 11:00:57 crc kubenswrapper[4824]: I1006 11:00:57.493481 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rtfvs\" (UniqueName: \"kubernetes.io/projected/fa6ffdef-9f65-4e2c-ae14-24ffdb8ceff9-kube-api-access-rtfvs\") pod \"fa6ffdef-9f65-4e2c-ae14-24ffdb8ceff9\" (UID: \"fa6ffdef-9f65-4e2c-ae14-24ffdb8ceff9\") " Oct 06 11:00:57 crc kubenswrapper[4824]: I1006 11:00:57.494311 4824 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/fa6ffdef-9f65-4e2c-ae14-24ffdb8ceff9-host\") on node \"crc\" DevicePath \"\"" Oct 06 11:00:57 crc kubenswrapper[4824]: I1006 11:00:57.502322 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fa6ffdef-9f65-4e2c-ae14-24ffdb8ceff9-kube-api-access-rtfvs" (OuterVolumeSpecName: "kube-api-access-rtfvs") pod "fa6ffdef-9f65-4e2c-ae14-24ffdb8ceff9" (UID: "fa6ffdef-9f65-4e2c-ae14-24ffdb8ceff9"). InnerVolumeSpecName "kube-api-access-rtfvs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 11:00:57 crc kubenswrapper[4824]: I1006 11:00:57.573032 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-lh2p7" Oct 06 11:00:57 crc kubenswrapper[4824]: I1006 11:00:57.573112 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-lh2p7" Oct 06 11:00:57 crc kubenswrapper[4824]: I1006 11:00:57.597461 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rtfvs\" (UniqueName: \"kubernetes.io/projected/fa6ffdef-9f65-4e2c-ae14-24ffdb8ceff9-kube-api-access-rtfvs\") on node \"crc\" DevicePath \"\"" Oct 06 11:00:57 crc kubenswrapper[4824]: I1006 11:00:57.626429 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-lh2p7" Oct 06 11:00:58 crc kubenswrapper[4824]: I1006 11:00:58.338695 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8e88bcf55188658933bd0c65c684146db3e8c654d242763c6d30b5ef8d6abff1" Oct 06 11:00:58 crc kubenswrapper[4824]: I1006 11:00:58.338732 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-v7b5k/crc-debug-vbchb" Oct 06 11:00:58 crc kubenswrapper[4824]: I1006 11:00:58.438198 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-lh2p7" Oct 06 11:00:58 crc kubenswrapper[4824]: I1006 11:00:58.542836 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-lh2p7"] Oct 06 11:00:59 crc kubenswrapper[4824]: I1006 11:00:59.052617 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-v7b5k/crc-debug-spt6k"] Oct 06 11:00:59 crc kubenswrapper[4824]: E1006 11:00:59.053627 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa6ffdef-9f65-4e2c-ae14-24ffdb8ceff9" containerName="container-00" Oct 06 11:00:59 crc kubenswrapper[4824]: I1006 11:00:59.053650 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa6ffdef-9f65-4e2c-ae14-24ffdb8ceff9" containerName="container-00" Oct 06 11:00:59 crc kubenswrapper[4824]: I1006 11:00:59.053860 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="fa6ffdef-9f65-4e2c-ae14-24ffdb8ceff9" containerName="container-00" Oct 06 11:00:59 crc kubenswrapper[4824]: I1006 11:00:59.054817 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-v7b5k/crc-debug-spt6k" Oct 06 11:00:59 crc kubenswrapper[4824]: I1006 11:00:59.235090 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pw7p7\" (UniqueName: \"kubernetes.io/projected/e91477ed-d23b-4a31-9dab-d55657afbd3b-kube-api-access-pw7p7\") pod \"crc-debug-spt6k\" (UID: \"e91477ed-d23b-4a31-9dab-d55657afbd3b\") " pod="openshift-must-gather-v7b5k/crc-debug-spt6k" Oct 06 11:00:59 crc kubenswrapper[4824]: I1006 11:00:59.235352 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e91477ed-d23b-4a31-9dab-d55657afbd3b-host\") pod \"crc-debug-spt6k\" (UID: \"e91477ed-d23b-4a31-9dab-d55657afbd3b\") " pod="openshift-must-gather-v7b5k/crc-debug-spt6k" Oct 06 11:00:59 crc kubenswrapper[4824]: I1006 11:00:59.287655 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fa6ffdef-9f65-4e2c-ae14-24ffdb8ceff9" path="/var/lib/kubelet/pods/fa6ffdef-9f65-4e2c-ae14-24ffdb8ceff9/volumes" Oct 06 11:00:59 crc kubenswrapper[4824]: I1006 11:00:59.337345 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e91477ed-d23b-4a31-9dab-d55657afbd3b-host\") pod \"crc-debug-spt6k\" (UID: \"e91477ed-d23b-4a31-9dab-d55657afbd3b\") " pod="openshift-must-gather-v7b5k/crc-debug-spt6k" Oct 06 11:00:59 crc kubenswrapper[4824]: I1006 11:00:59.337411 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pw7p7\" (UniqueName: \"kubernetes.io/projected/e91477ed-d23b-4a31-9dab-d55657afbd3b-kube-api-access-pw7p7\") pod \"crc-debug-spt6k\" (UID: \"e91477ed-d23b-4a31-9dab-d55657afbd3b\") " pod="openshift-must-gather-v7b5k/crc-debug-spt6k" Oct 06 11:00:59 crc kubenswrapper[4824]: I1006 11:00:59.337506 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e91477ed-d23b-4a31-9dab-d55657afbd3b-host\") pod \"crc-debug-spt6k\" (UID: \"e91477ed-d23b-4a31-9dab-d55657afbd3b\") " pod="openshift-must-gather-v7b5k/crc-debug-spt6k" Oct 06 11:00:59 crc kubenswrapper[4824]: I1006 11:00:59.364672 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pw7p7\" (UniqueName: \"kubernetes.io/projected/e91477ed-d23b-4a31-9dab-d55657afbd3b-kube-api-access-pw7p7\") pod \"crc-debug-spt6k\" (UID: \"e91477ed-d23b-4a31-9dab-d55657afbd3b\") " pod="openshift-must-gather-v7b5k/crc-debug-spt6k" Oct 06 11:00:59 crc kubenswrapper[4824]: I1006 11:00:59.375523 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-v7b5k/crc-debug-spt6k" Oct 06 11:01:00 crc kubenswrapper[4824]: I1006 11:01:00.157821 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29329141-z4m2j"] Oct 06 11:01:00 crc kubenswrapper[4824]: I1006 11:01:00.159800 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29329141-z4m2j" Oct 06 11:01:00 crc kubenswrapper[4824]: I1006 11:01:00.185051 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29329141-z4m2j"] Oct 06 11:01:00 crc kubenswrapper[4824]: I1006 11:01:00.260240 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pfzhr\" (UniqueName: \"kubernetes.io/projected/46ee6fe4-f9a8-4f4a-a25c-0807817131d4-kube-api-access-pfzhr\") pod \"keystone-cron-29329141-z4m2j\" (UID: \"46ee6fe4-f9a8-4f4a-a25c-0807817131d4\") " pod="openstack/keystone-cron-29329141-z4m2j" Oct 06 11:01:00 crc kubenswrapper[4824]: I1006 11:01:00.260291 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46ee6fe4-f9a8-4f4a-a25c-0807817131d4-config-data\") pod \"keystone-cron-29329141-z4m2j\" (UID: \"46ee6fe4-f9a8-4f4a-a25c-0807817131d4\") " pod="openstack/keystone-cron-29329141-z4m2j" Oct 06 11:01:00 crc kubenswrapper[4824]: I1006 11:01:00.260408 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46ee6fe4-f9a8-4f4a-a25c-0807817131d4-combined-ca-bundle\") pod \"keystone-cron-29329141-z4m2j\" (UID: \"46ee6fe4-f9a8-4f4a-a25c-0807817131d4\") " pod="openstack/keystone-cron-29329141-z4m2j" Oct 06 11:01:00 crc kubenswrapper[4824]: I1006 11:01:00.260450 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/46ee6fe4-f9a8-4f4a-a25c-0807817131d4-fernet-keys\") pod \"keystone-cron-29329141-z4m2j\" (UID: \"46ee6fe4-f9a8-4f4a-a25c-0807817131d4\") " pod="openstack/keystone-cron-29329141-z4m2j" Oct 06 11:01:00 crc kubenswrapper[4824]: I1006 11:01:00.358187 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-v7b5k/crc-debug-spt6k" event={"ID":"e91477ed-d23b-4a31-9dab-d55657afbd3b","Type":"ContainerStarted","Data":"60d15bc379e2f400b7346e85651a8f0e968a4e13015c6e329a969e869bc7971c"} Oct 06 11:01:00 crc kubenswrapper[4824]: I1006 11:01:00.358235 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-v7b5k/crc-debug-spt6k" event={"ID":"e91477ed-d23b-4a31-9dab-d55657afbd3b","Type":"ContainerStarted","Data":"040218944b4997996394867b680ed633537bdd50cf85511194f4c3c68648dc29"} Oct 06 11:01:00 crc kubenswrapper[4824]: I1006 11:01:00.358320 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-lh2p7" podUID="20a76519-ef12-4100-b3ca-d9fc0bfa667d" containerName="registry-server" containerID="cri-o://0c46307182016d650173546873a7907a3d72bf56bf8c20ebadde8153139f4f65" gracePeriod=2 Oct 06 11:01:00 crc kubenswrapper[4824]: I1006 11:01:00.361811 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pfzhr\" (UniqueName: \"kubernetes.io/projected/46ee6fe4-f9a8-4f4a-a25c-0807817131d4-kube-api-access-pfzhr\") pod \"keystone-cron-29329141-z4m2j\" (UID: \"46ee6fe4-f9a8-4f4a-a25c-0807817131d4\") " pod="openstack/keystone-cron-29329141-z4m2j" Oct 06 11:01:00 crc kubenswrapper[4824]: I1006 11:01:00.361858 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46ee6fe4-f9a8-4f4a-a25c-0807817131d4-config-data\") pod \"keystone-cron-29329141-z4m2j\" (UID: \"46ee6fe4-f9a8-4f4a-a25c-0807817131d4\") " pod="openstack/keystone-cron-29329141-z4m2j" Oct 06 11:01:00 crc kubenswrapper[4824]: I1006 11:01:00.361969 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46ee6fe4-f9a8-4f4a-a25c-0807817131d4-combined-ca-bundle\") pod \"keystone-cron-29329141-z4m2j\" (UID: \"46ee6fe4-f9a8-4f4a-a25c-0807817131d4\") " pod="openstack/keystone-cron-29329141-z4m2j" Oct 06 11:01:00 crc kubenswrapper[4824]: I1006 11:01:00.362043 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/46ee6fe4-f9a8-4f4a-a25c-0807817131d4-fernet-keys\") pod \"keystone-cron-29329141-z4m2j\" (UID: \"46ee6fe4-f9a8-4f4a-a25c-0807817131d4\") " pod="openstack/keystone-cron-29329141-z4m2j" Oct 06 11:01:00 crc kubenswrapper[4824]: I1006 11:01:00.370757 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/46ee6fe4-f9a8-4f4a-a25c-0807817131d4-fernet-keys\") pod \"keystone-cron-29329141-z4m2j\" (UID: \"46ee6fe4-f9a8-4f4a-a25c-0807817131d4\") " pod="openstack/keystone-cron-29329141-z4m2j" Oct 06 11:01:00 crc kubenswrapper[4824]: I1006 11:01:00.374768 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46ee6fe4-f9a8-4f4a-a25c-0807817131d4-combined-ca-bundle\") pod \"keystone-cron-29329141-z4m2j\" (UID: \"46ee6fe4-f9a8-4f4a-a25c-0807817131d4\") " pod="openstack/keystone-cron-29329141-z4m2j" Oct 06 11:01:00 crc kubenswrapper[4824]: I1006 11:01:00.384089 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46ee6fe4-f9a8-4f4a-a25c-0807817131d4-config-data\") pod \"keystone-cron-29329141-z4m2j\" (UID: \"46ee6fe4-f9a8-4f4a-a25c-0807817131d4\") " pod="openstack/keystone-cron-29329141-z4m2j" Oct 06 11:01:00 crc kubenswrapper[4824]: I1006 11:01:00.396885 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pfzhr\" (UniqueName: \"kubernetes.io/projected/46ee6fe4-f9a8-4f4a-a25c-0807817131d4-kube-api-access-pfzhr\") pod \"keystone-cron-29329141-z4m2j\" (UID: \"46ee6fe4-f9a8-4f4a-a25c-0807817131d4\") " pod="openstack/keystone-cron-29329141-z4m2j" Oct 06 11:01:00 crc kubenswrapper[4824]: I1006 11:01:00.410533 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-v7b5k/crc-debug-spt6k" podStartSLOduration=1.410510545 podStartE2EDuration="1.410510545s" podCreationTimestamp="2025-10-06 11:00:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 11:01:00.405194483 +0000 UTC m=+3769.769617364" watchObservedRunningTime="2025-10-06 11:01:00.410510545 +0000 UTC m=+3769.774933406" Oct 06 11:01:00 crc kubenswrapper[4824]: I1006 11:01:00.500530 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29329141-z4m2j" Oct 06 11:01:01 crc kubenswrapper[4824]: I1006 11:01:01.145038 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lh2p7" Oct 06 11:01:01 crc kubenswrapper[4824]: I1006 11:01:01.228972 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hd5pm\" (UniqueName: \"kubernetes.io/projected/20a76519-ef12-4100-b3ca-d9fc0bfa667d-kube-api-access-hd5pm\") pod \"20a76519-ef12-4100-b3ca-d9fc0bfa667d\" (UID: \"20a76519-ef12-4100-b3ca-d9fc0bfa667d\") " Oct 06 11:01:01 crc kubenswrapper[4824]: I1006 11:01:01.229222 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/20a76519-ef12-4100-b3ca-d9fc0bfa667d-utilities\") pod \"20a76519-ef12-4100-b3ca-d9fc0bfa667d\" (UID: \"20a76519-ef12-4100-b3ca-d9fc0bfa667d\") " Oct 06 11:01:01 crc kubenswrapper[4824]: I1006 11:01:01.229250 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/20a76519-ef12-4100-b3ca-d9fc0bfa667d-catalog-content\") pod \"20a76519-ef12-4100-b3ca-d9fc0bfa667d\" (UID: \"20a76519-ef12-4100-b3ca-d9fc0bfa667d\") " Oct 06 11:01:01 crc kubenswrapper[4824]: I1006 11:01:01.231193 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/20a76519-ef12-4100-b3ca-d9fc0bfa667d-utilities" (OuterVolumeSpecName: "utilities") pod "20a76519-ef12-4100-b3ca-d9fc0bfa667d" (UID: "20a76519-ef12-4100-b3ca-d9fc0bfa667d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 11:01:01 crc kubenswrapper[4824]: I1006 11:01:01.238194 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20a76519-ef12-4100-b3ca-d9fc0bfa667d-kube-api-access-hd5pm" (OuterVolumeSpecName: "kube-api-access-hd5pm") pod "20a76519-ef12-4100-b3ca-d9fc0bfa667d" (UID: "20a76519-ef12-4100-b3ca-d9fc0bfa667d"). InnerVolumeSpecName "kube-api-access-hd5pm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 11:01:01 crc kubenswrapper[4824]: I1006 11:01:01.304635 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29329141-z4m2j"] Oct 06 11:01:01 crc kubenswrapper[4824]: I1006 11:01:01.336769 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/20a76519-ef12-4100-b3ca-d9fc0bfa667d-utilities\") on node \"crc\" DevicePath \"\"" Oct 06 11:01:01 crc kubenswrapper[4824]: I1006 11:01:01.336818 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hd5pm\" (UniqueName: \"kubernetes.io/projected/20a76519-ef12-4100-b3ca-d9fc0bfa667d-kube-api-access-hd5pm\") on node \"crc\" DevicePath \"\"" Oct 06 11:01:01 crc kubenswrapper[4824]: I1006 11:01:01.386641 4824 generic.go:334] "Generic (PLEG): container finished" podID="20a76519-ef12-4100-b3ca-d9fc0bfa667d" containerID="0c46307182016d650173546873a7907a3d72bf56bf8c20ebadde8153139f4f65" exitCode=0 Oct 06 11:01:01 crc kubenswrapper[4824]: I1006 11:01:01.386703 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lh2p7" event={"ID":"20a76519-ef12-4100-b3ca-d9fc0bfa667d","Type":"ContainerDied","Data":"0c46307182016d650173546873a7907a3d72bf56bf8c20ebadde8153139f4f65"} Oct 06 11:01:01 crc kubenswrapper[4824]: I1006 11:01:01.386789 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lh2p7" event={"ID":"20a76519-ef12-4100-b3ca-d9fc0bfa667d","Type":"ContainerDied","Data":"fcc78842344bed6915e0ed3c133013ff1d27d208de033e805a454b759cd36b9e"} Oct 06 11:01:01 crc kubenswrapper[4824]: I1006 11:01:01.386824 4824 scope.go:117] "RemoveContainer" containerID="0c46307182016d650173546873a7907a3d72bf56bf8c20ebadde8153139f4f65" Oct 06 11:01:01 crc kubenswrapper[4824]: I1006 11:01:01.387323 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lh2p7" Oct 06 11:01:01 crc kubenswrapper[4824]: I1006 11:01:01.390063 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29329141-z4m2j" event={"ID":"46ee6fe4-f9a8-4f4a-a25c-0807817131d4","Type":"ContainerStarted","Data":"b8283822c43ecea5142f4b8ae0074e50a9ec69ddd66f29d4dde2a80ae06ae9b3"} Oct 06 11:01:01 crc kubenswrapper[4824]: I1006 11:01:01.423158 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/20a76519-ef12-4100-b3ca-d9fc0bfa667d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "20a76519-ef12-4100-b3ca-d9fc0bfa667d" (UID: "20a76519-ef12-4100-b3ca-d9fc0bfa667d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 11:01:01 crc kubenswrapper[4824]: I1006 11:01:01.438851 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/20a76519-ef12-4100-b3ca-d9fc0bfa667d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 06 11:01:01 crc kubenswrapper[4824]: I1006 11:01:01.453258 4824 scope.go:117] "RemoveContainer" containerID="f47d0a633a6952e97f2365e41478122c32f7a2e5f97a264d31bbb088c2838eaa" Oct 06 11:01:01 crc kubenswrapper[4824]: I1006 11:01:01.524055 4824 scope.go:117] "RemoveContainer" containerID="131d230d1f3fd55553a371e632c02a2ebdca9d3a889d217f3063c0edaf9e1c48" Oct 06 11:01:01 crc kubenswrapper[4824]: I1006 11:01:01.576749 4824 scope.go:117] "RemoveContainer" containerID="0c46307182016d650173546873a7907a3d72bf56bf8c20ebadde8153139f4f65" Oct 06 11:01:01 crc kubenswrapper[4824]: E1006 11:01:01.577274 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0c46307182016d650173546873a7907a3d72bf56bf8c20ebadde8153139f4f65\": container with ID starting with 0c46307182016d650173546873a7907a3d72bf56bf8c20ebadde8153139f4f65 not found: ID does not exist" containerID="0c46307182016d650173546873a7907a3d72bf56bf8c20ebadde8153139f4f65" Oct 06 11:01:01 crc kubenswrapper[4824]: I1006 11:01:01.577334 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0c46307182016d650173546873a7907a3d72bf56bf8c20ebadde8153139f4f65"} err="failed to get container status \"0c46307182016d650173546873a7907a3d72bf56bf8c20ebadde8153139f4f65\": rpc error: code = NotFound desc = could not find container \"0c46307182016d650173546873a7907a3d72bf56bf8c20ebadde8153139f4f65\": container with ID starting with 0c46307182016d650173546873a7907a3d72bf56bf8c20ebadde8153139f4f65 not found: ID does not exist" Oct 06 11:01:01 crc kubenswrapper[4824]: I1006 11:01:01.577760 4824 scope.go:117] "RemoveContainer" containerID="f47d0a633a6952e97f2365e41478122c32f7a2e5f97a264d31bbb088c2838eaa" Oct 06 11:01:01 crc kubenswrapper[4824]: E1006 11:01:01.578349 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f47d0a633a6952e97f2365e41478122c32f7a2e5f97a264d31bbb088c2838eaa\": container with ID starting with f47d0a633a6952e97f2365e41478122c32f7a2e5f97a264d31bbb088c2838eaa not found: ID does not exist" containerID="f47d0a633a6952e97f2365e41478122c32f7a2e5f97a264d31bbb088c2838eaa" Oct 06 11:01:01 crc kubenswrapper[4824]: I1006 11:01:01.578645 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f47d0a633a6952e97f2365e41478122c32f7a2e5f97a264d31bbb088c2838eaa"} err="failed to get container status \"f47d0a633a6952e97f2365e41478122c32f7a2e5f97a264d31bbb088c2838eaa\": rpc error: code = NotFound desc = could not find container \"f47d0a633a6952e97f2365e41478122c32f7a2e5f97a264d31bbb088c2838eaa\": container with ID starting with f47d0a633a6952e97f2365e41478122c32f7a2e5f97a264d31bbb088c2838eaa not found: ID does not exist" Oct 06 11:01:01 crc kubenswrapper[4824]: I1006 11:01:01.578671 4824 scope.go:117] "RemoveContainer" containerID="131d230d1f3fd55553a371e632c02a2ebdca9d3a889d217f3063c0edaf9e1c48" Oct 06 11:01:01 crc kubenswrapper[4824]: E1006 11:01:01.580665 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"131d230d1f3fd55553a371e632c02a2ebdca9d3a889d217f3063c0edaf9e1c48\": container with ID starting with 131d230d1f3fd55553a371e632c02a2ebdca9d3a889d217f3063c0edaf9e1c48 not found: ID does not exist" containerID="131d230d1f3fd55553a371e632c02a2ebdca9d3a889d217f3063c0edaf9e1c48" Oct 06 11:01:01 crc kubenswrapper[4824]: I1006 11:01:01.580738 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"131d230d1f3fd55553a371e632c02a2ebdca9d3a889d217f3063c0edaf9e1c48"} err="failed to get container status \"131d230d1f3fd55553a371e632c02a2ebdca9d3a889d217f3063c0edaf9e1c48\": rpc error: code = NotFound desc = could not find container \"131d230d1f3fd55553a371e632c02a2ebdca9d3a889d217f3063c0edaf9e1c48\": container with ID starting with 131d230d1f3fd55553a371e632c02a2ebdca9d3a889d217f3063c0edaf9e1c48 not found: ID does not exist" Oct 06 11:01:01 crc kubenswrapper[4824]: I1006 11:01:01.724603 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-lh2p7"] Oct 06 11:01:01 crc kubenswrapper[4824]: I1006 11:01:01.734682 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-lh2p7"] Oct 06 11:01:02 crc kubenswrapper[4824]: I1006 11:01:02.404892 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29329141-z4m2j" event={"ID":"46ee6fe4-f9a8-4f4a-a25c-0807817131d4","Type":"ContainerStarted","Data":"228dc20c4ab0850d504daa6fe8382a851d0fa370dbc6d3e58d607559b016910c"} Oct 06 11:01:02 crc kubenswrapper[4824]: I1006 11:01:02.410142 4824 generic.go:334] "Generic (PLEG): container finished" podID="e91477ed-d23b-4a31-9dab-d55657afbd3b" containerID="60d15bc379e2f400b7346e85651a8f0e968a4e13015c6e329a969e869bc7971c" exitCode=0 Oct 06 11:01:02 crc kubenswrapper[4824]: I1006 11:01:02.410226 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-v7b5k/crc-debug-spt6k" event={"ID":"e91477ed-d23b-4a31-9dab-d55657afbd3b","Type":"ContainerDied","Data":"60d15bc379e2f400b7346e85651a8f0e968a4e13015c6e329a969e869bc7971c"} Oct 06 11:01:02 crc kubenswrapper[4824]: I1006 11:01:02.434195 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29329141-z4m2j" podStartSLOduration=2.434148872 podStartE2EDuration="2.434148872s" podCreationTimestamp="2025-10-06 11:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 11:01:02.423318692 +0000 UTC m=+3771.787741553" watchObservedRunningTime="2025-10-06 11:01:02.434148872 +0000 UTC m=+3771.798571993" Oct 06 11:01:03 crc kubenswrapper[4824]: I1006 11:01:03.285842 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20a76519-ef12-4100-b3ca-d9fc0bfa667d" path="/var/lib/kubelet/pods/20a76519-ef12-4100-b3ca-d9fc0bfa667d/volumes" Oct 06 11:01:03 crc kubenswrapper[4824]: I1006 11:01:03.546404 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-v7b5k/crc-debug-spt6k" Oct 06 11:01:03 crc kubenswrapper[4824]: I1006 11:01:03.682710 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pw7p7\" (UniqueName: \"kubernetes.io/projected/e91477ed-d23b-4a31-9dab-d55657afbd3b-kube-api-access-pw7p7\") pod \"e91477ed-d23b-4a31-9dab-d55657afbd3b\" (UID: \"e91477ed-d23b-4a31-9dab-d55657afbd3b\") " Oct 06 11:01:03 crc kubenswrapper[4824]: I1006 11:01:03.682968 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e91477ed-d23b-4a31-9dab-d55657afbd3b-host\") pod \"e91477ed-d23b-4a31-9dab-d55657afbd3b\" (UID: \"e91477ed-d23b-4a31-9dab-d55657afbd3b\") " Oct 06 11:01:03 crc kubenswrapper[4824]: I1006 11:01:03.683537 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e91477ed-d23b-4a31-9dab-d55657afbd3b-host" (OuterVolumeSpecName: "host") pod "e91477ed-d23b-4a31-9dab-d55657afbd3b" (UID: "e91477ed-d23b-4a31-9dab-d55657afbd3b"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 06 11:01:03 crc kubenswrapper[4824]: I1006 11:01:03.711273 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e91477ed-d23b-4a31-9dab-d55657afbd3b-kube-api-access-pw7p7" (OuterVolumeSpecName: "kube-api-access-pw7p7") pod "e91477ed-d23b-4a31-9dab-d55657afbd3b" (UID: "e91477ed-d23b-4a31-9dab-d55657afbd3b"). InnerVolumeSpecName "kube-api-access-pw7p7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 11:01:03 crc kubenswrapper[4824]: I1006 11:01:03.786713 4824 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e91477ed-d23b-4a31-9dab-d55657afbd3b-host\") on node \"crc\" DevicePath \"\"" Oct 06 11:01:03 crc kubenswrapper[4824]: I1006 11:01:03.786747 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pw7p7\" (UniqueName: \"kubernetes.io/projected/e91477ed-d23b-4a31-9dab-d55657afbd3b-kube-api-access-pw7p7\") on node \"crc\" DevicePath \"\"" Oct 06 11:01:04 crc kubenswrapper[4824]: I1006 11:01:04.439783 4824 generic.go:334] "Generic (PLEG): container finished" podID="46ee6fe4-f9a8-4f4a-a25c-0807817131d4" containerID="228dc20c4ab0850d504daa6fe8382a851d0fa370dbc6d3e58d607559b016910c" exitCode=0 Oct 06 11:01:04 crc kubenswrapper[4824]: I1006 11:01:04.440171 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29329141-z4m2j" event={"ID":"46ee6fe4-f9a8-4f4a-a25c-0807817131d4","Type":"ContainerDied","Data":"228dc20c4ab0850d504daa6fe8382a851d0fa370dbc6d3e58d607559b016910c"} Oct 06 11:01:04 crc kubenswrapper[4824]: I1006 11:01:04.443537 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-v7b5k/crc-debug-spt6k" event={"ID":"e91477ed-d23b-4a31-9dab-d55657afbd3b","Type":"ContainerDied","Data":"040218944b4997996394867b680ed633537bdd50cf85511194f4c3c68648dc29"} Oct 06 11:01:04 crc kubenswrapper[4824]: I1006 11:01:04.443591 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="040218944b4997996394867b680ed633537bdd50cf85511194f4c3c68648dc29" Oct 06 11:01:04 crc kubenswrapper[4824]: I1006 11:01:04.443662 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-v7b5k/crc-debug-spt6k" Oct 06 11:01:05 crc kubenswrapper[4824]: I1006 11:01:05.873179 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29329141-z4m2j" Oct 06 11:01:06 crc kubenswrapper[4824]: I1006 11:01:06.027214 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46ee6fe4-f9a8-4f4a-a25c-0807817131d4-config-data\") pod \"46ee6fe4-f9a8-4f4a-a25c-0807817131d4\" (UID: \"46ee6fe4-f9a8-4f4a-a25c-0807817131d4\") " Oct 06 11:01:06 crc kubenswrapper[4824]: I1006 11:01:06.027292 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46ee6fe4-f9a8-4f4a-a25c-0807817131d4-combined-ca-bundle\") pod \"46ee6fe4-f9a8-4f4a-a25c-0807817131d4\" (UID: \"46ee6fe4-f9a8-4f4a-a25c-0807817131d4\") " Oct 06 11:01:06 crc kubenswrapper[4824]: I1006 11:01:06.027324 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pfzhr\" (UniqueName: \"kubernetes.io/projected/46ee6fe4-f9a8-4f4a-a25c-0807817131d4-kube-api-access-pfzhr\") pod \"46ee6fe4-f9a8-4f4a-a25c-0807817131d4\" (UID: \"46ee6fe4-f9a8-4f4a-a25c-0807817131d4\") " Oct 06 11:01:06 crc kubenswrapper[4824]: I1006 11:01:06.027474 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/46ee6fe4-f9a8-4f4a-a25c-0807817131d4-fernet-keys\") pod \"46ee6fe4-f9a8-4f4a-a25c-0807817131d4\" (UID: \"46ee6fe4-f9a8-4f4a-a25c-0807817131d4\") " Oct 06 11:01:06 crc kubenswrapper[4824]: I1006 11:01:06.039455 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/46ee6fe4-f9a8-4f4a-a25c-0807817131d4-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "46ee6fe4-f9a8-4f4a-a25c-0807817131d4" (UID: "46ee6fe4-f9a8-4f4a-a25c-0807817131d4"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 11:01:06 crc kubenswrapper[4824]: I1006 11:01:06.045171 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/46ee6fe4-f9a8-4f4a-a25c-0807817131d4-kube-api-access-pfzhr" (OuterVolumeSpecName: "kube-api-access-pfzhr") pod "46ee6fe4-f9a8-4f4a-a25c-0807817131d4" (UID: "46ee6fe4-f9a8-4f4a-a25c-0807817131d4"). InnerVolumeSpecName "kube-api-access-pfzhr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 11:01:06 crc kubenswrapper[4824]: I1006 11:01:06.063208 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/46ee6fe4-f9a8-4f4a-a25c-0807817131d4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "46ee6fe4-f9a8-4f4a-a25c-0807817131d4" (UID: "46ee6fe4-f9a8-4f4a-a25c-0807817131d4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 11:01:06 crc kubenswrapper[4824]: I1006 11:01:06.091037 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/46ee6fe4-f9a8-4f4a-a25c-0807817131d4-config-data" (OuterVolumeSpecName: "config-data") pod "46ee6fe4-f9a8-4f4a-a25c-0807817131d4" (UID: "46ee6fe4-f9a8-4f4a-a25c-0807817131d4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 11:01:06 crc kubenswrapper[4824]: I1006 11:01:06.129387 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46ee6fe4-f9a8-4f4a-a25c-0807817131d4-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 11:01:06 crc kubenswrapper[4824]: I1006 11:01:06.129425 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46ee6fe4-f9a8-4f4a-a25c-0807817131d4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 11:01:06 crc kubenswrapper[4824]: I1006 11:01:06.129442 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pfzhr\" (UniqueName: \"kubernetes.io/projected/46ee6fe4-f9a8-4f4a-a25c-0807817131d4-kube-api-access-pfzhr\") on node \"crc\" DevicePath \"\"" Oct 06 11:01:06 crc kubenswrapper[4824]: I1006 11:01:06.129451 4824 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/46ee6fe4-f9a8-4f4a-a25c-0807817131d4-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 06 11:01:06 crc kubenswrapper[4824]: I1006 11:01:06.468476 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29329141-z4m2j" event={"ID":"46ee6fe4-f9a8-4f4a-a25c-0807817131d4","Type":"ContainerDied","Data":"b8283822c43ecea5142f4b8ae0074e50a9ec69ddd66f29d4dde2a80ae06ae9b3"} Oct 06 11:01:06 crc kubenswrapper[4824]: I1006 11:01:06.468541 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b8283822c43ecea5142f4b8ae0074e50a9ec69ddd66f29d4dde2a80ae06ae9b3" Oct 06 11:01:06 crc kubenswrapper[4824]: I1006 11:01:06.468564 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29329141-z4m2j" Oct 06 11:01:07 crc kubenswrapper[4824]: I1006 11:01:07.914427 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-v7b5k/crc-debug-spt6k"] Oct 06 11:01:07 crc kubenswrapper[4824]: I1006 11:01:07.923427 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-v7b5k/crc-debug-spt6k"] Oct 06 11:01:09 crc kubenswrapper[4824]: I1006 11:01:09.146355 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-v7b5k/crc-debug-p6rbb"] Oct 06 11:01:09 crc kubenswrapper[4824]: E1006 11:01:09.147514 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20a76519-ef12-4100-b3ca-d9fc0bfa667d" containerName="extract-utilities" Oct 06 11:01:09 crc kubenswrapper[4824]: I1006 11:01:09.147539 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="20a76519-ef12-4100-b3ca-d9fc0bfa667d" containerName="extract-utilities" Oct 06 11:01:09 crc kubenswrapper[4824]: E1006 11:01:09.147587 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20a76519-ef12-4100-b3ca-d9fc0bfa667d" containerName="registry-server" Oct 06 11:01:09 crc kubenswrapper[4824]: I1006 11:01:09.147599 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="20a76519-ef12-4100-b3ca-d9fc0bfa667d" containerName="registry-server" Oct 06 11:01:09 crc kubenswrapper[4824]: E1006 11:01:09.147628 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46ee6fe4-f9a8-4f4a-a25c-0807817131d4" containerName="keystone-cron" Oct 06 11:01:09 crc kubenswrapper[4824]: I1006 11:01:09.147639 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="46ee6fe4-f9a8-4f4a-a25c-0807817131d4" containerName="keystone-cron" Oct 06 11:01:09 crc kubenswrapper[4824]: E1006 11:01:09.147665 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e91477ed-d23b-4a31-9dab-d55657afbd3b" containerName="container-00" Oct 06 11:01:09 crc kubenswrapper[4824]: I1006 11:01:09.147675 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="e91477ed-d23b-4a31-9dab-d55657afbd3b" containerName="container-00" Oct 06 11:01:09 crc kubenswrapper[4824]: E1006 11:01:09.147728 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20a76519-ef12-4100-b3ca-d9fc0bfa667d" containerName="extract-content" Oct 06 11:01:09 crc kubenswrapper[4824]: I1006 11:01:09.147738 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="20a76519-ef12-4100-b3ca-d9fc0bfa667d" containerName="extract-content" Oct 06 11:01:09 crc kubenswrapper[4824]: I1006 11:01:09.148061 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="20a76519-ef12-4100-b3ca-d9fc0bfa667d" containerName="registry-server" Oct 06 11:01:09 crc kubenswrapper[4824]: I1006 11:01:09.148090 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="46ee6fe4-f9a8-4f4a-a25c-0807817131d4" containerName="keystone-cron" Oct 06 11:01:09 crc kubenswrapper[4824]: I1006 11:01:09.148109 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="e91477ed-d23b-4a31-9dab-d55657afbd3b" containerName="container-00" Oct 06 11:01:09 crc kubenswrapper[4824]: I1006 11:01:09.149164 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-v7b5k/crc-debug-p6rbb" Oct 06 11:01:09 crc kubenswrapper[4824]: I1006 11:01:09.289082 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e91477ed-d23b-4a31-9dab-d55657afbd3b" path="/var/lib/kubelet/pods/e91477ed-d23b-4a31-9dab-d55657afbd3b/volumes" Oct 06 11:01:09 crc kubenswrapper[4824]: I1006 11:01:09.297550 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a6fafaeb-02b6-44cf-9ee5-5c6c1996988b-host\") pod \"crc-debug-p6rbb\" (UID: \"a6fafaeb-02b6-44cf-9ee5-5c6c1996988b\") " pod="openshift-must-gather-v7b5k/crc-debug-p6rbb" Oct 06 11:01:09 crc kubenswrapper[4824]: I1006 11:01:09.297618 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mf2vr\" (UniqueName: \"kubernetes.io/projected/a6fafaeb-02b6-44cf-9ee5-5c6c1996988b-kube-api-access-mf2vr\") pod \"crc-debug-p6rbb\" (UID: \"a6fafaeb-02b6-44cf-9ee5-5c6c1996988b\") " pod="openshift-must-gather-v7b5k/crc-debug-p6rbb" Oct 06 11:01:09 crc kubenswrapper[4824]: I1006 11:01:09.399840 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a6fafaeb-02b6-44cf-9ee5-5c6c1996988b-host\") pod \"crc-debug-p6rbb\" (UID: \"a6fafaeb-02b6-44cf-9ee5-5c6c1996988b\") " pod="openshift-must-gather-v7b5k/crc-debug-p6rbb" Oct 06 11:01:09 crc kubenswrapper[4824]: I1006 11:01:09.399923 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mf2vr\" (UniqueName: \"kubernetes.io/projected/a6fafaeb-02b6-44cf-9ee5-5c6c1996988b-kube-api-access-mf2vr\") pod \"crc-debug-p6rbb\" (UID: \"a6fafaeb-02b6-44cf-9ee5-5c6c1996988b\") " pod="openshift-must-gather-v7b5k/crc-debug-p6rbb" Oct 06 11:01:09 crc kubenswrapper[4824]: I1006 11:01:09.400137 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a6fafaeb-02b6-44cf-9ee5-5c6c1996988b-host\") pod \"crc-debug-p6rbb\" (UID: \"a6fafaeb-02b6-44cf-9ee5-5c6c1996988b\") " pod="openshift-must-gather-v7b5k/crc-debug-p6rbb" Oct 06 11:01:09 crc kubenswrapper[4824]: I1006 11:01:09.439412 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mf2vr\" (UniqueName: \"kubernetes.io/projected/a6fafaeb-02b6-44cf-9ee5-5c6c1996988b-kube-api-access-mf2vr\") pod \"crc-debug-p6rbb\" (UID: \"a6fafaeb-02b6-44cf-9ee5-5c6c1996988b\") " pod="openshift-must-gather-v7b5k/crc-debug-p6rbb" Oct 06 11:01:09 crc kubenswrapper[4824]: I1006 11:01:09.476599 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-v7b5k/crc-debug-p6rbb" Oct 06 11:01:10 crc kubenswrapper[4824]: I1006 11:01:10.546383 4824 generic.go:334] "Generic (PLEG): container finished" podID="a6fafaeb-02b6-44cf-9ee5-5c6c1996988b" containerID="73855056e6335fc9be9969e8c2d828e52a40844eb538285376c8fc14d3cf08bd" exitCode=0 Oct 06 11:01:10 crc kubenswrapper[4824]: I1006 11:01:10.546582 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-v7b5k/crc-debug-p6rbb" event={"ID":"a6fafaeb-02b6-44cf-9ee5-5c6c1996988b","Type":"ContainerDied","Data":"73855056e6335fc9be9969e8c2d828e52a40844eb538285376c8fc14d3cf08bd"} Oct 06 11:01:10 crc kubenswrapper[4824]: I1006 11:01:10.546906 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-v7b5k/crc-debug-p6rbb" event={"ID":"a6fafaeb-02b6-44cf-9ee5-5c6c1996988b","Type":"ContainerStarted","Data":"cc7bdba9d6f5cdc501a59be53153c6a88b8a4bfab57726044e6294d4ffc29e73"} Oct 06 11:01:10 crc kubenswrapper[4824]: I1006 11:01:10.601020 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-v7b5k/crc-debug-p6rbb"] Oct 06 11:01:10 crc kubenswrapper[4824]: I1006 11:01:10.617592 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-v7b5k/crc-debug-p6rbb"] Oct 06 11:01:11 crc kubenswrapper[4824]: I1006 11:01:11.662249 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-v7b5k/crc-debug-p6rbb" Oct 06 11:01:11 crc kubenswrapper[4824]: I1006 11:01:11.761366 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a6fafaeb-02b6-44cf-9ee5-5c6c1996988b-host\") pod \"a6fafaeb-02b6-44cf-9ee5-5c6c1996988b\" (UID: \"a6fafaeb-02b6-44cf-9ee5-5c6c1996988b\") " Oct 06 11:01:11 crc kubenswrapper[4824]: I1006 11:01:11.761439 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mf2vr\" (UniqueName: \"kubernetes.io/projected/a6fafaeb-02b6-44cf-9ee5-5c6c1996988b-kube-api-access-mf2vr\") pod \"a6fafaeb-02b6-44cf-9ee5-5c6c1996988b\" (UID: \"a6fafaeb-02b6-44cf-9ee5-5c6c1996988b\") " Oct 06 11:01:11 crc kubenswrapper[4824]: I1006 11:01:11.762693 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a6fafaeb-02b6-44cf-9ee5-5c6c1996988b-host" (OuterVolumeSpecName: "host") pod "a6fafaeb-02b6-44cf-9ee5-5c6c1996988b" (UID: "a6fafaeb-02b6-44cf-9ee5-5c6c1996988b"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 06 11:01:11 crc kubenswrapper[4824]: I1006 11:01:11.767513 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a6fafaeb-02b6-44cf-9ee5-5c6c1996988b-kube-api-access-mf2vr" (OuterVolumeSpecName: "kube-api-access-mf2vr") pod "a6fafaeb-02b6-44cf-9ee5-5c6c1996988b" (UID: "a6fafaeb-02b6-44cf-9ee5-5c6c1996988b"). InnerVolumeSpecName "kube-api-access-mf2vr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 11:01:11 crc kubenswrapper[4824]: I1006 11:01:11.864119 4824 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a6fafaeb-02b6-44cf-9ee5-5c6c1996988b-host\") on node \"crc\" DevicePath \"\"" Oct 06 11:01:11 crc kubenswrapper[4824]: I1006 11:01:11.864154 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mf2vr\" (UniqueName: \"kubernetes.io/projected/a6fafaeb-02b6-44cf-9ee5-5c6c1996988b-kube-api-access-mf2vr\") on node \"crc\" DevicePath \"\"" Oct 06 11:01:12 crc kubenswrapper[4824]: I1006 11:01:12.426369 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_a842b0520c38c46f4cb51faa08e79e3c8193222ff536aee10057c0eb8cqhcsj_65505c4e-e82d-485f-abc6-33be1406fd9a/util/0.log" Oct 06 11:01:12 crc kubenswrapper[4824]: I1006 11:01:12.568569 4824 scope.go:117] "RemoveContainer" containerID="73855056e6335fc9be9969e8c2d828e52a40844eb538285376c8fc14d3cf08bd" Oct 06 11:01:12 crc kubenswrapper[4824]: I1006 11:01:12.568594 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-v7b5k/crc-debug-p6rbb" Oct 06 11:01:12 crc kubenswrapper[4824]: I1006 11:01:12.608554 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_a842b0520c38c46f4cb51faa08e79e3c8193222ff536aee10057c0eb8cqhcsj_65505c4e-e82d-485f-abc6-33be1406fd9a/util/0.log" Oct 06 11:01:12 crc kubenswrapper[4824]: I1006 11:01:12.612266 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_a842b0520c38c46f4cb51faa08e79e3c8193222ff536aee10057c0eb8cqhcsj_65505c4e-e82d-485f-abc6-33be1406fd9a/pull/0.log" Oct 06 11:01:12 crc kubenswrapper[4824]: I1006 11:01:12.684347 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_a842b0520c38c46f4cb51faa08e79e3c8193222ff536aee10057c0eb8cqhcsj_65505c4e-e82d-485f-abc6-33be1406fd9a/pull/0.log" Oct 06 11:01:12 crc kubenswrapper[4824]: I1006 11:01:12.897223 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_a842b0520c38c46f4cb51faa08e79e3c8193222ff536aee10057c0eb8cqhcsj_65505c4e-e82d-485f-abc6-33be1406fd9a/util/0.log" Oct 06 11:01:12 crc kubenswrapper[4824]: I1006 11:01:12.922402 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_a842b0520c38c46f4cb51faa08e79e3c8193222ff536aee10057c0eb8cqhcsj_65505c4e-e82d-485f-abc6-33be1406fd9a/extract/0.log" Oct 06 11:01:12 crc kubenswrapper[4824]: I1006 11:01:12.935693 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_a842b0520c38c46f4cb51faa08e79e3c8193222ff536aee10057c0eb8cqhcsj_65505c4e-e82d-485f-abc6-33be1406fd9a/pull/0.log" Oct 06 11:01:13 crc kubenswrapper[4824]: I1006 11:01:13.101211 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-5f7c849b98-fn8nc_e60e7a0d-ccc6-4b1a-b645-a4802b21d48e/kube-rbac-proxy/0.log" Oct 06 11:01:13 crc kubenswrapper[4824]: I1006 11:01:13.118498 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-7d4d4f8d-4kn7z_b3c0daed-de3b-415e-9166-756cb0f5cab7/kube-rbac-proxy/0.log" Oct 06 11:01:13 crc kubenswrapper[4824]: I1006 11:01:13.236890 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-5f7c849b98-fn8nc_e60e7a0d-ccc6-4b1a-b645-a4802b21d48e/manager/0.log" Oct 06 11:01:13 crc kubenswrapper[4824]: I1006 11:01:13.286384 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a6fafaeb-02b6-44cf-9ee5-5c6c1996988b" path="/var/lib/kubelet/pods/a6fafaeb-02b6-44cf-9ee5-5c6c1996988b/volumes" Oct 06 11:01:13 crc kubenswrapper[4824]: I1006 11:01:13.355711 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-7d4d4f8d-4kn7z_b3c0daed-de3b-415e-9166-756cb0f5cab7/manager/0.log" Oct 06 11:01:13 crc kubenswrapper[4824]: I1006 11:01:13.432276 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-75dfd9b554-hktsj_4ca61d7c-741e-4028-8a75-d26a7b4cbd3a/kube-rbac-proxy/0.log" Oct 06 11:01:13 crc kubenswrapper[4824]: I1006 11:01:13.454651 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-75dfd9b554-hktsj_4ca61d7c-741e-4028-8a75-d26a7b4cbd3a/manager/0.log" Oct 06 11:01:13 crc kubenswrapper[4824]: I1006 11:01:13.585885 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-5568b5d68-r996x_e8ad16ea-027c-4ca7-accd-928ba6faf830/kube-rbac-proxy/0.log" Oct 06 11:01:13 crc kubenswrapper[4824]: I1006 11:01:13.765950 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-5568b5d68-r996x_e8ad16ea-027c-4ca7-accd-928ba6faf830/manager/0.log" Oct 06 11:01:13 crc kubenswrapper[4824]: I1006 11:01:13.811245 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-8f58bc9db-tp4b2_76419802-f7a7-4270-9460-3cc1a9f7f667/kube-rbac-proxy/0.log" Oct 06 11:01:13 crc kubenswrapper[4824]: I1006 11:01:13.818296 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-8f58bc9db-tp4b2_76419802-f7a7-4270-9460-3cc1a9f7f667/manager/0.log" Oct 06 11:01:13 crc kubenswrapper[4824]: I1006 11:01:13.918745 4824 patch_prober.go:28] interesting pod/machine-config-daemon-khgzw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 11:01:13 crc kubenswrapper[4824]: I1006 11:01:13.918831 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 11:01:13 crc kubenswrapper[4824]: I1006 11:01:13.988853 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-54876c876f-swsgf_67824860-53b9-4c9c-95a1-955d9139f6e8/kube-rbac-proxy/0.log" Oct 06 11:01:14 crc kubenswrapper[4824]: I1006 11:01:14.028885 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-54876c876f-swsgf_67824860-53b9-4c9c-95a1-955d9139f6e8/manager/0.log" Oct 06 11:01:14 crc kubenswrapper[4824]: I1006 11:01:14.122125 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-658588b8c9-6884h_98cddaff-3395-473c-afd4-02c849951d51/kube-rbac-proxy/0.log" Oct 06 11:01:14 crc kubenswrapper[4824]: I1006 11:01:14.349922 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-699b87f775-8mvml_09c09b76-0deb-44f1-bc93-b4479b7baa36/manager/0.log" Oct 06 11:01:14 crc kubenswrapper[4824]: I1006 11:01:14.352897 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-699b87f775-8mvml_09c09b76-0deb-44f1-bc93-b4479b7baa36/kube-rbac-proxy/0.log" Oct 06 11:01:14 crc kubenswrapper[4824]: I1006 11:01:14.355116 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-658588b8c9-6884h_98cddaff-3395-473c-afd4-02c849951d51/manager/0.log" Oct 06 11:01:14 crc kubenswrapper[4824]: I1006 11:01:14.573667 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-655d88ccb9-t8sjr_dd382562-c508-4b17-ae00-5db8abe20596/kube-rbac-proxy/0.log" Oct 06 11:01:14 crc kubenswrapper[4824]: I1006 11:01:14.622599 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-655d88ccb9-t8sjr_dd382562-c508-4b17-ae00-5db8abe20596/manager/0.log" Oct 06 11:01:14 crc kubenswrapper[4824]: I1006 11:01:14.697743 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-65d89cfd9f-gts97_21123da0-cde6-45dd-9f50-f3e44a1f78c2/kube-rbac-proxy/0.log" Oct 06 11:01:14 crc kubenswrapper[4824]: I1006 11:01:14.790143 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-65d89cfd9f-gts97_21123da0-cde6-45dd-9f50-f3e44a1f78c2/manager/0.log" Oct 06 11:01:14 crc kubenswrapper[4824]: I1006 11:01:14.883757 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-6cd6d7bdf5-4zc5b_84a43a6c-6456-47ed-ae93-e4476d3cc978/kube-rbac-proxy/0.log" Oct 06 11:01:14 crc kubenswrapper[4824]: I1006 11:01:14.949945 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-6cd6d7bdf5-4zc5b_84a43a6c-6456-47ed-ae93-e4476d3cc978/manager/0.log" Oct 06 11:01:15 crc kubenswrapper[4824]: I1006 11:01:15.062277 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-8d984cc4d-82bf8_36dbefcc-0dac-4435-aaac-5248cd2eb209/kube-rbac-proxy/0.log" Oct 06 11:01:15 crc kubenswrapper[4824]: I1006 11:01:15.132098 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-8d984cc4d-82bf8_36dbefcc-0dac-4435-aaac-5248cd2eb209/manager/0.log" Oct 06 11:01:15 crc kubenswrapper[4824]: I1006 11:01:15.267466 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-7c7fc454ff-fjjsf_dff987b5-f7ce-4eb5-9778-3cb34663824a/kube-rbac-proxy/0.log" Oct 06 11:01:15 crc kubenswrapper[4824]: I1006 11:01:15.340218 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-7c7fc454ff-fjjsf_dff987b5-f7ce-4eb5-9778-3cb34663824a/manager/0.log" Oct 06 11:01:15 crc kubenswrapper[4824]: I1006 11:01:15.364036 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-7468f855d8-kpbcv_c0da600c-cfd5-405a-82ca-ebbf3c474e44/kube-rbac-proxy/0.log" Oct 06 11:01:15 crc kubenswrapper[4824]: I1006 11:01:15.538166 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-7468f855d8-kpbcv_c0da600c-cfd5-405a-82ca-ebbf3c474e44/manager/0.log" Oct 06 11:01:15 crc kubenswrapper[4824]: I1006 11:01:15.542209 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-5dfbbd665cxcwhp_fa5da9bb-805d-420f-adc6-e948cf910b21/kube-rbac-proxy/0.log" Oct 06 11:01:15 crc kubenswrapper[4824]: I1006 11:01:15.570869 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-5dfbbd665cxcwhp_fa5da9bb-805d-420f-adc6-e948cf910b21/manager/0.log" Oct 06 11:01:15 crc kubenswrapper[4824]: I1006 11:01:15.728084 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-844bbc956-tjxqg_993a8c7b-2aac-4de6-934d-b591cb6bfffe/kube-rbac-proxy/0.log" Oct 06 11:01:15 crc kubenswrapper[4824]: I1006 11:01:15.972269 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-58f9547c95-dncvl_c1fed37a-a4ba-43eb-82fa-5b04f6a39ead/kube-rbac-proxy/0.log" Oct 06 11:01:16 crc kubenswrapper[4824]: I1006 11:01:16.139432 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-58f9547c95-dncvl_c1fed37a-a4ba-43eb-82fa-5b04f6a39ead/operator/0.log" Oct 06 11:01:16 crc kubenswrapper[4824]: I1006 11:01:16.205950 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-dlhm7_bd9ffa83-7e81-4322-9507-e92aeec0276d/registry-server/0.log" Oct 06 11:01:16 crc kubenswrapper[4824]: I1006 11:01:16.369665 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-579449c7d5-w7k8s_12802c32-cb1d-452b-8240-9a5e1e40a163/kube-rbac-proxy/0.log" Oct 06 11:01:16 crc kubenswrapper[4824]: I1006 11:01:16.485694 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-579449c7d5-w7k8s_12802c32-cb1d-452b-8240-9a5e1e40a163/manager/0.log" Oct 06 11:01:16 crc kubenswrapper[4824]: I1006 11:01:16.620922 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-54689d9f88-8qfvp_1778e737-b572-417b-931d-0eec04a89ac9/kube-rbac-proxy/0.log" Oct 06 11:01:16 crc kubenswrapper[4824]: I1006 11:01:16.725399 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-54689d9f88-8qfvp_1778e737-b572-417b-931d-0eec04a89ac9/manager/0.log" Oct 06 11:01:16 crc kubenswrapper[4824]: I1006 11:01:16.797965 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-5f97d8c699-c8wx2_1f00bc57-93de-431a-adcf-d1e1221121a1/operator/0.log" Oct 06 11:01:16 crc kubenswrapper[4824]: I1006 11:01:16.987887 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-6859f9b676-n84pw_b32a09cc-901f-41f2-9912-628bbc33da7b/kube-rbac-proxy/0.log" Oct 06 11:01:17 crc kubenswrapper[4824]: I1006 11:01:17.001750 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-6859f9b676-n84pw_b32a09cc-901f-41f2-9912-628bbc33da7b/manager/0.log" Oct 06 11:01:17 crc kubenswrapper[4824]: I1006 11:01:17.053607 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-844bbc956-tjxqg_993a8c7b-2aac-4de6-934d-b591cb6bfffe/manager/0.log" Oct 06 11:01:17 crc kubenswrapper[4824]: I1006 11:01:17.126642 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-84c678c567-wzzrh_7c4f21a3-07f8-46d8-bd7b-e2a24bf280d8/kube-rbac-proxy/0.log" Oct 06 11:01:17 crc kubenswrapper[4824]: I1006 11:01:17.280351 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-84c678c567-wzzrh_7c4f21a3-07f8-46d8-bd7b-e2a24bf280d8/manager/0.log" Oct 06 11:01:17 crc kubenswrapper[4824]: I1006 11:01:17.305676 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5cd5cb47d7-6qgln_755275e4-c198-49df-90b5-0688a4fb8228/manager/0.log" Oct 06 11:01:17 crc kubenswrapper[4824]: I1006 11:01:17.313412 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5cd5cb47d7-6qgln_755275e4-c198-49df-90b5-0688a4fb8228/kube-rbac-proxy/0.log" Oct 06 11:01:17 crc kubenswrapper[4824]: I1006 11:01:17.448871 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-6cbc6dd547-dglbk_2a48640f-172c-4e24-8c75-5c36d26ae1aa/kube-rbac-proxy/0.log" Oct 06 11:01:17 crc kubenswrapper[4824]: I1006 11:01:17.502850 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-6cbc6dd547-dglbk_2a48640f-172c-4e24-8c75-5c36d26ae1aa/manager/0.log" Oct 06 11:01:36 crc kubenswrapper[4824]: I1006 11:01:36.903342 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-fn7dj_2a76ccaa-9249-46ba-bbc5-dbbbc33b06bb/control-plane-machine-set-operator/0.log" Oct 06 11:01:37 crc kubenswrapper[4824]: I1006 11:01:37.125638 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-fjdwv_b349b0af-0ba8-409f-a915-7fb06b2e3774/kube-rbac-proxy/0.log" Oct 06 11:01:37 crc kubenswrapper[4824]: I1006 11:01:37.159646 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-fjdwv_b349b0af-0ba8-409f-a915-7fb06b2e3774/machine-api-operator/0.log" Oct 06 11:01:43 crc kubenswrapper[4824]: I1006 11:01:43.915610 4824 patch_prober.go:28] interesting pod/machine-config-daemon-khgzw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 11:01:43 crc kubenswrapper[4824]: I1006 11:01:43.916479 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 11:01:52 crc kubenswrapper[4824]: I1006 11:01:52.029939 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-bdn2f_bde1f651-2eda-4d2d-8ab5-5fcdcd4580a4/cert-manager-controller/0.log" Oct 06 11:01:52 crc kubenswrapper[4824]: I1006 11:01:52.250238 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-htx98_d6f420d5-ed6d-4201-b78f-0bfa304e3b6b/cert-manager-webhook/0.log" Oct 06 11:01:52 crc kubenswrapper[4824]: I1006 11:01:52.261331 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-bbqpq_b5759913-9cb4-4012-bffb-fe2f8085543c/cert-manager-cainjector/0.log" Oct 06 11:02:06 crc kubenswrapper[4824]: I1006 11:02:06.469440 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-6b874cbd85-7nwh2_00271786-ab42-4507-a545-8c5bdf1b976d/nmstate-console-plugin/0.log" Oct 06 11:02:06 crc kubenswrapper[4824]: I1006 11:02:06.680811 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-jbv5q_3048d3eb-7f41-4b26-92b3-6d60f5a3fa18/nmstate-handler/0.log" Oct 06 11:02:06 crc kubenswrapper[4824]: I1006 11:02:06.738620 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-6vncz_355b7d6a-3ab4-4d16-8923-f15646b3af54/nmstate-metrics/0.log" Oct 06 11:02:06 crc kubenswrapper[4824]: I1006 11:02:06.750144 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-6vncz_355b7d6a-3ab4-4d16-8923-f15646b3af54/kube-rbac-proxy/0.log" Oct 06 11:02:06 crc kubenswrapper[4824]: I1006 11:02:06.946481 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-858ddd8f98-grzj6_230d69e7-8f29-41a3-af2f-b1b93c58e8c2/nmstate-operator/0.log" Oct 06 11:02:06 crc kubenswrapper[4824]: I1006 11:02:06.994369 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-6cdbc54649-b7dd2_b706331c-d771-4212-950f-6e3fd16faa43/nmstate-webhook/0.log" Oct 06 11:02:13 crc kubenswrapper[4824]: I1006 11:02:13.915332 4824 patch_prober.go:28] interesting pod/machine-config-daemon-khgzw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 11:02:13 crc kubenswrapper[4824]: I1006 11:02:13.917805 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 11:02:13 crc kubenswrapper[4824]: I1006 11:02:13.917948 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" Oct 06 11:02:13 crc kubenswrapper[4824]: I1006 11:02:13.919493 4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5a47a5d2e5b1ba0ad6ad09ca0bd7f5568851664618ce5ad15f9b3204fecd76c5"} pod="openshift-machine-config-operator/machine-config-daemon-khgzw" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 06 11:02:13 crc kubenswrapper[4824]: I1006 11:02:13.919675 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" containerName="machine-config-daemon" containerID="cri-o://5a47a5d2e5b1ba0ad6ad09ca0bd7f5568851664618ce5ad15f9b3204fecd76c5" gracePeriod=600 Oct 06 11:02:14 crc kubenswrapper[4824]: E1006 11:02:14.040237 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-khgzw_openshift-machine-config-operator(e1140c8c-93f1-4fce-8c06-a69561ff8a6d)\"" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" Oct 06 11:02:14 crc kubenswrapper[4824]: I1006 11:02:14.234004 4824 generic.go:334] "Generic (PLEG): container finished" podID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" containerID="5a47a5d2e5b1ba0ad6ad09ca0bd7f5568851664618ce5ad15f9b3204fecd76c5" exitCode=0 Oct 06 11:02:14 crc kubenswrapper[4824]: I1006 11:02:14.234051 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" event={"ID":"e1140c8c-93f1-4fce-8c06-a69561ff8a6d","Type":"ContainerDied","Data":"5a47a5d2e5b1ba0ad6ad09ca0bd7f5568851664618ce5ad15f9b3204fecd76c5"} Oct 06 11:02:14 crc kubenswrapper[4824]: I1006 11:02:14.234088 4824 scope.go:117] "RemoveContainer" containerID="d960e8d823dfd2a405439bf2eac79b76e3ee46509993c3394eac4ba53ab10143" Oct 06 11:02:14 crc kubenswrapper[4824]: I1006 11:02:14.234900 4824 scope.go:117] "RemoveContainer" containerID="5a47a5d2e5b1ba0ad6ad09ca0bd7f5568851664618ce5ad15f9b3204fecd76c5" Oct 06 11:02:14 crc kubenswrapper[4824]: E1006 11:02:14.235286 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-khgzw_openshift-machine-config-operator(e1140c8c-93f1-4fce-8c06-a69561ff8a6d)\"" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" Oct 06 11:02:22 crc kubenswrapper[4824]: I1006 11:02:22.069243 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-sldcd_129eb734-6e0d-4785-b25a-4c68af80d8c6/kube-rbac-proxy/0.log" Oct 06 11:02:22 crc kubenswrapper[4824]: I1006 11:02:22.125331 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-sldcd_129eb734-6e0d-4785-b25a-4c68af80d8c6/controller/0.log" Oct 06 11:02:22 crc kubenswrapper[4824]: I1006 11:02:22.286003 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-64bf5d555-d9kq5_e0836691-2fe5-4e7c-a374-abe71aeb2278/frr-k8s-webhook-server/0.log" Oct 06 11:02:22 crc kubenswrapper[4824]: I1006 11:02:22.348089 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-x6lsw_2f2ec220-4ca6-4b43-9e89-fdae758ab063/cp-frr-files/0.log" Oct 06 11:02:22 crc kubenswrapper[4824]: I1006 11:02:22.558786 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-x6lsw_2f2ec220-4ca6-4b43-9e89-fdae758ab063/cp-reloader/0.log" Oct 06 11:02:22 crc kubenswrapper[4824]: I1006 11:02:22.574454 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-x6lsw_2f2ec220-4ca6-4b43-9e89-fdae758ab063/cp-reloader/0.log" Oct 06 11:02:22 crc kubenswrapper[4824]: I1006 11:02:22.577213 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-x6lsw_2f2ec220-4ca6-4b43-9e89-fdae758ab063/cp-frr-files/0.log" Oct 06 11:02:22 crc kubenswrapper[4824]: I1006 11:02:22.578351 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-x6lsw_2f2ec220-4ca6-4b43-9e89-fdae758ab063/cp-metrics/0.log" Oct 06 11:02:22 crc kubenswrapper[4824]: I1006 11:02:22.785634 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-x6lsw_2f2ec220-4ca6-4b43-9e89-fdae758ab063/cp-frr-files/0.log" Oct 06 11:02:22 crc kubenswrapper[4824]: I1006 11:02:22.800911 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-x6lsw_2f2ec220-4ca6-4b43-9e89-fdae758ab063/cp-reloader/0.log" Oct 06 11:02:22 crc kubenswrapper[4824]: I1006 11:02:22.825598 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-x6lsw_2f2ec220-4ca6-4b43-9e89-fdae758ab063/cp-metrics/0.log" Oct 06 11:02:22 crc kubenswrapper[4824]: I1006 11:02:22.827432 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-x6lsw_2f2ec220-4ca6-4b43-9e89-fdae758ab063/cp-metrics/0.log" Oct 06 11:02:23 crc kubenswrapper[4824]: I1006 11:02:23.053912 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-x6lsw_2f2ec220-4ca6-4b43-9e89-fdae758ab063/cp-reloader/0.log" Oct 06 11:02:23 crc kubenswrapper[4824]: I1006 11:02:23.071595 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-x6lsw_2f2ec220-4ca6-4b43-9e89-fdae758ab063/cp-frr-files/0.log" Oct 06 11:02:23 crc kubenswrapper[4824]: I1006 11:02:23.112169 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-x6lsw_2f2ec220-4ca6-4b43-9e89-fdae758ab063/cp-metrics/0.log" Oct 06 11:02:23 crc kubenswrapper[4824]: I1006 11:02:23.114039 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-x6lsw_2f2ec220-4ca6-4b43-9e89-fdae758ab063/controller/0.log" Oct 06 11:02:23 crc kubenswrapper[4824]: I1006 11:02:23.321926 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-x6lsw_2f2ec220-4ca6-4b43-9e89-fdae758ab063/frr-metrics/0.log" Oct 06 11:02:23 crc kubenswrapper[4824]: I1006 11:02:23.349859 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-x6lsw_2f2ec220-4ca6-4b43-9e89-fdae758ab063/kube-rbac-proxy-frr/0.log" Oct 06 11:02:23 crc kubenswrapper[4824]: I1006 11:02:23.390277 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-x6lsw_2f2ec220-4ca6-4b43-9e89-fdae758ab063/kube-rbac-proxy/0.log" Oct 06 11:02:23 crc kubenswrapper[4824]: I1006 11:02:23.598640 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-x6lsw_2f2ec220-4ca6-4b43-9e89-fdae758ab063/reloader/0.log" Oct 06 11:02:23 crc kubenswrapper[4824]: I1006 11:02:23.685924 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-57744c8fb4-85v7z_d8226542-256f-4e73-9d78-0d8f261108d4/manager/0.log" Oct 06 11:02:23 crc kubenswrapper[4824]: I1006 11:02:23.907874 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-56788496c6-xdgkc_1d2065c1-e8a8-4dfd-bf6e-a8f701cbe060/webhook-server/0.log" Oct 06 11:02:24 crc kubenswrapper[4824]: I1006 11:02:24.202476 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-tcgrw_5c0a8aa4-2b23-4265-8b07-e075b355fc6c/kube-rbac-proxy/0.log" Oct 06 11:02:24 crc kubenswrapper[4824]: I1006 11:02:24.745440 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-tcgrw_5c0a8aa4-2b23-4265-8b07-e075b355fc6c/speaker/0.log" Oct 06 11:02:24 crc kubenswrapper[4824]: I1006 11:02:24.946865 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-x6lsw_2f2ec220-4ca6-4b43-9e89-fdae758ab063/frr/0.log" Oct 06 11:02:29 crc kubenswrapper[4824]: I1006 11:02:29.275491 4824 scope.go:117] "RemoveContainer" containerID="5a47a5d2e5b1ba0ad6ad09ca0bd7f5568851664618ce5ad15f9b3204fecd76c5" Oct 06 11:02:29 crc kubenswrapper[4824]: E1006 11:02:29.276506 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-khgzw_openshift-machine-config-operator(e1140c8c-93f1-4fce-8c06-a69561ff8a6d)\"" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" Oct 06 11:02:38 crc kubenswrapper[4824]: I1006 11:02:38.255085 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2frg7w_b186e32b-a095-4c20-816c-9e83ea7e261f/util/0.log" Oct 06 11:02:38 crc kubenswrapper[4824]: I1006 11:02:38.500321 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2frg7w_b186e32b-a095-4c20-816c-9e83ea7e261f/util/0.log" Oct 06 11:02:38 crc kubenswrapper[4824]: I1006 11:02:38.531677 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2frg7w_b186e32b-a095-4c20-816c-9e83ea7e261f/pull/0.log" Oct 06 11:02:38 crc kubenswrapper[4824]: I1006 11:02:38.539371 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2frg7w_b186e32b-a095-4c20-816c-9e83ea7e261f/pull/0.log" Oct 06 11:02:38 crc kubenswrapper[4824]: I1006 11:02:38.752859 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2frg7w_b186e32b-a095-4c20-816c-9e83ea7e261f/extract/0.log" Oct 06 11:02:38 crc kubenswrapper[4824]: I1006 11:02:38.768029 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2frg7w_b186e32b-a095-4c20-816c-9e83ea7e261f/pull/0.log" Oct 06 11:02:38 crc kubenswrapper[4824]: I1006 11:02:38.811543 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2frg7w_b186e32b-a095-4c20-816c-9e83ea7e261f/util/0.log" Oct 06 11:02:38 crc kubenswrapper[4824]: I1006 11:02:38.941265 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-mblh7_2dc6af8a-35a7-4fa8-9e03-eb6df841f127/extract-utilities/0.log" Oct 06 11:02:39 crc kubenswrapper[4824]: I1006 11:02:39.177056 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-mblh7_2dc6af8a-35a7-4fa8-9e03-eb6df841f127/extract-content/0.log" Oct 06 11:02:39 crc kubenswrapper[4824]: I1006 11:02:39.184260 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-mblh7_2dc6af8a-35a7-4fa8-9e03-eb6df841f127/extract-utilities/0.log" Oct 06 11:02:39 crc kubenswrapper[4824]: I1006 11:02:39.195698 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-mblh7_2dc6af8a-35a7-4fa8-9e03-eb6df841f127/extract-content/0.log" Oct 06 11:02:39 crc kubenswrapper[4824]: I1006 11:02:39.403761 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-mblh7_2dc6af8a-35a7-4fa8-9e03-eb6df841f127/extract-utilities/0.log" Oct 06 11:02:39 crc kubenswrapper[4824]: I1006 11:02:39.430085 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-mblh7_2dc6af8a-35a7-4fa8-9e03-eb6df841f127/extract-content/0.log" Oct 06 11:02:39 crc kubenswrapper[4824]: I1006 11:02:39.680377 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-s2nbh_4519644a-379b-451a-9dd9-fb528ced61bf/extract-utilities/0.log" Oct 06 11:02:39 crc kubenswrapper[4824]: I1006 11:02:39.898333 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-s2nbh_4519644a-379b-451a-9dd9-fb528ced61bf/extract-utilities/0.log" Oct 06 11:02:39 crc kubenswrapper[4824]: I1006 11:02:39.953050 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-s2nbh_4519644a-379b-451a-9dd9-fb528ced61bf/extract-content/0.log" Oct 06 11:02:39 crc kubenswrapper[4824]: I1006 11:02:39.995895 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-s2nbh_4519644a-379b-451a-9dd9-fb528ced61bf/extract-content/0.log" Oct 06 11:02:40 crc kubenswrapper[4824]: I1006 11:02:40.204106 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-mblh7_2dc6af8a-35a7-4fa8-9e03-eb6df841f127/registry-server/0.log" Oct 06 11:02:40 crc kubenswrapper[4824]: I1006 11:02:40.289747 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-s2nbh_4519644a-379b-451a-9dd9-fb528ced61bf/extract-content/0.log" Oct 06 11:02:40 crc kubenswrapper[4824]: I1006 11:02:40.314558 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-s2nbh_4519644a-379b-451a-9dd9-fb528ced61bf/extract-utilities/0.log" Oct 06 11:02:40 crc kubenswrapper[4824]: I1006 11:02:40.554847 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cxdt8w_5e47aa0e-c660-4f88-8fb1-8c873268879d/util/0.log" Oct 06 11:02:40 crc kubenswrapper[4824]: I1006 11:02:40.860075 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cxdt8w_5e47aa0e-c660-4f88-8fb1-8c873268879d/pull/0.log" Oct 06 11:02:40 crc kubenswrapper[4824]: I1006 11:02:40.865546 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cxdt8w_5e47aa0e-c660-4f88-8fb1-8c873268879d/util/0.log" Oct 06 11:02:40 crc kubenswrapper[4824]: I1006 11:02:40.916094 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cxdt8w_5e47aa0e-c660-4f88-8fb1-8c873268879d/pull/0.log" Oct 06 11:02:41 crc kubenswrapper[4824]: I1006 11:02:41.175421 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-s2nbh_4519644a-379b-451a-9dd9-fb528ced61bf/registry-server/0.log" Oct 06 11:02:41 crc kubenswrapper[4824]: I1006 11:02:41.180758 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cxdt8w_5e47aa0e-c660-4f88-8fb1-8c873268879d/pull/0.log" Oct 06 11:02:41 crc kubenswrapper[4824]: I1006 11:02:41.194875 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cxdt8w_5e47aa0e-c660-4f88-8fb1-8c873268879d/util/0.log" Oct 06 11:02:41 crc kubenswrapper[4824]: I1006 11:02:41.230243 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cxdt8w_5e47aa0e-c660-4f88-8fb1-8c873268879d/extract/0.log" Oct 06 11:02:41 crc kubenswrapper[4824]: I1006 11:02:41.386231 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-db6vn_a3a514b8-f221-465e-bbb2-3c88bcb4f622/marketplace-operator/0.log" Oct 06 11:02:41 crc kubenswrapper[4824]: I1006 11:02:41.505427 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-xdszq_504bf3d4-9589-4a5e-ad25-e16f783e0115/extract-utilities/0.log" Oct 06 11:02:41 crc kubenswrapper[4824]: I1006 11:02:41.674202 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-xdszq_504bf3d4-9589-4a5e-ad25-e16f783e0115/extract-utilities/0.log" Oct 06 11:02:41 crc kubenswrapper[4824]: I1006 11:02:41.699031 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-xdszq_504bf3d4-9589-4a5e-ad25-e16f783e0115/extract-content/0.log" Oct 06 11:02:41 crc kubenswrapper[4824]: I1006 11:02:41.741206 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-xdszq_504bf3d4-9589-4a5e-ad25-e16f783e0115/extract-content/0.log" Oct 06 11:02:41 crc kubenswrapper[4824]: I1006 11:02:41.907935 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-xdszq_504bf3d4-9589-4a5e-ad25-e16f783e0115/extract-utilities/0.log" Oct 06 11:02:41 crc kubenswrapper[4824]: I1006 11:02:41.924821 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-xdszq_504bf3d4-9589-4a5e-ad25-e16f783e0115/extract-content/0.log" Oct 06 11:02:42 crc kubenswrapper[4824]: I1006 11:02:42.070071 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-xdszq_504bf3d4-9589-4a5e-ad25-e16f783e0115/registry-server/0.log" Oct 06 11:02:42 crc kubenswrapper[4824]: I1006 11:02:42.157454 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-d8s5f_bcb79b86-65c0-44c1-9ce2-588ad66206a6/extract-utilities/0.log" Oct 06 11:02:42 crc kubenswrapper[4824]: I1006 11:02:42.314869 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-d8s5f_bcb79b86-65c0-44c1-9ce2-588ad66206a6/extract-utilities/0.log" Oct 06 11:02:42 crc kubenswrapper[4824]: I1006 11:02:42.338865 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-d8s5f_bcb79b86-65c0-44c1-9ce2-588ad66206a6/extract-content/0.log" Oct 06 11:02:42 crc kubenswrapper[4824]: I1006 11:02:42.347067 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-d8s5f_bcb79b86-65c0-44c1-9ce2-588ad66206a6/extract-content/0.log" Oct 06 11:02:42 crc kubenswrapper[4824]: I1006 11:02:42.522892 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-d8s5f_bcb79b86-65c0-44c1-9ce2-588ad66206a6/extract-utilities/0.log" Oct 06 11:02:42 crc kubenswrapper[4824]: I1006 11:02:42.568060 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-d8s5f_bcb79b86-65c0-44c1-9ce2-588ad66206a6/extract-content/0.log" Oct 06 11:02:42 crc kubenswrapper[4824]: I1006 11:02:42.786514 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-d8s5f_bcb79b86-65c0-44c1-9ce2-588ad66206a6/registry-server/0.log" Oct 06 11:02:44 crc kubenswrapper[4824]: I1006 11:02:44.274129 4824 scope.go:117] "RemoveContainer" containerID="5a47a5d2e5b1ba0ad6ad09ca0bd7f5568851664618ce5ad15f9b3204fecd76c5" Oct 06 11:02:44 crc kubenswrapper[4824]: E1006 11:02:44.274688 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-khgzw_openshift-machine-config-operator(e1140c8c-93f1-4fce-8c06-a69561ff8a6d)\"" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" Oct 06 11:02:57 crc kubenswrapper[4824]: I1006 11:02:57.274255 4824 scope.go:117] "RemoveContainer" containerID="5a47a5d2e5b1ba0ad6ad09ca0bd7f5568851664618ce5ad15f9b3204fecd76c5" Oct 06 11:02:57 crc kubenswrapper[4824]: E1006 11:02:57.276300 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-khgzw_openshift-machine-config-operator(e1140c8c-93f1-4fce-8c06-a69561ff8a6d)\"" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" Oct 06 11:03:10 crc kubenswrapper[4824]: I1006 11:03:10.274556 4824 scope.go:117] "RemoveContainer" containerID="5a47a5d2e5b1ba0ad6ad09ca0bd7f5568851664618ce5ad15f9b3204fecd76c5" Oct 06 11:03:10 crc kubenswrapper[4824]: E1006 11:03:10.275508 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-khgzw_openshift-machine-config-operator(e1140c8c-93f1-4fce-8c06-a69561ff8a6d)\"" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" Oct 06 11:03:21 crc kubenswrapper[4824]: I1006 11:03:21.291255 4824 scope.go:117] "RemoveContainer" containerID="5a47a5d2e5b1ba0ad6ad09ca0bd7f5568851664618ce5ad15f9b3204fecd76c5" Oct 06 11:03:21 crc kubenswrapper[4824]: E1006 11:03:21.292359 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-khgzw_openshift-machine-config-operator(e1140c8c-93f1-4fce-8c06-a69561ff8a6d)\"" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" Oct 06 11:03:36 crc kubenswrapper[4824]: I1006 11:03:36.275803 4824 scope.go:117] "RemoveContainer" containerID="5a47a5d2e5b1ba0ad6ad09ca0bd7f5568851664618ce5ad15f9b3204fecd76c5" Oct 06 11:03:36 crc kubenswrapper[4824]: E1006 11:03:36.277424 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-khgzw_openshift-machine-config-operator(e1140c8c-93f1-4fce-8c06-a69561ff8a6d)\"" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" Oct 06 11:03:41 crc kubenswrapper[4824]: I1006 11:03:41.789128 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-fxhmj"] Oct 06 11:03:41 crc kubenswrapper[4824]: E1006 11:03:41.789998 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6fafaeb-02b6-44cf-9ee5-5c6c1996988b" containerName="container-00" Oct 06 11:03:41 crc kubenswrapper[4824]: I1006 11:03:41.790012 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6fafaeb-02b6-44cf-9ee5-5c6c1996988b" containerName="container-00" Oct 06 11:03:41 crc kubenswrapper[4824]: I1006 11:03:41.790197 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="a6fafaeb-02b6-44cf-9ee5-5c6c1996988b" containerName="container-00" Oct 06 11:03:41 crc kubenswrapper[4824]: I1006 11:03:41.793989 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fxhmj" Oct 06 11:03:41 crc kubenswrapper[4824]: I1006 11:03:41.804390 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-fxhmj"] Oct 06 11:03:41 crc kubenswrapper[4824]: I1006 11:03:41.848788 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lcnzn\" (UniqueName: \"kubernetes.io/projected/78440166-ae09-45fc-950e-2f8b244f1d8a-kube-api-access-lcnzn\") pod \"community-operators-fxhmj\" (UID: \"78440166-ae09-45fc-950e-2f8b244f1d8a\") " pod="openshift-marketplace/community-operators-fxhmj" Oct 06 11:03:41 crc kubenswrapper[4824]: I1006 11:03:41.849291 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/78440166-ae09-45fc-950e-2f8b244f1d8a-catalog-content\") pod \"community-operators-fxhmj\" (UID: \"78440166-ae09-45fc-950e-2f8b244f1d8a\") " pod="openshift-marketplace/community-operators-fxhmj" Oct 06 11:03:41 crc kubenswrapper[4824]: I1006 11:03:41.849840 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/78440166-ae09-45fc-950e-2f8b244f1d8a-utilities\") pod \"community-operators-fxhmj\" (UID: \"78440166-ae09-45fc-950e-2f8b244f1d8a\") " pod="openshift-marketplace/community-operators-fxhmj" Oct 06 11:03:41 crc kubenswrapper[4824]: I1006 11:03:41.952057 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/78440166-ae09-45fc-950e-2f8b244f1d8a-utilities\") pod \"community-operators-fxhmj\" (UID: \"78440166-ae09-45fc-950e-2f8b244f1d8a\") " pod="openshift-marketplace/community-operators-fxhmj" Oct 06 11:03:41 crc kubenswrapper[4824]: I1006 11:03:41.952130 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lcnzn\" (UniqueName: \"kubernetes.io/projected/78440166-ae09-45fc-950e-2f8b244f1d8a-kube-api-access-lcnzn\") pod \"community-operators-fxhmj\" (UID: \"78440166-ae09-45fc-950e-2f8b244f1d8a\") " pod="openshift-marketplace/community-operators-fxhmj" Oct 06 11:03:41 crc kubenswrapper[4824]: I1006 11:03:41.952217 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/78440166-ae09-45fc-950e-2f8b244f1d8a-catalog-content\") pod \"community-operators-fxhmj\" (UID: \"78440166-ae09-45fc-950e-2f8b244f1d8a\") " pod="openshift-marketplace/community-operators-fxhmj" Oct 06 11:03:41 crc kubenswrapper[4824]: I1006 11:03:41.952768 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/78440166-ae09-45fc-950e-2f8b244f1d8a-utilities\") pod \"community-operators-fxhmj\" (UID: \"78440166-ae09-45fc-950e-2f8b244f1d8a\") " pod="openshift-marketplace/community-operators-fxhmj" Oct 06 11:03:41 crc kubenswrapper[4824]: I1006 11:03:41.952808 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/78440166-ae09-45fc-950e-2f8b244f1d8a-catalog-content\") pod \"community-operators-fxhmj\" (UID: \"78440166-ae09-45fc-950e-2f8b244f1d8a\") " pod="openshift-marketplace/community-operators-fxhmj" Oct 06 11:03:41 crc kubenswrapper[4824]: I1006 11:03:41.975680 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lcnzn\" (UniqueName: \"kubernetes.io/projected/78440166-ae09-45fc-950e-2f8b244f1d8a-kube-api-access-lcnzn\") pod \"community-operators-fxhmj\" (UID: \"78440166-ae09-45fc-950e-2f8b244f1d8a\") " pod="openshift-marketplace/community-operators-fxhmj" Oct 06 11:03:42 crc kubenswrapper[4824]: I1006 11:03:42.116767 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fxhmj" Oct 06 11:03:42 crc kubenswrapper[4824]: I1006 11:03:42.770497 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-fxhmj"] Oct 06 11:03:43 crc kubenswrapper[4824]: I1006 11:03:43.245048 4824 generic.go:334] "Generic (PLEG): container finished" podID="78440166-ae09-45fc-950e-2f8b244f1d8a" containerID="1a6e7e6f37201cd2667c01cf41ff34662c8aafbf971d12424b087a2a9165bcc4" exitCode=0 Oct 06 11:03:43 crc kubenswrapper[4824]: I1006 11:03:43.245192 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fxhmj" event={"ID":"78440166-ae09-45fc-950e-2f8b244f1d8a","Type":"ContainerDied","Data":"1a6e7e6f37201cd2667c01cf41ff34662c8aafbf971d12424b087a2a9165bcc4"} Oct 06 11:03:43 crc kubenswrapper[4824]: I1006 11:03:43.245483 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fxhmj" event={"ID":"78440166-ae09-45fc-950e-2f8b244f1d8a","Type":"ContainerStarted","Data":"b31c89d3b808aa4f6c3cee204ed698fdd545e2c8c427feab3bb6f6fa3eac9593"} Oct 06 11:03:43 crc kubenswrapper[4824]: I1006 11:03:43.254372 4824 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 06 11:03:44 crc kubenswrapper[4824]: I1006 11:03:44.260366 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fxhmj" event={"ID":"78440166-ae09-45fc-950e-2f8b244f1d8a","Type":"ContainerStarted","Data":"f35a01e43fa3d34a51697539ce5b936bedc03ed400fd0859d7fcaf38d5a61a42"} Oct 06 11:03:45 crc kubenswrapper[4824]: I1006 11:03:45.307445 4824 generic.go:334] "Generic (PLEG): container finished" podID="78440166-ae09-45fc-950e-2f8b244f1d8a" containerID="f35a01e43fa3d34a51697539ce5b936bedc03ed400fd0859d7fcaf38d5a61a42" exitCode=0 Oct 06 11:03:45 crc kubenswrapper[4824]: I1006 11:03:45.309827 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fxhmj" event={"ID":"78440166-ae09-45fc-950e-2f8b244f1d8a","Type":"ContainerDied","Data":"f35a01e43fa3d34a51697539ce5b936bedc03ed400fd0859d7fcaf38d5a61a42"} Oct 06 11:03:47 crc kubenswrapper[4824]: I1006 11:03:47.351035 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fxhmj" event={"ID":"78440166-ae09-45fc-950e-2f8b244f1d8a","Type":"ContainerStarted","Data":"6f34e43b77cded2748c96117247246ee85888bf79005c17ad312e49a72ec7286"} Oct 06 11:03:47 crc kubenswrapper[4824]: I1006 11:03:47.378063 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-fxhmj" podStartSLOduration=3.126043408 podStartE2EDuration="6.378033499s" podCreationTimestamp="2025-10-06 11:03:41 +0000 UTC" firstStartedPulling="2025-10-06 11:03:43.254102881 +0000 UTC m=+3932.618525742" lastFinishedPulling="2025-10-06 11:03:46.506092972 +0000 UTC m=+3935.870515833" observedRunningTime="2025-10-06 11:03:47.375234269 +0000 UTC m=+3936.739657140" watchObservedRunningTime="2025-10-06 11:03:47.378033499 +0000 UTC m=+3936.742456370" Oct 06 11:03:50 crc kubenswrapper[4824]: I1006 11:03:50.274186 4824 scope.go:117] "RemoveContainer" containerID="5a47a5d2e5b1ba0ad6ad09ca0bd7f5568851664618ce5ad15f9b3204fecd76c5" Oct 06 11:03:50 crc kubenswrapper[4824]: E1006 11:03:50.274907 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-khgzw_openshift-machine-config-operator(e1140c8c-93f1-4fce-8c06-a69561ff8a6d)\"" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" Oct 06 11:03:52 crc kubenswrapper[4824]: I1006 11:03:52.117160 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-fxhmj" Oct 06 11:03:52 crc kubenswrapper[4824]: I1006 11:03:52.117740 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-fxhmj" Oct 06 11:03:52 crc kubenswrapper[4824]: I1006 11:03:52.175303 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-fxhmj" Oct 06 11:03:52 crc kubenswrapper[4824]: I1006 11:03:52.473562 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-fxhmj" Oct 06 11:03:52 crc kubenswrapper[4824]: I1006 11:03:52.596745 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-fxhmj"] Oct 06 11:03:54 crc kubenswrapper[4824]: I1006 11:03:54.425475 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-fxhmj" podUID="78440166-ae09-45fc-950e-2f8b244f1d8a" containerName="registry-server" containerID="cri-o://6f34e43b77cded2748c96117247246ee85888bf79005c17ad312e49a72ec7286" gracePeriod=2 Oct 06 11:03:55 crc kubenswrapper[4824]: I1006 11:03:55.155184 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fxhmj" Oct 06 11:03:55 crc kubenswrapper[4824]: I1006 11:03:55.286266 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/78440166-ae09-45fc-950e-2f8b244f1d8a-catalog-content\") pod \"78440166-ae09-45fc-950e-2f8b244f1d8a\" (UID: \"78440166-ae09-45fc-950e-2f8b244f1d8a\") " Oct 06 11:03:55 crc kubenswrapper[4824]: I1006 11:03:55.286693 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/78440166-ae09-45fc-950e-2f8b244f1d8a-utilities\") pod \"78440166-ae09-45fc-950e-2f8b244f1d8a\" (UID: \"78440166-ae09-45fc-950e-2f8b244f1d8a\") " Oct 06 11:03:55 crc kubenswrapper[4824]: I1006 11:03:55.286749 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lcnzn\" (UniqueName: \"kubernetes.io/projected/78440166-ae09-45fc-950e-2f8b244f1d8a-kube-api-access-lcnzn\") pod \"78440166-ae09-45fc-950e-2f8b244f1d8a\" (UID: \"78440166-ae09-45fc-950e-2f8b244f1d8a\") " Oct 06 11:03:55 crc kubenswrapper[4824]: I1006 11:03:55.288874 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/78440166-ae09-45fc-950e-2f8b244f1d8a-utilities" (OuterVolumeSpecName: "utilities") pod "78440166-ae09-45fc-950e-2f8b244f1d8a" (UID: "78440166-ae09-45fc-950e-2f8b244f1d8a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 11:03:55 crc kubenswrapper[4824]: I1006 11:03:55.298182 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/78440166-ae09-45fc-950e-2f8b244f1d8a-kube-api-access-lcnzn" (OuterVolumeSpecName: "kube-api-access-lcnzn") pod "78440166-ae09-45fc-950e-2f8b244f1d8a" (UID: "78440166-ae09-45fc-950e-2f8b244f1d8a"). InnerVolumeSpecName "kube-api-access-lcnzn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 11:03:55 crc kubenswrapper[4824]: I1006 11:03:55.368031 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/78440166-ae09-45fc-950e-2f8b244f1d8a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "78440166-ae09-45fc-950e-2f8b244f1d8a" (UID: "78440166-ae09-45fc-950e-2f8b244f1d8a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 11:03:55 crc kubenswrapper[4824]: I1006 11:03:55.389811 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/78440166-ae09-45fc-950e-2f8b244f1d8a-utilities\") on node \"crc\" DevicePath \"\"" Oct 06 11:03:55 crc kubenswrapper[4824]: I1006 11:03:55.389871 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lcnzn\" (UniqueName: \"kubernetes.io/projected/78440166-ae09-45fc-950e-2f8b244f1d8a-kube-api-access-lcnzn\") on node \"crc\" DevicePath \"\"" Oct 06 11:03:55 crc kubenswrapper[4824]: I1006 11:03:55.389885 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/78440166-ae09-45fc-950e-2f8b244f1d8a-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 06 11:03:55 crc kubenswrapper[4824]: I1006 11:03:55.446608 4824 generic.go:334] "Generic (PLEG): container finished" podID="78440166-ae09-45fc-950e-2f8b244f1d8a" containerID="6f34e43b77cded2748c96117247246ee85888bf79005c17ad312e49a72ec7286" exitCode=0 Oct 06 11:03:55 crc kubenswrapper[4824]: I1006 11:03:55.446673 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fxhmj" event={"ID":"78440166-ae09-45fc-950e-2f8b244f1d8a","Type":"ContainerDied","Data":"6f34e43b77cded2748c96117247246ee85888bf79005c17ad312e49a72ec7286"} Oct 06 11:03:55 crc kubenswrapper[4824]: I1006 11:03:55.446713 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fxhmj" event={"ID":"78440166-ae09-45fc-950e-2f8b244f1d8a","Type":"ContainerDied","Data":"b31c89d3b808aa4f6c3cee204ed698fdd545e2c8c427feab3bb6f6fa3eac9593"} Oct 06 11:03:55 crc kubenswrapper[4824]: I1006 11:03:55.446738 4824 scope.go:117] "RemoveContainer" containerID="6f34e43b77cded2748c96117247246ee85888bf79005c17ad312e49a72ec7286" Oct 06 11:03:55 crc kubenswrapper[4824]: I1006 11:03:55.446926 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fxhmj" Oct 06 11:03:55 crc kubenswrapper[4824]: I1006 11:03:55.492589 4824 scope.go:117] "RemoveContainer" containerID="f35a01e43fa3d34a51697539ce5b936bedc03ed400fd0859d7fcaf38d5a61a42" Oct 06 11:03:55 crc kubenswrapper[4824]: I1006 11:03:55.496662 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-fxhmj"] Oct 06 11:03:55 crc kubenswrapper[4824]: I1006 11:03:55.509883 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-fxhmj"] Oct 06 11:03:55 crc kubenswrapper[4824]: I1006 11:03:55.529500 4824 scope.go:117] "RemoveContainer" containerID="1a6e7e6f37201cd2667c01cf41ff34662c8aafbf971d12424b087a2a9165bcc4" Oct 06 11:03:55 crc kubenswrapper[4824]: I1006 11:03:55.586235 4824 scope.go:117] "RemoveContainer" containerID="6f34e43b77cded2748c96117247246ee85888bf79005c17ad312e49a72ec7286" Oct 06 11:03:55 crc kubenswrapper[4824]: E1006 11:03:55.586880 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6f34e43b77cded2748c96117247246ee85888bf79005c17ad312e49a72ec7286\": container with ID starting with 6f34e43b77cded2748c96117247246ee85888bf79005c17ad312e49a72ec7286 not found: ID does not exist" containerID="6f34e43b77cded2748c96117247246ee85888bf79005c17ad312e49a72ec7286" Oct 06 11:03:55 crc kubenswrapper[4824]: I1006 11:03:55.586919 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6f34e43b77cded2748c96117247246ee85888bf79005c17ad312e49a72ec7286"} err="failed to get container status \"6f34e43b77cded2748c96117247246ee85888bf79005c17ad312e49a72ec7286\": rpc error: code = NotFound desc = could not find container \"6f34e43b77cded2748c96117247246ee85888bf79005c17ad312e49a72ec7286\": container with ID starting with 6f34e43b77cded2748c96117247246ee85888bf79005c17ad312e49a72ec7286 not found: ID does not exist" Oct 06 11:03:55 crc kubenswrapper[4824]: I1006 11:03:55.586963 4824 scope.go:117] "RemoveContainer" containerID="f35a01e43fa3d34a51697539ce5b936bedc03ed400fd0859d7fcaf38d5a61a42" Oct 06 11:03:55 crc kubenswrapper[4824]: E1006 11:03:55.587814 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f35a01e43fa3d34a51697539ce5b936bedc03ed400fd0859d7fcaf38d5a61a42\": container with ID starting with f35a01e43fa3d34a51697539ce5b936bedc03ed400fd0859d7fcaf38d5a61a42 not found: ID does not exist" containerID="f35a01e43fa3d34a51697539ce5b936bedc03ed400fd0859d7fcaf38d5a61a42" Oct 06 11:03:55 crc kubenswrapper[4824]: I1006 11:03:55.587853 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f35a01e43fa3d34a51697539ce5b936bedc03ed400fd0859d7fcaf38d5a61a42"} err="failed to get container status \"f35a01e43fa3d34a51697539ce5b936bedc03ed400fd0859d7fcaf38d5a61a42\": rpc error: code = NotFound desc = could not find container \"f35a01e43fa3d34a51697539ce5b936bedc03ed400fd0859d7fcaf38d5a61a42\": container with ID starting with f35a01e43fa3d34a51697539ce5b936bedc03ed400fd0859d7fcaf38d5a61a42 not found: ID does not exist" Oct 06 11:03:55 crc kubenswrapper[4824]: I1006 11:03:55.587879 4824 scope.go:117] "RemoveContainer" containerID="1a6e7e6f37201cd2667c01cf41ff34662c8aafbf971d12424b087a2a9165bcc4" Oct 06 11:03:55 crc kubenswrapper[4824]: E1006 11:03:55.588393 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1a6e7e6f37201cd2667c01cf41ff34662c8aafbf971d12424b087a2a9165bcc4\": container with ID starting with 1a6e7e6f37201cd2667c01cf41ff34662c8aafbf971d12424b087a2a9165bcc4 not found: ID does not exist" containerID="1a6e7e6f37201cd2667c01cf41ff34662c8aafbf971d12424b087a2a9165bcc4" Oct 06 11:03:55 crc kubenswrapper[4824]: I1006 11:03:55.588428 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1a6e7e6f37201cd2667c01cf41ff34662c8aafbf971d12424b087a2a9165bcc4"} err="failed to get container status \"1a6e7e6f37201cd2667c01cf41ff34662c8aafbf971d12424b087a2a9165bcc4\": rpc error: code = NotFound desc = could not find container \"1a6e7e6f37201cd2667c01cf41ff34662c8aafbf971d12424b087a2a9165bcc4\": container with ID starting with 1a6e7e6f37201cd2667c01cf41ff34662c8aafbf971d12424b087a2a9165bcc4 not found: ID does not exist" Oct 06 11:03:57 crc kubenswrapper[4824]: I1006 11:03:57.291653 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="78440166-ae09-45fc-950e-2f8b244f1d8a" path="/var/lib/kubelet/pods/78440166-ae09-45fc-950e-2f8b244f1d8a/volumes" Oct 06 11:04:05 crc kubenswrapper[4824]: I1006 11:04:05.275465 4824 scope.go:117] "RemoveContainer" containerID="5a47a5d2e5b1ba0ad6ad09ca0bd7f5568851664618ce5ad15f9b3204fecd76c5" Oct 06 11:04:05 crc kubenswrapper[4824]: E1006 11:04:05.277323 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-khgzw_openshift-machine-config-operator(e1140c8c-93f1-4fce-8c06-a69561ff8a6d)\"" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" Oct 06 11:04:18 crc kubenswrapper[4824]: I1006 11:04:18.274792 4824 scope.go:117] "RemoveContainer" containerID="5a47a5d2e5b1ba0ad6ad09ca0bd7f5568851664618ce5ad15f9b3204fecd76c5" Oct 06 11:04:18 crc kubenswrapper[4824]: E1006 11:04:18.276020 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-khgzw_openshift-machine-config-operator(e1140c8c-93f1-4fce-8c06-a69561ff8a6d)\"" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" Oct 06 11:04:32 crc kubenswrapper[4824]: I1006 11:04:32.274269 4824 scope.go:117] "RemoveContainer" containerID="5a47a5d2e5b1ba0ad6ad09ca0bd7f5568851664618ce5ad15f9b3204fecd76c5" Oct 06 11:04:32 crc kubenswrapper[4824]: E1006 11:04:32.277108 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-khgzw_openshift-machine-config-operator(e1140c8c-93f1-4fce-8c06-a69561ff8a6d)\"" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" Oct 06 11:04:44 crc kubenswrapper[4824]: I1006 11:04:44.276669 4824 scope.go:117] "RemoveContainer" containerID="5a47a5d2e5b1ba0ad6ad09ca0bd7f5568851664618ce5ad15f9b3204fecd76c5" Oct 06 11:04:44 crc kubenswrapper[4824]: E1006 11:04:44.277837 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-khgzw_openshift-machine-config-operator(e1140c8c-93f1-4fce-8c06-a69561ff8a6d)\"" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" Oct 06 11:04:51 crc kubenswrapper[4824]: I1006 11:04:51.422823 4824 scope.go:117] "RemoveContainer" containerID="22fdcfb9df44916a40b6ce06b4258b25582a99185978d15eb47c665ed73431a1" Oct 06 11:04:57 crc kubenswrapper[4824]: I1006 11:04:57.274678 4824 scope.go:117] "RemoveContainer" containerID="5a47a5d2e5b1ba0ad6ad09ca0bd7f5568851664618ce5ad15f9b3204fecd76c5" Oct 06 11:04:57 crc kubenswrapper[4824]: E1006 11:04:57.276164 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-khgzw_openshift-machine-config-operator(e1140c8c-93f1-4fce-8c06-a69561ff8a6d)\"" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" Oct 06 11:04:58 crc kubenswrapper[4824]: I1006 11:04:58.249160 4824 generic.go:334] "Generic (PLEG): container finished" podID="2a18f37a-4010-4742-8811-a2fe992fe4f1" containerID="a97e4ecc6d64ffef527b64c876826b3236ed56d7ed64f580f35f4e350b2b7331" exitCode=0 Oct 06 11:04:58 crc kubenswrapper[4824]: I1006 11:04:58.249340 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-v7b5k/must-gather-kg2w2" event={"ID":"2a18f37a-4010-4742-8811-a2fe992fe4f1","Type":"ContainerDied","Data":"a97e4ecc6d64ffef527b64c876826b3236ed56d7ed64f580f35f4e350b2b7331"} Oct 06 11:04:58 crc kubenswrapper[4824]: I1006 11:04:58.250581 4824 scope.go:117] "RemoveContainer" containerID="a97e4ecc6d64ffef527b64c876826b3236ed56d7ed64f580f35f4e350b2b7331" Oct 06 11:04:59 crc kubenswrapper[4824]: I1006 11:04:59.186470 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-v7b5k_must-gather-kg2w2_2a18f37a-4010-4742-8811-a2fe992fe4f1/gather/0.log" Oct 06 11:05:08 crc kubenswrapper[4824]: I1006 11:05:08.288935 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-v7b5k/must-gather-kg2w2"] Oct 06 11:05:08 crc kubenswrapper[4824]: I1006 11:05:08.290195 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-v7b5k/must-gather-kg2w2" podUID="2a18f37a-4010-4742-8811-a2fe992fe4f1" containerName="copy" containerID="cri-o://e1bed53b164b7508a7a0ac7d00bc63f9c58c23644ccfa0d031e8357cff0078dc" gracePeriod=2 Oct 06 11:05:08 crc kubenswrapper[4824]: I1006 11:05:08.302172 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-v7b5k/must-gather-kg2w2"] Oct 06 11:05:08 crc kubenswrapper[4824]: I1006 11:05:08.821818 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-v7b5k_must-gather-kg2w2_2a18f37a-4010-4742-8811-a2fe992fe4f1/copy/0.log" Oct 06 11:05:08 crc kubenswrapper[4824]: I1006 11:05:08.822971 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-v7b5k/must-gather-kg2w2" Oct 06 11:05:08 crc kubenswrapper[4824]: I1006 11:05:08.939588 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/2a18f37a-4010-4742-8811-a2fe992fe4f1-must-gather-output\") pod \"2a18f37a-4010-4742-8811-a2fe992fe4f1\" (UID: \"2a18f37a-4010-4742-8811-a2fe992fe4f1\") " Oct 06 11:05:08 crc kubenswrapper[4824]: I1006 11:05:08.939918 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4gf44\" (UniqueName: \"kubernetes.io/projected/2a18f37a-4010-4742-8811-a2fe992fe4f1-kube-api-access-4gf44\") pod \"2a18f37a-4010-4742-8811-a2fe992fe4f1\" (UID: \"2a18f37a-4010-4742-8811-a2fe992fe4f1\") " Oct 06 11:05:08 crc kubenswrapper[4824]: I1006 11:05:08.950291 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2a18f37a-4010-4742-8811-a2fe992fe4f1-kube-api-access-4gf44" (OuterVolumeSpecName: "kube-api-access-4gf44") pod "2a18f37a-4010-4742-8811-a2fe992fe4f1" (UID: "2a18f37a-4010-4742-8811-a2fe992fe4f1"). InnerVolumeSpecName "kube-api-access-4gf44". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 11:05:09 crc kubenswrapper[4824]: I1006 11:05:09.042255 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4gf44\" (UniqueName: \"kubernetes.io/projected/2a18f37a-4010-4742-8811-a2fe992fe4f1-kube-api-access-4gf44\") on node \"crc\" DevicePath \"\"" Oct 06 11:05:09 crc kubenswrapper[4824]: I1006 11:05:09.175090 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2a18f37a-4010-4742-8811-a2fe992fe4f1-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "2a18f37a-4010-4742-8811-a2fe992fe4f1" (UID: "2a18f37a-4010-4742-8811-a2fe992fe4f1"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 11:05:09 crc kubenswrapper[4824]: I1006 11:05:09.247958 4824 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/2a18f37a-4010-4742-8811-a2fe992fe4f1-must-gather-output\") on node \"crc\" DevicePath \"\"" Oct 06 11:05:09 crc kubenswrapper[4824]: I1006 11:05:09.283812 4824 scope.go:117] "RemoveContainer" containerID="5a47a5d2e5b1ba0ad6ad09ca0bd7f5568851664618ce5ad15f9b3204fecd76c5" Oct 06 11:05:09 crc kubenswrapper[4824]: E1006 11:05:09.284145 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-khgzw_openshift-machine-config-operator(e1140c8c-93f1-4fce-8c06-a69561ff8a6d)\"" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" Oct 06 11:05:09 crc kubenswrapper[4824]: I1006 11:05:09.291589 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2a18f37a-4010-4742-8811-a2fe992fe4f1" path="/var/lib/kubelet/pods/2a18f37a-4010-4742-8811-a2fe992fe4f1/volumes" Oct 06 11:05:09 crc kubenswrapper[4824]: I1006 11:05:09.372046 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-v7b5k_must-gather-kg2w2_2a18f37a-4010-4742-8811-a2fe992fe4f1/copy/0.log" Oct 06 11:05:09 crc kubenswrapper[4824]: I1006 11:05:09.372572 4824 generic.go:334] "Generic (PLEG): container finished" podID="2a18f37a-4010-4742-8811-a2fe992fe4f1" containerID="e1bed53b164b7508a7a0ac7d00bc63f9c58c23644ccfa0d031e8357cff0078dc" exitCode=143 Oct 06 11:05:09 crc kubenswrapper[4824]: I1006 11:05:09.372623 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-v7b5k/must-gather-kg2w2" Oct 06 11:05:09 crc kubenswrapper[4824]: I1006 11:05:09.372685 4824 scope.go:117] "RemoveContainer" containerID="e1bed53b164b7508a7a0ac7d00bc63f9c58c23644ccfa0d031e8357cff0078dc" Oct 06 11:05:09 crc kubenswrapper[4824]: I1006 11:05:09.400398 4824 scope.go:117] "RemoveContainer" containerID="a97e4ecc6d64ffef527b64c876826b3236ed56d7ed64f580f35f4e350b2b7331" Oct 06 11:05:09 crc kubenswrapper[4824]: I1006 11:05:09.458182 4824 scope.go:117] "RemoveContainer" containerID="e1bed53b164b7508a7a0ac7d00bc63f9c58c23644ccfa0d031e8357cff0078dc" Oct 06 11:05:09 crc kubenswrapper[4824]: E1006 11:05:09.458831 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e1bed53b164b7508a7a0ac7d00bc63f9c58c23644ccfa0d031e8357cff0078dc\": container with ID starting with e1bed53b164b7508a7a0ac7d00bc63f9c58c23644ccfa0d031e8357cff0078dc not found: ID does not exist" containerID="e1bed53b164b7508a7a0ac7d00bc63f9c58c23644ccfa0d031e8357cff0078dc" Oct 06 11:05:09 crc kubenswrapper[4824]: I1006 11:05:09.458872 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e1bed53b164b7508a7a0ac7d00bc63f9c58c23644ccfa0d031e8357cff0078dc"} err="failed to get container status \"e1bed53b164b7508a7a0ac7d00bc63f9c58c23644ccfa0d031e8357cff0078dc\": rpc error: code = NotFound desc = could not find container \"e1bed53b164b7508a7a0ac7d00bc63f9c58c23644ccfa0d031e8357cff0078dc\": container with ID starting with e1bed53b164b7508a7a0ac7d00bc63f9c58c23644ccfa0d031e8357cff0078dc not found: ID does not exist" Oct 06 11:05:09 crc kubenswrapper[4824]: I1006 11:05:09.458894 4824 scope.go:117] "RemoveContainer" containerID="a97e4ecc6d64ffef527b64c876826b3236ed56d7ed64f580f35f4e350b2b7331" Oct 06 11:05:09 crc kubenswrapper[4824]: E1006 11:05:09.459618 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a97e4ecc6d64ffef527b64c876826b3236ed56d7ed64f580f35f4e350b2b7331\": container with ID starting with a97e4ecc6d64ffef527b64c876826b3236ed56d7ed64f580f35f4e350b2b7331 not found: ID does not exist" containerID="a97e4ecc6d64ffef527b64c876826b3236ed56d7ed64f580f35f4e350b2b7331" Oct 06 11:05:09 crc kubenswrapper[4824]: I1006 11:05:09.459684 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a97e4ecc6d64ffef527b64c876826b3236ed56d7ed64f580f35f4e350b2b7331"} err="failed to get container status \"a97e4ecc6d64ffef527b64c876826b3236ed56d7ed64f580f35f4e350b2b7331\": rpc error: code = NotFound desc = could not find container \"a97e4ecc6d64ffef527b64c876826b3236ed56d7ed64f580f35f4e350b2b7331\": container with ID starting with a97e4ecc6d64ffef527b64c876826b3236ed56d7ed64f580f35f4e350b2b7331 not found: ID does not exist" Oct 06 11:05:23 crc kubenswrapper[4824]: I1006 11:05:23.274426 4824 scope.go:117] "RemoveContainer" containerID="5a47a5d2e5b1ba0ad6ad09ca0bd7f5568851664618ce5ad15f9b3204fecd76c5" Oct 06 11:05:23 crc kubenswrapper[4824]: E1006 11:05:23.275782 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-khgzw_openshift-machine-config-operator(e1140c8c-93f1-4fce-8c06-a69561ff8a6d)\"" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" Oct 06 11:05:38 crc kubenswrapper[4824]: I1006 11:05:38.275768 4824 scope.go:117] "RemoveContainer" containerID="5a47a5d2e5b1ba0ad6ad09ca0bd7f5568851664618ce5ad15f9b3204fecd76c5" Oct 06 11:05:38 crc kubenswrapper[4824]: E1006 11:05:38.277085 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-khgzw_openshift-machine-config-operator(e1140c8c-93f1-4fce-8c06-a69561ff8a6d)\"" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" Oct 06 11:05:51 crc kubenswrapper[4824]: I1006 11:05:51.280941 4824 scope.go:117] "RemoveContainer" containerID="5a47a5d2e5b1ba0ad6ad09ca0bd7f5568851664618ce5ad15f9b3204fecd76c5" Oct 06 11:05:51 crc kubenswrapper[4824]: E1006 11:05:51.282221 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-khgzw_openshift-machine-config-operator(e1140c8c-93f1-4fce-8c06-a69561ff8a6d)\"" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" Oct 06 11:05:55 crc kubenswrapper[4824]: I1006 11:05:55.079543 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-x5ksk/must-gather-zn6wq"] Oct 06 11:05:55 crc kubenswrapper[4824]: E1006 11:05:55.080906 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a18f37a-4010-4742-8811-a2fe992fe4f1" containerName="copy" Oct 06 11:05:55 crc kubenswrapper[4824]: I1006 11:05:55.080925 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a18f37a-4010-4742-8811-a2fe992fe4f1" containerName="copy" Oct 06 11:05:55 crc kubenswrapper[4824]: E1006 11:05:55.080945 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78440166-ae09-45fc-950e-2f8b244f1d8a" containerName="extract-content" Oct 06 11:05:55 crc kubenswrapper[4824]: I1006 11:05:55.080954 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="78440166-ae09-45fc-950e-2f8b244f1d8a" containerName="extract-content" Oct 06 11:05:55 crc kubenswrapper[4824]: E1006 11:05:55.081014 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a18f37a-4010-4742-8811-a2fe992fe4f1" containerName="gather" Oct 06 11:05:55 crc kubenswrapper[4824]: I1006 11:05:55.081024 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a18f37a-4010-4742-8811-a2fe992fe4f1" containerName="gather" Oct 06 11:05:55 crc kubenswrapper[4824]: E1006 11:05:55.081033 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78440166-ae09-45fc-950e-2f8b244f1d8a" containerName="extract-utilities" Oct 06 11:05:55 crc kubenswrapper[4824]: I1006 11:05:55.081042 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="78440166-ae09-45fc-950e-2f8b244f1d8a" containerName="extract-utilities" Oct 06 11:05:55 crc kubenswrapper[4824]: E1006 11:05:55.081083 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78440166-ae09-45fc-950e-2f8b244f1d8a" containerName="registry-server" Oct 06 11:05:55 crc kubenswrapper[4824]: I1006 11:05:55.081091 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="78440166-ae09-45fc-950e-2f8b244f1d8a" containerName="registry-server" Oct 06 11:05:55 crc kubenswrapper[4824]: I1006 11:05:55.081364 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a18f37a-4010-4742-8811-a2fe992fe4f1" containerName="copy" Oct 06 11:05:55 crc kubenswrapper[4824]: I1006 11:05:55.081382 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a18f37a-4010-4742-8811-a2fe992fe4f1" containerName="gather" Oct 06 11:05:55 crc kubenswrapper[4824]: I1006 11:05:55.081397 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="78440166-ae09-45fc-950e-2f8b244f1d8a" containerName="registry-server" Oct 06 11:05:55 crc kubenswrapper[4824]: I1006 11:05:55.082890 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-x5ksk/must-gather-zn6wq" Oct 06 11:05:55 crc kubenswrapper[4824]: I1006 11:05:55.089696 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-x5ksk"/"kube-root-ca.crt" Oct 06 11:05:55 crc kubenswrapper[4824]: I1006 11:05:55.089859 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-x5ksk"/"openshift-service-ca.crt" Oct 06 11:05:55 crc kubenswrapper[4824]: I1006 11:05:55.138074 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-x5ksk/must-gather-zn6wq"] Oct 06 11:05:55 crc kubenswrapper[4824]: I1006 11:05:55.210870 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/215303b4-ee21-44b7-a2a3-cd3ff4c6265e-must-gather-output\") pod \"must-gather-zn6wq\" (UID: \"215303b4-ee21-44b7-a2a3-cd3ff4c6265e\") " pod="openshift-must-gather-x5ksk/must-gather-zn6wq" Oct 06 11:05:55 crc kubenswrapper[4824]: I1006 11:05:55.210963 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-64g8x\" (UniqueName: \"kubernetes.io/projected/215303b4-ee21-44b7-a2a3-cd3ff4c6265e-kube-api-access-64g8x\") pod \"must-gather-zn6wq\" (UID: \"215303b4-ee21-44b7-a2a3-cd3ff4c6265e\") " pod="openshift-must-gather-x5ksk/must-gather-zn6wq" Oct 06 11:05:55 crc kubenswrapper[4824]: I1006 11:05:55.313084 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/215303b4-ee21-44b7-a2a3-cd3ff4c6265e-must-gather-output\") pod \"must-gather-zn6wq\" (UID: \"215303b4-ee21-44b7-a2a3-cd3ff4c6265e\") " pod="openshift-must-gather-x5ksk/must-gather-zn6wq" Oct 06 11:05:55 crc kubenswrapper[4824]: I1006 11:05:55.313157 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-64g8x\" (UniqueName: \"kubernetes.io/projected/215303b4-ee21-44b7-a2a3-cd3ff4c6265e-kube-api-access-64g8x\") pod \"must-gather-zn6wq\" (UID: \"215303b4-ee21-44b7-a2a3-cd3ff4c6265e\") " pod="openshift-must-gather-x5ksk/must-gather-zn6wq" Oct 06 11:05:55 crc kubenswrapper[4824]: I1006 11:05:55.313596 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/215303b4-ee21-44b7-a2a3-cd3ff4c6265e-must-gather-output\") pod \"must-gather-zn6wq\" (UID: \"215303b4-ee21-44b7-a2a3-cd3ff4c6265e\") " pod="openshift-must-gather-x5ksk/must-gather-zn6wq" Oct 06 11:05:55 crc kubenswrapper[4824]: I1006 11:05:55.338189 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-64g8x\" (UniqueName: \"kubernetes.io/projected/215303b4-ee21-44b7-a2a3-cd3ff4c6265e-kube-api-access-64g8x\") pod \"must-gather-zn6wq\" (UID: \"215303b4-ee21-44b7-a2a3-cd3ff4c6265e\") " pod="openshift-must-gather-x5ksk/must-gather-zn6wq" Oct 06 11:05:55 crc kubenswrapper[4824]: I1006 11:05:55.432803 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-x5ksk/must-gather-zn6wq" Oct 06 11:05:55 crc kubenswrapper[4824]: I1006 11:05:55.980430 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-x5ksk/must-gather-zn6wq"] Oct 06 11:05:56 crc kubenswrapper[4824]: I1006 11:05:56.996178 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-x5ksk/must-gather-zn6wq" event={"ID":"215303b4-ee21-44b7-a2a3-cd3ff4c6265e","Type":"ContainerStarted","Data":"018ec10e14724ca059b1c230ec932142c368814b8f1315ec80e308d2f7d9a3bf"} Oct 06 11:05:56 crc kubenswrapper[4824]: I1006 11:05:56.996695 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-x5ksk/must-gather-zn6wq" event={"ID":"215303b4-ee21-44b7-a2a3-cd3ff4c6265e","Type":"ContainerStarted","Data":"551a46d712b0750162e9fdff376ae3c1fdecf89e52dd8b5b60fe053ec4f4b731"} Oct 06 11:05:56 crc kubenswrapper[4824]: I1006 11:05:56.996719 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-x5ksk/must-gather-zn6wq" event={"ID":"215303b4-ee21-44b7-a2a3-cd3ff4c6265e","Type":"ContainerStarted","Data":"a40f5e3756f5f0923a4fb501d5ff6b5254ad2bb885ae2f3322e8f0fc90363080"} Oct 06 11:05:57 crc kubenswrapper[4824]: I1006 11:05:57.028746 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-x5ksk/must-gather-zn6wq" podStartSLOduration=2.028716045 podStartE2EDuration="2.028716045s" podCreationTimestamp="2025-10-06 11:05:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 11:05:57.017502766 +0000 UTC m=+4066.381925627" watchObservedRunningTime="2025-10-06 11:05:57.028716045 +0000 UTC m=+4066.393138946" Oct 06 11:06:01 crc kubenswrapper[4824]: I1006 11:06:01.076160 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-x5ksk/crc-debug-bc9hs"] Oct 06 11:06:01 crc kubenswrapper[4824]: I1006 11:06:01.081515 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-x5ksk/crc-debug-bc9hs" Oct 06 11:06:01 crc kubenswrapper[4824]: I1006 11:06:01.085001 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-x5ksk"/"default-dockercfg-5c5x6" Oct 06 11:06:01 crc kubenswrapper[4824]: I1006 11:06:01.155705 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/acffb1f0-1d70-4905-9648-e2782b876ef8-host\") pod \"crc-debug-bc9hs\" (UID: \"acffb1f0-1d70-4905-9648-e2782b876ef8\") " pod="openshift-must-gather-x5ksk/crc-debug-bc9hs" Oct 06 11:06:01 crc kubenswrapper[4824]: I1006 11:06:01.156338 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zdkqs\" (UniqueName: \"kubernetes.io/projected/acffb1f0-1d70-4905-9648-e2782b876ef8-kube-api-access-zdkqs\") pod \"crc-debug-bc9hs\" (UID: \"acffb1f0-1d70-4905-9648-e2782b876ef8\") " pod="openshift-must-gather-x5ksk/crc-debug-bc9hs" Oct 06 11:06:01 crc kubenswrapper[4824]: I1006 11:06:01.258852 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/acffb1f0-1d70-4905-9648-e2782b876ef8-host\") pod \"crc-debug-bc9hs\" (UID: \"acffb1f0-1d70-4905-9648-e2782b876ef8\") " pod="openshift-must-gather-x5ksk/crc-debug-bc9hs" Oct 06 11:06:01 crc kubenswrapper[4824]: I1006 11:06:01.259218 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zdkqs\" (UniqueName: \"kubernetes.io/projected/acffb1f0-1d70-4905-9648-e2782b876ef8-kube-api-access-zdkqs\") pod \"crc-debug-bc9hs\" (UID: \"acffb1f0-1d70-4905-9648-e2782b876ef8\") " pod="openshift-must-gather-x5ksk/crc-debug-bc9hs" Oct 06 11:06:01 crc kubenswrapper[4824]: I1006 11:06:01.259023 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/acffb1f0-1d70-4905-9648-e2782b876ef8-host\") pod \"crc-debug-bc9hs\" (UID: \"acffb1f0-1d70-4905-9648-e2782b876ef8\") " pod="openshift-must-gather-x5ksk/crc-debug-bc9hs" Oct 06 11:06:01 crc kubenswrapper[4824]: I1006 11:06:01.283822 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zdkqs\" (UniqueName: \"kubernetes.io/projected/acffb1f0-1d70-4905-9648-e2782b876ef8-kube-api-access-zdkqs\") pod \"crc-debug-bc9hs\" (UID: \"acffb1f0-1d70-4905-9648-e2782b876ef8\") " pod="openshift-must-gather-x5ksk/crc-debug-bc9hs" Oct 06 11:06:01 crc kubenswrapper[4824]: I1006 11:06:01.403855 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-x5ksk/crc-debug-bc9hs" Oct 06 11:06:01 crc kubenswrapper[4824]: W1006 11:06:01.456317 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podacffb1f0_1d70_4905_9648_e2782b876ef8.slice/crio-53d5081c0e1fc5829eb5a2b2121edf9d4b7ce50e323a84399eeb30155ed9ad45 WatchSource:0}: Error finding container 53d5081c0e1fc5829eb5a2b2121edf9d4b7ce50e323a84399eeb30155ed9ad45: Status 404 returned error can't find the container with id 53d5081c0e1fc5829eb5a2b2121edf9d4b7ce50e323a84399eeb30155ed9ad45 Oct 06 11:06:02 crc kubenswrapper[4824]: I1006 11:06:02.051310 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-x5ksk/crc-debug-bc9hs" event={"ID":"acffb1f0-1d70-4905-9648-e2782b876ef8","Type":"ContainerStarted","Data":"ea6e43bb9cdcd06002cd0bc6be6805769696143892dae1049fc95d7f7f527d3e"} Oct 06 11:06:02 crc kubenswrapper[4824]: I1006 11:06:02.052223 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-x5ksk/crc-debug-bc9hs" event={"ID":"acffb1f0-1d70-4905-9648-e2782b876ef8","Type":"ContainerStarted","Data":"53d5081c0e1fc5829eb5a2b2121edf9d4b7ce50e323a84399eeb30155ed9ad45"} Oct 06 11:06:02 crc kubenswrapper[4824]: I1006 11:06:02.075441 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-x5ksk/crc-debug-bc9hs" podStartSLOduration=1.075416434 podStartE2EDuration="1.075416434s" podCreationTimestamp="2025-10-06 11:06:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 11:06:02.068604955 +0000 UTC m=+4071.433027806" watchObservedRunningTime="2025-10-06 11:06:02.075416434 +0000 UTC m=+4071.439839295" Oct 06 11:06:06 crc kubenswrapper[4824]: I1006 11:06:06.275320 4824 scope.go:117] "RemoveContainer" containerID="5a47a5d2e5b1ba0ad6ad09ca0bd7f5568851664618ce5ad15f9b3204fecd76c5" Oct 06 11:06:06 crc kubenswrapper[4824]: E1006 11:06:06.276352 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-khgzw_openshift-machine-config-operator(e1140c8c-93f1-4fce-8c06-a69561ff8a6d)\"" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" Oct 06 11:06:18 crc kubenswrapper[4824]: I1006 11:06:18.274900 4824 scope.go:117] "RemoveContainer" containerID="5a47a5d2e5b1ba0ad6ad09ca0bd7f5568851664618ce5ad15f9b3204fecd76c5" Oct 06 11:06:18 crc kubenswrapper[4824]: E1006 11:06:18.276048 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-khgzw_openshift-machine-config-operator(e1140c8c-93f1-4fce-8c06-a69561ff8a6d)\"" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" Oct 06 11:06:32 crc kubenswrapper[4824]: I1006 11:06:32.274412 4824 scope.go:117] "RemoveContainer" containerID="5a47a5d2e5b1ba0ad6ad09ca0bd7f5568851664618ce5ad15f9b3204fecd76c5" Oct 06 11:06:32 crc kubenswrapper[4824]: E1006 11:06:32.275258 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-khgzw_openshift-machine-config-operator(e1140c8c-93f1-4fce-8c06-a69561ff8a6d)\"" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" Oct 06 11:06:43 crc kubenswrapper[4824]: I1006 11:06:43.275247 4824 scope.go:117] "RemoveContainer" containerID="5a47a5d2e5b1ba0ad6ad09ca0bd7f5568851664618ce5ad15f9b3204fecd76c5" Oct 06 11:06:43 crc kubenswrapper[4824]: E1006 11:06:43.276448 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-khgzw_openshift-machine-config-operator(e1140c8c-93f1-4fce-8c06-a69561ff8a6d)\"" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" Oct 06 11:06:57 crc kubenswrapper[4824]: I1006 11:06:57.274388 4824 scope.go:117] "RemoveContainer" containerID="5a47a5d2e5b1ba0ad6ad09ca0bd7f5568851664618ce5ad15f9b3204fecd76c5" Oct 06 11:06:57 crc kubenswrapper[4824]: E1006 11:06:57.275457 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-khgzw_openshift-machine-config-operator(e1140c8c-93f1-4fce-8c06-a69561ff8a6d)\"" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" Oct 06 11:07:12 crc kubenswrapper[4824]: I1006 11:07:12.275000 4824 scope.go:117] "RemoveContainer" containerID="5a47a5d2e5b1ba0ad6ad09ca0bd7f5568851664618ce5ad15f9b3204fecd76c5" Oct 06 11:07:12 crc kubenswrapper[4824]: E1006 11:07:12.276069 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-khgzw_openshift-machine-config-operator(e1140c8c-93f1-4fce-8c06-a69561ff8a6d)\"" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" Oct 06 11:07:19 crc kubenswrapper[4824]: I1006 11:07:19.408832 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-78c986c6fb-bzcht_b8a8eafe-1a7a-4eca-b61c-2d744512a54b/barbican-api/0.log" Oct 06 11:07:19 crc kubenswrapper[4824]: I1006 11:07:19.462517 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-78c986c6fb-bzcht_b8a8eafe-1a7a-4eca-b61c-2d744512a54b/barbican-api-log/0.log" Oct 06 11:07:19 crc kubenswrapper[4824]: I1006 11:07:19.679753 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-67979769c4-7fg5z_5e844107-f8f2-48e8-9731-84eff0f8c752/barbican-keystone-listener/0.log" Oct 06 11:07:19 crc kubenswrapper[4824]: I1006 11:07:19.708687 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-67979769c4-7fg5z_5e844107-f8f2-48e8-9731-84eff0f8c752/barbican-keystone-listener-log/0.log" Oct 06 11:07:19 crc kubenswrapper[4824]: I1006 11:07:19.931901 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-659dd56d6c-qpxtd_fb46f73f-9d47-4189-990a-051fed587851/barbican-worker/0.log" Oct 06 11:07:19 crc kubenswrapper[4824]: I1006 11:07:19.937310 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-659dd56d6c-qpxtd_fb46f73f-9d47-4189-990a-051fed587851/barbican-worker-log/0.log" Oct 06 11:07:20 crc kubenswrapper[4824]: I1006 11:07:20.147445 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-mlckl_c995dce1-6fa5-456b-b984-b397dcc9f9dc/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Oct 06 11:07:20 crc kubenswrapper[4824]: I1006 11:07:20.353567 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_369bbc94-27b7-4016-9a43-1e20b28f4323/ceilometer-central-agent/0.log" Oct 06 11:07:20 crc kubenswrapper[4824]: I1006 11:07:20.386875 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_369bbc94-27b7-4016-9a43-1e20b28f4323/proxy-httpd/0.log" Oct 06 11:07:20 crc kubenswrapper[4824]: I1006 11:07:20.402363 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_369bbc94-27b7-4016-9a43-1e20b28f4323/ceilometer-notification-agent/0.log" Oct 06 11:07:20 crc kubenswrapper[4824]: I1006 11:07:20.534705 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_369bbc94-27b7-4016-9a43-1e20b28f4323/sg-core/0.log" Oct 06 11:07:20 crc kubenswrapper[4824]: I1006 11:07:20.650154 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_13a74e55-383d-4105-9c9c-aafb63b75c39/cinder-api/0.log" Oct 06 11:07:20 crc kubenswrapper[4824]: I1006 11:07:20.755264 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_13a74e55-383d-4105-9c9c-aafb63b75c39/cinder-api-log/0.log" Oct 06 11:07:21 crc kubenswrapper[4824]: I1006 11:07:21.455253 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_18194590-7963-41e8-9253-dcf486d2cea4/probe/0.log" Oct 06 11:07:21 crc kubenswrapper[4824]: I1006 11:07:21.516992 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_18194590-7963-41e8-9253-dcf486d2cea4/cinder-scheduler/0.log" Oct 06 11:07:21 crc kubenswrapper[4824]: I1006 11:07:21.753818 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-xtlm8_1961267a-e5c0-469e-9f0e-9d4edbc8e64c/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Oct 06 11:07:21 crc kubenswrapper[4824]: I1006 11:07:21.912652 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-nm76g_87320828-94da-481c-b903-e7d478e3df65/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 06 11:07:22 crc kubenswrapper[4824]: I1006 11:07:22.032909 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-shvtw_3a56b28d-58ae-4cd4-a06a-b942a9365de7/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 06 11:07:22 crc kubenswrapper[4824]: I1006 11:07:22.252839 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-cb6ffcf87-xt4l8_7648fd80-2fcf-4ddd-bb32-1dd33c86a0fb/init/0.log" Oct 06 11:07:22 crc kubenswrapper[4824]: I1006 11:07:22.398526 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-cb6ffcf87-xt4l8_7648fd80-2fcf-4ddd-bb32-1dd33c86a0fb/init/0.log" Oct 06 11:07:22 crc kubenswrapper[4824]: I1006 11:07:22.473324 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-cb6ffcf87-xt4l8_7648fd80-2fcf-4ddd-bb32-1dd33c86a0fb/dnsmasq-dns/0.log" Oct 06 11:07:22 crc kubenswrapper[4824]: I1006 11:07:22.507050 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-kcvhh_8ea3d60c-80c7-4163-9f81-c0ec20e758c9/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Oct 06 11:07:22 crc kubenswrapper[4824]: I1006 11:07:22.695439 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_4b875c3d-d6b9-47c6-ae58-4ba7b78fb338/glance-httpd/0.log" Oct 06 11:07:22 crc kubenswrapper[4824]: I1006 11:07:22.716837 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_4b875c3d-d6b9-47c6-ae58-4ba7b78fb338/glance-log/0.log" Oct 06 11:07:22 crc kubenswrapper[4824]: I1006 11:07:22.889252 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_7d31f37b-142f-4f46-962a-e806e3d1269c/glance-httpd/0.log" Oct 06 11:07:22 crc kubenswrapper[4824]: I1006 11:07:22.917044 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_7d31f37b-142f-4f46-962a-e806e3d1269c/glance-log/0.log" Oct 06 11:07:23 crc kubenswrapper[4824]: I1006 11:07:23.274762 4824 scope.go:117] "RemoveContainer" containerID="5a47a5d2e5b1ba0ad6ad09ca0bd7f5568851664618ce5ad15f9b3204fecd76c5" Oct 06 11:07:23 crc kubenswrapper[4824]: I1006 11:07:23.767226 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-jqgjd_54e35b51-aeb1-4e2f-9be0-11230cb89fbe/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Oct 06 11:07:23 crc kubenswrapper[4824]: I1006 11:07:23.910435 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-68fdc5dd48-gfq4d_194bef17-98ac-413f-a91e-0abc9835f2f6/horizon/0.log" Oct 06 11:07:23 crc kubenswrapper[4824]: I1006 11:07:23.990406 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" event={"ID":"e1140c8c-93f1-4fce-8c06-a69561ff8a6d","Type":"ContainerStarted","Data":"4fa20ecee3b1c9e1a0579597134d4ac9934abd73fbbea1c39577d842a5d800c4"} Oct 06 11:07:24 crc kubenswrapper[4824]: I1006 11:07:24.034218 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-k82nk_7db64f14-df4b-4519-81c5-f4e03e053925/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 06 11:07:24 crc kubenswrapper[4824]: I1006 11:07:24.268025 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-68fdc5dd48-gfq4d_194bef17-98ac-413f-a91e-0abc9835f2f6/horizon-log/0.log" Oct 06 11:07:24 crc kubenswrapper[4824]: I1006 11:07:24.345523 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29329141-z4m2j_46ee6fe4-f9a8-4f4a-a25c-0807817131d4/keystone-cron/0.log" Oct 06 11:07:24 crc kubenswrapper[4824]: I1006 11:07:24.553066 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_58776586-f2ad-4b0e-bf7e-303324ffe7f2/kube-state-metrics/0.log" Oct 06 11:07:24 crc kubenswrapper[4824]: I1006 11:07:24.606608 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-fbf8cb574-lrj2t_1015a252-b07f-4398-9423-41daf44c33b6/keystone-api/0.log" Oct 06 11:07:24 crc kubenswrapper[4824]: I1006 11:07:24.764511 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-8s745_4555aa06-f941-492e-9a1c-89a7c1a5c50f/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Oct 06 11:07:25 crc kubenswrapper[4824]: I1006 11:07:25.393990 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-c7c794649-hcxzr_2919256c-ce36-4092-898d-024f10692bef/neutron-api/0.log" Oct 06 11:07:25 crc kubenswrapper[4824]: I1006 11:07:25.477796 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-c7c794649-hcxzr_2919256c-ce36-4092-898d-024f10692bef/neutron-httpd/0.log" Oct 06 11:07:25 crc kubenswrapper[4824]: I1006 11:07:25.658767 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-kcljj_7e9c21f0-458e-493c-8dee-3ca1b591c60b/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Oct 06 11:07:26 crc kubenswrapper[4824]: I1006 11:07:26.432669 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_20fa260b-d7c3-4f58-82e3-070c45d494f4/nova-api-log/0.log" Oct 06 11:07:26 crc kubenswrapper[4824]: I1006 11:07:26.773304 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_20fa260b-d7c3-4f58-82e3-070c45d494f4/nova-api-api/0.log" Oct 06 11:07:26 crc kubenswrapper[4824]: I1006 11:07:26.774750 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_2e76a912-576d-41a3-86b0-f7fc6a6814a2/nova-cell0-conductor-conductor/0.log" Oct 06 11:07:27 crc kubenswrapper[4824]: I1006 11:07:27.309760 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_e57703dd-1748-4d3d-8f2f-e4edc0901a76/nova-cell1-conductor-conductor/0.log" Oct 06 11:07:27 crc kubenswrapper[4824]: I1006 11:07:27.338431 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_db4434eb-50c9-4004-8ff3-752cffaa01b4/nova-cell1-novncproxy-novncproxy/0.log" Oct 06 11:07:27 crc kubenswrapper[4824]: I1006 11:07:27.673055 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-59jnp_910d1ffe-8746-4412-bba0-74bf7ff0e1ea/nova-edpm-deployment-openstack-edpm-ipam/0.log" Oct 06 11:07:27 crc kubenswrapper[4824]: I1006 11:07:27.680038 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_754abe5c-3e1b-4a4a-b5a7-619326c06846/nova-metadata-log/0.log" Oct 06 11:07:28 crc kubenswrapper[4824]: I1006 11:07:28.290610 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_1166f1ef-43be-42e7-9d5f-84252fccfb69/nova-scheduler-scheduler/0.log" Oct 06 11:07:28 crc kubenswrapper[4824]: I1006 11:07:28.509911 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_1510d553-3505-4bd2-9666-c95cecc43d01/mysql-bootstrap/0.log" Oct 06 11:07:28 crc kubenswrapper[4824]: I1006 11:07:28.747612 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_1510d553-3505-4bd2-9666-c95cecc43d01/galera/0.log" Oct 06 11:07:28 crc kubenswrapper[4824]: I1006 11:07:28.786255 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_1510d553-3505-4bd2-9666-c95cecc43d01/mysql-bootstrap/0.log" Oct 06 11:07:29 crc kubenswrapper[4824]: I1006 11:07:29.126303 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_58f6ad3c-7a20-4be6-8ffb-cadff1c6adef/mysql-bootstrap/0.log" Oct 06 11:07:29 crc kubenswrapper[4824]: I1006 11:07:29.310927 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_58f6ad3c-7a20-4be6-8ffb-cadff1c6adef/mysql-bootstrap/0.log" Oct 06 11:07:29 crc kubenswrapper[4824]: I1006 11:07:29.409496 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_58f6ad3c-7a20-4be6-8ffb-cadff1c6adef/galera/0.log" Oct 06 11:07:29 crc kubenswrapper[4824]: I1006 11:07:29.564812 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_754abe5c-3e1b-4a4a-b5a7-619326c06846/nova-metadata-metadata/0.log" Oct 06 11:07:29 crc kubenswrapper[4824]: I1006 11:07:29.670873 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_6cdeb7d5-11b7-4e30-92fd-5c88b658ea92/openstackclient/0.log" Oct 06 11:07:29 crc kubenswrapper[4824]: I1006 11:07:29.856307 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-48qsh_03d76a8a-df87-4a48-8fb9-e6a502b37ae8/ovn-controller/0.log" Oct 06 11:07:30 crc kubenswrapper[4824]: I1006 11:07:30.043943 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-77kzd_419276d8-4ecf-4133-a98e-6e9373438668/openstack-network-exporter/0.log" Oct 06 11:07:30 crc kubenswrapper[4824]: I1006 11:07:30.199080 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-v22g5_59148e21-e79a-466f-bed5-24671942c24f/ovsdb-server-init/0.log" Oct 06 11:07:30 crc kubenswrapper[4824]: I1006 11:07:30.472364 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-v22g5_59148e21-e79a-466f-bed5-24671942c24f/ovsdb-server-init/0.log" Oct 06 11:07:30 crc kubenswrapper[4824]: I1006 11:07:30.485663 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-v22g5_59148e21-e79a-466f-bed5-24671942c24f/ovsdb-server/0.log" Oct 06 11:07:30 crc kubenswrapper[4824]: I1006 11:07:30.523762 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-v22g5_59148e21-e79a-466f-bed5-24671942c24f/ovs-vswitchd/0.log" Oct 06 11:07:30 crc kubenswrapper[4824]: I1006 11:07:30.740187 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-fn42c_97deb005-9b94-45a5-8d7f-84147ae5807c/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Oct 06 11:07:30 crc kubenswrapper[4824]: I1006 11:07:30.997679 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_43002b2d-db5d-488e-9791-56388700cf9f/openstack-network-exporter/0.log" Oct 06 11:07:31 crc kubenswrapper[4824]: I1006 11:07:31.035567 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_43002b2d-db5d-488e-9791-56388700cf9f/ovn-northd/0.log" Oct 06 11:07:31 crc kubenswrapper[4824]: I1006 11:07:31.308358 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_a39bac9c-ff87-4d4a-ad96-fa5618ca4e40/openstack-network-exporter/0.log" Oct 06 11:07:31 crc kubenswrapper[4824]: I1006 11:07:31.316123 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_a39bac9c-ff87-4d4a-ad96-fa5618ca4e40/ovsdbserver-nb/0.log" Oct 06 11:07:31 crc kubenswrapper[4824]: I1006 11:07:31.555785 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_b6303c51-6db3-4ab6-aeab-edf39e88fdec/ovsdbserver-sb/0.log" Oct 06 11:07:31 crc kubenswrapper[4824]: I1006 11:07:31.557505 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_b6303c51-6db3-4ab6-aeab-edf39e88fdec/openstack-network-exporter/0.log" Oct 06 11:07:31 crc kubenswrapper[4824]: I1006 11:07:31.868672 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-5f446cffd4-v2pxw_1d3b99b4-b383-4c33-ae51-b61c611a5b3d/placement-api/0.log" Oct 06 11:07:31 crc kubenswrapper[4824]: I1006 11:07:31.935007 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-5f446cffd4-v2pxw_1d3b99b4-b383-4c33-ae51-b61c611a5b3d/placement-log/0.log" Oct 06 11:07:32 crc kubenswrapper[4824]: I1006 11:07:32.117598 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_0295abaf-72dd-46d7-9125-543921a6ef61/setup-container/0.log" Oct 06 11:07:32 crc kubenswrapper[4824]: I1006 11:07:32.411928 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_0295abaf-72dd-46d7-9125-543921a6ef61/setup-container/0.log" Oct 06 11:07:32 crc kubenswrapper[4824]: I1006 11:07:32.440396 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_0295abaf-72dd-46d7-9125-543921a6ef61/rabbitmq/0.log" Oct 06 11:07:32 crc kubenswrapper[4824]: I1006 11:07:32.673852 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_9b7a49d4-5482-4da9-aad2-4aabf9584534/setup-container/0.log" Oct 06 11:07:32 crc kubenswrapper[4824]: I1006 11:07:32.884118 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_9b7a49d4-5482-4da9-aad2-4aabf9584534/rabbitmq/0.log" Oct 06 11:07:32 crc kubenswrapper[4824]: I1006 11:07:32.892324 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_9b7a49d4-5482-4da9-aad2-4aabf9584534/setup-container/0.log" Oct 06 11:07:33 crc kubenswrapper[4824]: I1006 11:07:33.126293 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-97h9g_073c5011-b49a-4900-b00b-488164295d4d/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 06 11:07:33 crc kubenswrapper[4824]: I1006 11:07:33.231331 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-pqwkw_a043cdf1-f56a-415d-8914-dc689aca5a9f/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Oct 06 11:07:34 crc kubenswrapper[4824]: I1006 11:07:34.091219 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-9kpm9_4e90bc2c-77ea-40b7-87b1-82cd30dbb082/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Oct 06 11:07:34 crc kubenswrapper[4824]: I1006 11:07:34.343641 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-vwg8r_626b5d8f-93f2-4bbc-add6-a530ddf9a6fa/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 06 11:07:34 crc kubenswrapper[4824]: I1006 11:07:34.572599 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-d24cq_7dcfc0e7-18d6-4a54-9033-acf435eda511/ssh-known-hosts-edpm-deployment/0.log" Oct 06 11:07:34 crc kubenswrapper[4824]: I1006 11:07:34.822323 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-78d766d697-s7wzp_925ed890-0e01-4422-a2f8-4871bf90087d/proxy-server/0.log" Oct 06 11:07:34 crc kubenswrapper[4824]: I1006 11:07:34.830935 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-78d766d697-s7wzp_925ed890-0e01-4422-a2f8-4871bf90087d/proxy-httpd/0.log" Oct 06 11:07:35 crc kubenswrapper[4824]: I1006 11:07:35.089051 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-z6wvs_7f85d924-3fee-4cec-a9df-5a0e41567719/swift-ring-rebalance/0.log" Oct 06 11:07:35 crc kubenswrapper[4824]: I1006 11:07:35.193614 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6cdeba93-f1c9-487d-b776-5c2f1bb353d9/account-auditor/0.log" Oct 06 11:07:35 crc kubenswrapper[4824]: I1006 11:07:35.371615 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6cdeba93-f1c9-487d-b776-5c2f1bb353d9/account-reaper/0.log" Oct 06 11:07:35 crc kubenswrapper[4824]: I1006 11:07:35.446562 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6cdeba93-f1c9-487d-b776-5c2f1bb353d9/account-replicator/0.log" Oct 06 11:07:35 crc kubenswrapper[4824]: I1006 11:07:35.450059 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6cdeba93-f1c9-487d-b776-5c2f1bb353d9/account-server/0.log" Oct 06 11:07:35 crc kubenswrapper[4824]: I1006 11:07:35.559304 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6cdeba93-f1c9-487d-b776-5c2f1bb353d9/container-auditor/0.log" Oct 06 11:07:35 crc kubenswrapper[4824]: I1006 11:07:35.696820 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6cdeba93-f1c9-487d-b776-5c2f1bb353d9/container-replicator/0.log" Oct 06 11:07:35 crc kubenswrapper[4824]: I1006 11:07:35.769920 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6cdeba93-f1c9-487d-b776-5c2f1bb353d9/container-server/0.log" Oct 06 11:07:35 crc kubenswrapper[4824]: I1006 11:07:35.814681 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6cdeba93-f1c9-487d-b776-5c2f1bb353d9/container-updater/0.log" Oct 06 11:07:36 crc kubenswrapper[4824]: I1006 11:07:36.542942 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6cdeba93-f1c9-487d-b776-5c2f1bb353d9/object-auditor/0.log" Oct 06 11:07:36 crc kubenswrapper[4824]: I1006 11:07:36.550727 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6cdeba93-f1c9-487d-b776-5c2f1bb353d9/object-replicator/0.log" Oct 06 11:07:36 crc kubenswrapper[4824]: I1006 11:07:36.577621 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6cdeba93-f1c9-487d-b776-5c2f1bb353d9/object-expirer/0.log" Oct 06 11:07:36 crc kubenswrapper[4824]: I1006 11:07:36.792856 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6cdeba93-f1c9-487d-b776-5c2f1bb353d9/object-updater/0.log" Oct 06 11:07:36 crc kubenswrapper[4824]: I1006 11:07:36.805830 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6cdeba93-f1c9-487d-b776-5c2f1bb353d9/object-server/0.log" Oct 06 11:07:36 crc kubenswrapper[4824]: I1006 11:07:36.866923 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6cdeba93-f1c9-487d-b776-5c2f1bb353d9/rsync/0.log" Oct 06 11:07:37 crc kubenswrapper[4824]: I1006 11:07:37.039095 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6cdeba93-f1c9-487d-b776-5c2f1bb353d9/swift-recon-cron/0.log" Oct 06 11:07:37 crc kubenswrapper[4824]: I1006 11:07:37.150006 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-4br4q_3a410146-5f2b-48a0-824f-ee424f685694/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Oct 06 11:07:37 crc kubenswrapper[4824]: I1006 11:07:37.383830 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_7fd677dd-afa0-45a9-8198-ea78e02f4fb7/tempest-tests-tempest-tests-runner/0.log" Oct 06 11:07:37 crc kubenswrapper[4824]: I1006 11:07:37.641317 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_fbc3bf37-42b1-4a02-9922-2f8bad6f7ad5/test-operator-logs-container/0.log" Oct 06 11:07:37 crc kubenswrapper[4824]: I1006 11:07:37.698596 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-jm56f_9684787a-cab3-4930-9ada-f29df39d21a6/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Oct 06 11:07:46 crc kubenswrapper[4824]: I1006 11:07:46.777705 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_3a12653c-c0d0-46e3-b909-77bd2a7f6c4a/memcached/0.log" Oct 06 11:07:51 crc kubenswrapper[4824]: I1006 11:07:51.611573 4824 scope.go:117] "RemoveContainer" containerID="60d15bc379e2f400b7346e85651a8f0e968a4e13015c6e329a969e869bc7971c" Oct 06 11:08:07 crc kubenswrapper[4824]: I1006 11:08:07.510280 4824 generic.go:334] "Generic (PLEG): container finished" podID="acffb1f0-1d70-4905-9648-e2782b876ef8" containerID="ea6e43bb9cdcd06002cd0bc6be6805769696143892dae1049fc95d7f7f527d3e" exitCode=0 Oct 06 11:08:07 crc kubenswrapper[4824]: I1006 11:08:07.511319 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-x5ksk/crc-debug-bc9hs" event={"ID":"acffb1f0-1d70-4905-9648-e2782b876ef8","Type":"ContainerDied","Data":"ea6e43bb9cdcd06002cd0bc6be6805769696143892dae1049fc95d7f7f527d3e"} Oct 06 11:08:08 crc kubenswrapper[4824]: I1006 11:08:08.639082 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-x5ksk/crc-debug-bc9hs" Oct 06 11:08:08 crc kubenswrapper[4824]: I1006 11:08:08.689852 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-x5ksk/crc-debug-bc9hs"] Oct 06 11:08:08 crc kubenswrapper[4824]: I1006 11:08:08.700967 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-x5ksk/crc-debug-bc9hs"] Oct 06 11:08:08 crc kubenswrapper[4824]: I1006 11:08:08.801679 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zdkqs\" (UniqueName: \"kubernetes.io/projected/acffb1f0-1d70-4905-9648-e2782b876ef8-kube-api-access-zdkqs\") pod \"acffb1f0-1d70-4905-9648-e2782b876ef8\" (UID: \"acffb1f0-1d70-4905-9648-e2782b876ef8\") " Oct 06 11:08:08 crc kubenswrapper[4824]: I1006 11:08:08.801828 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/acffb1f0-1d70-4905-9648-e2782b876ef8-host\") pod \"acffb1f0-1d70-4905-9648-e2782b876ef8\" (UID: \"acffb1f0-1d70-4905-9648-e2782b876ef8\") " Oct 06 11:08:08 crc kubenswrapper[4824]: I1006 11:08:08.802445 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/acffb1f0-1d70-4905-9648-e2782b876ef8-host" (OuterVolumeSpecName: "host") pod "acffb1f0-1d70-4905-9648-e2782b876ef8" (UID: "acffb1f0-1d70-4905-9648-e2782b876ef8"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 06 11:08:08 crc kubenswrapper[4824]: I1006 11:08:08.803198 4824 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/acffb1f0-1d70-4905-9648-e2782b876ef8-host\") on node \"crc\" DevicePath \"\"" Oct 06 11:08:08 crc kubenswrapper[4824]: I1006 11:08:08.826291 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/acffb1f0-1d70-4905-9648-e2782b876ef8-kube-api-access-zdkqs" (OuterVolumeSpecName: "kube-api-access-zdkqs") pod "acffb1f0-1d70-4905-9648-e2782b876ef8" (UID: "acffb1f0-1d70-4905-9648-e2782b876ef8"). InnerVolumeSpecName "kube-api-access-zdkqs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 11:08:08 crc kubenswrapper[4824]: I1006 11:08:08.905133 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zdkqs\" (UniqueName: \"kubernetes.io/projected/acffb1f0-1d70-4905-9648-e2782b876ef8-kube-api-access-zdkqs\") on node \"crc\" DevicePath \"\"" Oct 06 11:08:09 crc kubenswrapper[4824]: I1006 11:08:09.288558 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="acffb1f0-1d70-4905-9648-e2782b876ef8" path="/var/lib/kubelet/pods/acffb1f0-1d70-4905-9648-e2782b876ef8/volumes" Oct 06 11:08:09 crc kubenswrapper[4824]: I1006 11:08:09.533044 4824 scope.go:117] "RemoveContainer" containerID="ea6e43bb9cdcd06002cd0bc6be6805769696143892dae1049fc95d7f7f527d3e" Oct 06 11:08:09 crc kubenswrapper[4824]: I1006 11:08:09.533327 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-x5ksk/crc-debug-bc9hs" Oct 06 11:08:09 crc kubenswrapper[4824]: I1006 11:08:09.953762 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-x5ksk/crc-debug-ktcsp"] Oct 06 11:08:09 crc kubenswrapper[4824]: E1006 11:08:09.955261 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="acffb1f0-1d70-4905-9648-e2782b876ef8" containerName="container-00" Oct 06 11:08:09 crc kubenswrapper[4824]: I1006 11:08:09.955294 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="acffb1f0-1d70-4905-9648-e2782b876ef8" containerName="container-00" Oct 06 11:08:09 crc kubenswrapper[4824]: I1006 11:08:09.956268 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="acffb1f0-1d70-4905-9648-e2782b876ef8" containerName="container-00" Oct 06 11:08:09 crc kubenswrapper[4824]: I1006 11:08:09.958211 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-x5ksk/crc-debug-ktcsp" Oct 06 11:08:09 crc kubenswrapper[4824]: I1006 11:08:09.969062 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-x5ksk"/"default-dockercfg-5c5x6" Oct 06 11:08:10 crc kubenswrapper[4824]: I1006 11:08:10.053387 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ced25660-0ce2-4c70-8b15-d57ff8d62c77-host\") pod \"crc-debug-ktcsp\" (UID: \"ced25660-0ce2-4c70-8b15-d57ff8d62c77\") " pod="openshift-must-gather-x5ksk/crc-debug-ktcsp" Oct 06 11:08:10 crc kubenswrapper[4824]: I1006 11:08:10.054116 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mb2ds\" (UniqueName: \"kubernetes.io/projected/ced25660-0ce2-4c70-8b15-d57ff8d62c77-kube-api-access-mb2ds\") pod \"crc-debug-ktcsp\" (UID: \"ced25660-0ce2-4c70-8b15-d57ff8d62c77\") " pod="openshift-must-gather-x5ksk/crc-debug-ktcsp" Oct 06 11:08:10 crc kubenswrapper[4824]: I1006 11:08:10.157173 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mb2ds\" (UniqueName: \"kubernetes.io/projected/ced25660-0ce2-4c70-8b15-d57ff8d62c77-kube-api-access-mb2ds\") pod \"crc-debug-ktcsp\" (UID: \"ced25660-0ce2-4c70-8b15-d57ff8d62c77\") " pod="openshift-must-gather-x5ksk/crc-debug-ktcsp" Oct 06 11:08:10 crc kubenswrapper[4824]: I1006 11:08:10.157336 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ced25660-0ce2-4c70-8b15-d57ff8d62c77-host\") pod \"crc-debug-ktcsp\" (UID: \"ced25660-0ce2-4c70-8b15-d57ff8d62c77\") " pod="openshift-must-gather-x5ksk/crc-debug-ktcsp" Oct 06 11:08:10 crc kubenswrapper[4824]: I1006 11:08:10.157503 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ced25660-0ce2-4c70-8b15-d57ff8d62c77-host\") pod \"crc-debug-ktcsp\" (UID: \"ced25660-0ce2-4c70-8b15-d57ff8d62c77\") " pod="openshift-must-gather-x5ksk/crc-debug-ktcsp" Oct 06 11:08:10 crc kubenswrapper[4824]: I1006 11:08:10.204288 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mb2ds\" (UniqueName: \"kubernetes.io/projected/ced25660-0ce2-4c70-8b15-d57ff8d62c77-kube-api-access-mb2ds\") pod \"crc-debug-ktcsp\" (UID: \"ced25660-0ce2-4c70-8b15-d57ff8d62c77\") " pod="openshift-must-gather-x5ksk/crc-debug-ktcsp" Oct 06 11:08:10 crc kubenswrapper[4824]: I1006 11:08:10.292648 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-x5ksk/crc-debug-ktcsp" Oct 06 11:08:10 crc kubenswrapper[4824]: I1006 11:08:10.547482 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-x5ksk/crc-debug-ktcsp" event={"ID":"ced25660-0ce2-4c70-8b15-d57ff8d62c77","Type":"ContainerStarted","Data":"4770e5f6ce1af7c589c2a2386484337bb32df7ebf6654617549a789395516a54"} Oct 06 11:08:11 crc kubenswrapper[4824]: I1006 11:08:11.566765 4824 generic.go:334] "Generic (PLEG): container finished" podID="ced25660-0ce2-4c70-8b15-d57ff8d62c77" containerID="ce99131703e891508801d432c329ab134661a49fe5c9d4ee1f064544d4eb835d" exitCode=0 Oct 06 11:08:11 crc kubenswrapper[4824]: I1006 11:08:11.567407 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-x5ksk/crc-debug-ktcsp" event={"ID":"ced25660-0ce2-4c70-8b15-d57ff8d62c77","Type":"ContainerDied","Data":"ce99131703e891508801d432c329ab134661a49fe5c9d4ee1f064544d4eb835d"} Oct 06 11:08:12 crc kubenswrapper[4824]: I1006 11:08:12.744872 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-x5ksk/crc-debug-ktcsp" Oct 06 11:08:12 crc kubenswrapper[4824]: I1006 11:08:12.916846 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ced25660-0ce2-4c70-8b15-d57ff8d62c77-host\") pod \"ced25660-0ce2-4c70-8b15-d57ff8d62c77\" (UID: \"ced25660-0ce2-4c70-8b15-d57ff8d62c77\") " Oct 06 11:08:12 crc kubenswrapper[4824]: I1006 11:08:12.916940 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mb2ds\" (UniqueName: \"kubernetes.io/projected/ced25660-0ce2-4c70-8b15-d57ff8d62c77-kube-api-access-mb2ds\") pod \"ced25660-0ce2-4c70-8b15-d57ff8d62c77\" (UID: \"ced25660-0ce2-4c70-8b15-d57ff8d62c77\") " Oct 06 11:08:12 crc kubenswrapper[4824]: I1006 11:08:12.918160 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ced25660-0ce2-4c70-8b15-d57ff8d62c77-host" (OuterVolumeSpecName: "host") pod "ced25660-0ce2-4c70-8b15-d57ff8d62c77" (UID: "ced25660-0ce2-4c70-8b15-d57ff8d62c77"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 06 11:08:12 crc kubenswrapper[4824]: I1006 11:08:12.934285 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ced25660-0ce2-4c70-8b15-d57ff8d62c77-kube-api-access-mb2ds" (OuterVolumeSpecName: "kube-api-access-mb2ds") pod "ced25660-0ce2-4c70-8b15-d57ff8d62c77" (UID: "ced25660-0ce2-4c70-8b15-d57ff8d62c77"). InnerVolumeSpecName "kube-api-access-mb2ds". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 11:08:13 crc kubenswrapper[4824]: I1006 11:08:13.019439 4824 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ced25660-0ce2-4c70-8b15-d57ff8d62c77-host\") on node \"crc\" DevicePath \"\"" Oct 06 11:08:13 crc kubenswrapper[4824]: I1006 11:08:13.019476 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mb2ds\" (UniqueName: \"kubernetes.io/projected/ced25660-0ce2-4c70-8b15-d57ff8d62c77-kube-api-access-mb2ds\") on node \"crc\" DevicePath \"\"" Oct 06 11:08:13 crc kubenswrapper[4824]: I1006 11:08:13.595355 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-x5ksk/crc-debug-ktcsp" event={"ID":"ced25660-0ce2-4c70-8b15-d57ff8d62c77","Type":"ContainerDied","Data":"4770e5f6ce1af7c589c2a2386484337bb32df7ebf6654617549a789395516a54"} Oct 06 11:08:13 crc kubenswrapper[4824]: I1006 11:08:13.595876 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4770e5f6ce1af7c589c2a2386484337bb32df7ebf6654617549a789395516a54" Oct 06 11:08:13 crc kubenswrapper[4824]: I1006 11:08:13.596277 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-x5ksk/crc-debug-ktcsp" Oct 06 11:08:15 crc kubenswrapper[4824]: I1006 11:08:15.337824 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-2fsf8"] Oct 06 11:08:15 crc kubenswrapper[4824]: E1006 11:08:15.338785 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ced25660-0ce2-4c70-8b15-d57ff8d62c77" containerName="container-00" Oct 06 11:08:15 crc kubenswrapper[4824]: I1006 11:08:15.338802 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="ced25660-0ce2-4c70-8b15-d57ff8d62c77" containerName="container-00" Oct 06 11:08:15 crc kubenswrapper[4824]: I1006 11:08:15.339053 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="ced25660-0ce2-4c70-8b15-d57ff8d62c77" containerName="container-00" Oct 06 11:08:15 crc kubenswrapper[4824]: I1006 11:08:15.340650 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2fsf8" Oct 06 11:08:15 crc kubenswrapper[4824]: I1006 11:08:15.354535 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-2fsf8"] Oct 06 11:08:15 crc kubenswrapper[4824]: I1006 11:08:15.473462 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c94c3574-b827-4045-869d-c9148e62cb46-utilities\") pod \"redhat-operators-2fsf8\" (UID: \"c94c3574-b827-4045-869d-c9148e62cb46\") " pod="openshift-marketplace/redhat-operators-2fsf8" Oct 06 11:08:15 crc kubenswrapper[4824]: I1006 11:08:15.473533 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zmnpr\" (UniqueName: \"kubernetes.io/projected/c94c3574-b827-4045-869d-c9148e62cb46-kube-api-access-zmnpr\") pod \"redhat-operators-2fsf8\" (UID: \"c94c3574-b827-4045-869d-c9148e62cb46\") " pod="openshift-marketplace/redhat-operators-2fsf8" Oct 06 11:08:15 crc kubenswrapper[4824]: I1006 11:08:15.473641 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c94c3574-b827-4045-869d-c9148e62cb46-catalog-content\") pod \"redhat-operators-2fsf8\" (UID: \"c94c3574-b827-4045-869d-c9148e62cb46\") " pod="openshift-marketplace/redhat-operators-2fsf8" Oct 06 11:08:15 crc kubenswrapper[4824]: I1006 11:08:15.575412 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c94c3574-b827-4045-869d-c9148e62cb46-catalog-content\") pod \"redhat-operators-2fsf8\" (UID: \"c94c3574-b827-4045-869d-c9148e62cb46\") " pod="openshift-marketplace/redhat-operators-2fsf8" Oct 06 11:08:15 crc kubenswrapper[4824]: I1006 11:08:15.575523 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c94c3574-b827-4045-869d-c9148e62cb46-utilities\") pod \"redhat-operators-2fsf8\" (UID: \"c94c3574-b827-4045-869d-c9148e62cb46\") " pod="openshift-marketplace/redhat-operators-2fsf8" Oct 06 11:08:15 crc kubenswrapper[4824]: I1006 11:08:15.575550 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zmnpr\" (UniqueName: \"kubernetes.io/projected/c94c3574-b827-4045-869d-c9148e62cb46-kube-api-access-zmnpr\") pod \"redhat-operators-2fsf8\" (UID: \"c94c3574-b827-4045-869d-c9148e62cb46\") " pod="openshift-marketplace/redhat-operators-2fsf8" Oct 06 11:08:15 crc kubenswrapper[4824]: I1006 11:08:15.576486 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c94c3574-b827-4045-869d-c9148e62cb46-utilities\") pod \"redhat-operators-2fsf8\" (UID: \"c94c3574-b827-4045-869d-c9148e62cb46\") " pod="openshift-marketplace/redhat-operators-2fsf8" Oct 06 11:08:15 crc kubenswrapper[4824]: I1006 11:08:15.576579 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c94c3574-b827-4045-869d-c9148e62cb46-catalog-content\") pod \"redhat-operators-2fsf8\" (UID: \"c94c3574-b827-4045-869d-c9148e62cb46\") " pod="openshift-marketplace/redhat-operators-2fsf8" Oct 06 11:08:15 crc kubenswrapper[4824]: I1006 11:08:15.600899 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zmnpr\" (UniqueName: \"kubernetes.io/projected/c94c3574-b827-4045-869d-c9148e62cb46-kube-api-access-zmnpr\") pod \"redhat-operators-2fsf8\" (UID: \"c94c3574-b827-4045-869d-c9148e62cb46\") " pod="openshift-marketplace/redhat-operators-2fsf8" Oct 06 11:08:15 crc kubenswrapper[4824]: I1006 11:08:15.684649 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2fsf8" Oct 06 11:08:16 crc kubenswrapper[4824]: I1006 11:08:16.321797 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-2fsf8"] Oct 06 11:08:16 crc kubenswrapper[4824]: I1006 11:08:16.634042 4824 generic.go:334] "Generic (PLEG): container finished" podID="c94c3574-b827-4045-869d-c9148e62cb46" containerID="585fefe35e4e5f30dc9a4a134301ff569dbf2fe2bd57e16995e08e33a29b62fe" exitCode=0 Oct 06 11:08:16 crc kubenswrapper[4824]: I1006 11:08:16.634110 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2fsf8" event={"ID":"c94c3574-b827-4045-869d-c9148e62cb46","Type":"ContainerDied","Data":"585fefe35e4e5f30dc9a4a134301ff569dbf2fe2bd57e16995e08e33a29b62fe"} Oct 06 11:08:16 crc kubenswrapper[4824]: I1006 11:08:16.634501 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2fsf8" event={"ID":"c94c3574-b827-4045-869d-c9148e62cb46","Type":"ContainerStarted","Data":"ef306fa45572f601b30716017d532db8d23d8bd370029fc9efad7a8c317ef1cc"} Oct 06 11:08:18 crc kubenswrapper[4824]: I1006 11:08:18.661634 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2fsf8" event={"ID":"c94c3574-b827-4045-869d-c9148e62cb46","Type":"ContainerStarted","Data":"a3e2116ff10f74ff5ef60ef85e83dc0d363e0dd959fba625952bc302c3db66f6"} Oct 06 11:08:19 crc kubenswrapper[4824]: I1006 11:08:19.313348 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-x5ksk/crc-debug-ktcsp"] Oct 06 11:08:19 crc kubenswrapper[4824]: I1006 11:08:19.325623 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-x5ksk/crc-debug-ktcsp"] Oct 06 11:08:20 crc kubenswrapper[4824]: I1006 11:08:20.646812 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-x5ksk/crc-debug-pltbs"] Oct 06 11:08:20 crc kubenswrapper[4824]: I1006 11:08:20.649056 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-x5ksk/crc-debug-pltbs" Oct 06 11:08:20 crc kubenswrapper[4824]: I1006 11:08:20.651887 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-x5ksk"/"default-dockercfg-5c5x6" Oct 06 11:08:20 crc kubenswrapper[4824]: I1006 11:08:20.684890 4824 generic.go:334] "Generic (PLEG): container finished" podID="c94c3574-b827-4045-869d-c9148e62cb46" containerID="a3e2116ff10f74ff5ef60ef85e83dc0d363e0dd959fba625952bc302c3db66f6" exitCode=0 Oct 06 11:08:20 crc kubenswrapper[4824]: I1006 11:08:20.684943 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2fsf8" event={"ID":"c94c3574-b827-4045-869d-c9148e62cb46","Type":"ContainerDied","Data":"a3e2116ff10f74ff5ef60ef85e83dc0d363e0dd959fba625952bc302c3db66f6"} Oct 06 11:08:20 crc kubenswrapper[4824]: I1006 11:08:20.794287 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xtlj9\" (UniqueName: \"kubernetes.io/projected/9c7079f3-84d9-481d-b5fd-090ad21f4678-kube-api-access-xtlj9\") pod \"crc-debug-pltbs\" (UID: \"9c7079f3-84d9-481d-b5fd-090ad21f4678\") " pod="openshift-must-gather-x5ksk/crc-debug-pltbs" Oct 06 11:08:20 crc kubenswrapper[4824]: I1006 11:08:20.794457 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9c7079f3-84d9-481d-b5fd-090ad21f4678-host\") pod \"crc-debug-pltbs\" (UID: \"9c7079f3-84d9-481d-b5fd-090ad21f4678\") " pod="openshift-must-gather-x5ksk/crc-debug-pltbs" Oct 06 11:08:20 crc kubenswrapper[4824]: I1006 11:08:20.897135 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xtlj9\" (UniqueName: \"kubernetes.io/projected/9c7079f3-84d9-481d-b5fd-090ad21f4678-kube-api-access-xtlj9\") pod \"crc-debug-pltbs\" (UID: \"9c7079f3-84d9-481d-b5fd-090ad21f4678\") " pod="openshift-must-gather-x5ksk/crc-debug-pltbs" Oct 06 11:08:20 crc kubenswrapper[4824]: I1006 11:08:20.897293 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9c7079f3-84d9-481d-b5fd-090ad21f4678-host\") pod \"crc-debug-pltbs\" (UID: \"9c7079f3-84d9-481d-b5fd-090ad21f4678\") " pod="openshift-must-gather-x5ksk/crc-debug-pltbs" Oct 06 11:08:20 crc kubenswrapper[4824]: I1006 11:08:20.897518 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9c7079f3-84d9-481d-b5fd-090ad21f4678-host\") pod \"crc-debug-pltbs\" (UID: \"9c7079f3-84d9-481d-b5fd-090ad21f4678\") " pod="openshift-must-gather-x5ksk/crc-debug-pltbs" Oct 06 11:08:20 crc kubenswrapper[4824]: I1006 11:08:20.928555 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xtlj9\" (UniqueName: \"kubernetes.io/projected/9c7079f3-84d9-481d-b5fd-090ad21f4678-kube-api-access-xtlj9\") pod \"crc-debug-pltbs\" (UID: \"9c7079f3-84d9-481d-b5fd-090ad21f4678\") " pod="openshift-must-gather-x5ksk/crc-debug-pltbs" Oct 06 11:08:20 crc kubenswrapper[4824]: I1006 11:08:20.978071 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-x5ksk/crc-debug-pltbs" Oct 06 11:08:21 crc kubenswrapper[4824]: W1006 11:08:21.013034 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9c7079f3_84d9_481d_b5fd_090ad21f4678.slice/crio-4e6362d9f8dbad3d592a234cd163788a1a89ac9526f21e829a1adf764d04ed98 WatchSource:0}: Error finding container 4e6362d9f8dbad3d592a234cd163788a1a89ac9526f21e829a1adf764d04ed98: Status 404 returned error can't find the container with id 4e6362d9f8dbad3d592a234cd163788a1a89ac9526f21e829a1adf764d04ed98 Oct 06 11:08:21 crc kubenswrapper[4824]: I1006 11:08:21.292230 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ced25660-0ce2-4c70-8b15-d57ff8d62c77" path="/var/lib/kubelet/pods/ced25660-0ce2-4c70-8b15-d57ff8d62c77/volumes" Oct 06 11:08:21 crc kubenswrapper[4824]: I1006 11:08:21.701861 4824 generic.go:334] "Generic (PLEG): container finished" podID="9c7079f3-84d9-481d-b5fd-090ad21f4678" containerID="ce0486bfbea8b804059ab65d9cc4f513314892322c326c6e8086e77d8b1139b6" exitCode=0 Oct 06 11:08:21 crc kubenswrapper[4824]: I1006 11:08:21.701963 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-x5ksk/crc-debug-pltbs" event={"ID":"9c7079f3-84d9-481d-b5fd-090ad21f4678","Type":"ContainerDied","Data":"ce0486bfbea8b804059ab65d9cc4f513314892322c326c6e8086e77d8b1139b6"} Oct 06 11:08:21 crc kubenswrapper[4824]: I1006 11:08:21.702032 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-x5ksk/crc-debug-pltbs" event={"ID":"9c7079f3-84d9-481d-b5fd-090ad21f4678","Type":"ContainerStarted","Data":"4e6362d9f8dbad3d592a234cd163788a1a89ac9526f21e829a1adf764d04ed98"} Oct 06 11:08:21 crc kubenswrapper[4824]: I1006 11:08:21.705292 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2fsf8" event={"ID":"c94c3574-b827-4045-869d-c9148e62cb46","Type":"ContainerStarted","Data":"d5fcb607eb9652f0b0b0277d746e580d5e7cc05f99f1260efb9fba869e4069a0"} Oct 06 11:08:21 crc kubenswrapper[4824]: I1006 11:08:21.751253 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-2fsf8" podStartSLOduration=2.224723524 podStartE2EDuration="6.751233168s" podCreationTimestamp="2025-10-06 11:08:15 +0000 UTC" firstStartedPulling="2025-10-06 11:08:16.636245852 +0000 UTC m=+4206.000668713" lastFinishedPulling="2025-10-06 11:08:21.162755496 +0000 UTC m=+4210.527178357" observedRunningTime="2025-10-06 11:08:21.749145527 +0000 UTC m=+4211.113568408" watchObservedRunningTime="2025-10-06 11:08:21.751233168 +0000 UTC m=+4211.115656039" Oct 06 11:08:21 crc kubenswrapper[4824]: I1006 11:08:21.784216 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-x5ksk/crc-debug-pltbs"] Oct 06 11:08:21 crc kubenswrapper[4824]: I1006 11:08:21.797767 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-x5ksk/crc-debug-pltbs"] Oct 06 11:08:22 crc kubenswrapper[4824]: I1006 11:08:22.850571 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-x5ksk/crc-debug-pltbs" Oct 06 11:08:22 crc kubenswrapper[4824]: I1006 11:08:22.947860 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9c7079f3-84d9-481d-b5fd-090ad21f4678-host\") pod \"9c7079f3-84d9-481d-b5fd-090ad21f4678\" (UID: \"9c7079f3-84d9-481d-b5fd-090ad21f4678\") " Oct 06 11:08:22 crc kubenswrapper[4824]: I1006 11:08:22.947961 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xtlj9\" (UniqueName: \"kubernetes.io/projected/9c7079f3-84d9-481d-b5fd-090ad21f4678-kube-api-access-xtlj9\") pod \"9c7079f3-84d9-481d-b5fd-090ad21f4678\" (UID: \"9c7079f3-84d9-481d-b5fd-090ad21f4678\") " Oct 06 11:08:22 crc kubenswrapper[4824]: I1006 11:08:22.948077 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9c7079f3-84d9-481d-b5fd-090ad21f4678-host" (OuterVolumeSpecName: "host") pod "9c7079f3-84d9-481d-b5fd-090ad21f4678" (UID: "9c7079f3-84d9-481d-b5fd-090ad21f4678"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 06 11:08:22 crc kubenswrapper[4824]: I1006 11:08:22.948868 4824 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9c7079f3-84d9-481d-b5fd-090ad21f4678-host\") on node \"crc\" DevicePath \"\"" Oct 06 11:08:22 crc kubenswrapper[4824]: I1006 11:08:22.969375 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9c7079f3-84d9-481d-b5fd-090ad21f4678-kube-api-access-xtlj9" (OuterVolumeSpecName: "kube-api-access-xtlj9") pod "9c7079f3-84d9-481d-b5fd-090ad21f4678" (UID: "9c7079f3-84d9-481d-b5fd-090ad21f4678"). InnerVolumeSpecName "kube-api-access-xtlj9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 11:08:23 crc kubenswrapper[4824]: I1006 11:08:23.051077 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xtlj9\" (UniqueName: \"kubernetes.io/projected/9c7079f3-84d9-481d-b5fd-090ad21f4678-kube-api-access-xtlj9\") on node \"crc\" DevicePath \"\"" Oct 06 11:08:23 crc kubenswrapper[4824]: I1006 11:08:23.285393 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9c7079f3-84d9-481d-b5fd-090ad21f4678" path="/var/lib/kubelet/pods/9c7079f3-84d9-481d-b5fd-090ad21f4678/volumes" Oct 06 11:08:23 crc kubenswrapper[4824]: I1006 11:08:23.733873 4824 scope.go:117] "RemoveContainer" containerID="ce0486bfbea8b804059ab65d9cc4f513314892322c326c6e8086e77d8b1139b6" Oct 06 11:08:23 crc kubenswrapper[4824]: I1006 11:08:23.734056 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-x5ksk/crc-debug-pltbs" Oct 06 11:08:23 crc kubenswrapper[4824]: I1006 11:08:23.892148 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_a842b0520c38c46f4cb51faa08e79e3c8193222ff536aee10057c0eb8cqhcsj_65505c4e-e82d-485f-abc6-33be1406fd9a/util/0.log" Oct 06 11:08:24 crc kubenswrapper[4824]: I1006 11:08:24.109814 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_a842b0520c38c46f4cb51faa08e79e3c8193222ff536aee10057c0eb8cqhcsj_65505c4e-e82d-485f-abc6-33be1406fd9a/pull/0.log" Oct 06 11:08:24 crc kubenswrapper[4824]: I1006 11:08:24.160851 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_a842b0520c38c46f4cb51faa08e79e3c8193222ff536aee10057c0eb8cqhcsj_65505c4e-e82d-485f-abc6-33be1406fd9a/pull/0.log" Oct 06 11:08:24 crc kubenswrapper[4824]: I1006 11:08:24.216759 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_a842b0520c38c46f4cb51faa08e79e3c8193222ff536aee10057c0eb8cqhcsj_65505c4e-e82d-485f-abc6-33be1406fd9a/util/0.log" Oct 06 11:08:24 crc kubenswrapper[4824]: I1006 11:08:24.339658 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_a842b0520c38c46f4cb51faa08e79e3c8193222ff536aee10057c0eb8cqhcsj_65505c4e-e82d-485f-abc6-33be1406fd9a/pull/0.log" Oct 06 11:08:24 crc kubenswrapper[4824]: I1006 11:08:24.359187 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_a842b0520c38c46f4cb51faa08e79e3c8193222ff536aee10057c0eb8cqhcsj_65505c4e-e82d-485f-abc6-33be1406fd9a/util/0.log" Oct 06 11:08:24 crc kubenswrapper[4824]: I1006 11:08:24.477039 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_a842b0520c38c46f4cb51faa08e79e3c8193222ff536aee10057c0eb8cqhcsj_65505c4e-e82d-485f-abc6-33be1406fd9a/extract/0.log" Oct 06 11:08:24 crc kubenswrapper[4824]: I1006 11:08:24.576273 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-5f7c849b98-fn8nc_e60e7a0d-ccc6-4b1a-b645-a4802b21d48e/kube-rbac-proxy/0.log" Oct 06 11:08:25 crc kubenswrapper[4824]: I1006 11:08:25.198466 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-7d4d4f8d-4kn7z_b3c0daed-de3b-415e-9166-756cb0f5cab7/kube-rbac-proxy/0.log" Oct 06 11:08:25 crc kubenswrapper[4824]: I1006 11:08:25.258837 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-5f7c849b98-fn8nc_e60e7a0d-ccc6-4b1a-b645-a4802b21d48e/manager/0.log" Oct 06 11:08:25 crc kubenswrapper[4824]: I1006 11:08:25.359683 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-7d4d4f8d-4kn7z_b3c0daed-de3b-415e-9166-756cb0f5cab7/manager/0.log" Oct 06 11:08:25 crc kubenswrapper[4824]: I1006 11:08:25.686051 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-2fsf8" Oct 06 11:08:25 crc kubenswrapper[4824]: I1006 11:08:25.686141 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-2fsf8" Oct 06 11:08:25 crc kubenswrapper[4824]: I1006 11:08:25.762635 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-75dfd9b554-hktsj_4ca61d7c-741e-4028-8a75-d26a7b4cbd3a/kube-rbac-proxy/0.log" Oct 06 11:08:25 crc kubenswrapper[4824]: I1006 11:08:25.862366 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-75dfd9b554-hktsj_4ca61d7c-741e-4028-8a75-d26a7b4cbd3a/manager/0.log" Oct 06 11:08:26 crc kubenswrapper[4824]: I1006 11:08:26.004504 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-5568b5d68-r996x_e8ad16ea-027c-4ca7-accd-928ba6faf830/kube-rbac-proxy/0.log" Oct 06 11:08:26 crc kubenswrapper[4824]: I1006 11:08:26.107640 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-5568b5d68-r996x_e8ad16ea-027c-4ca7-accd-928ba6faf830/manager/0.log" Oct 06 11:08:26 crc kubenswrapper[4824]: I1006 11:08:26.131510 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-8f58bc9db-tp4b2_76419802-f7a7-4270-9460-3cc1a9f7f667/kube-rbac-proxy/0.log" Oct 06 11:08:26 crc kubenswrapper[4824]: I1006 11:08:26.270259 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-8f58bc9db-tp4b2_76419802-f7a7-4270-9460-3cc1a9f7f667/manager/0.log" Oct 06 11:08:26 crc kubenswrapper[4824]: I1006 11:08:26.399432 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-54876c876f-swsgf_67824860-53b9-4c9c-95a1-955d9139f6e8/kube-rbac-proxy/0.log" Oct 06 11:08:26 crc kubenswrapper[4824]: I1006 11:08:26.536772 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-54876c876f-swsgf_67824860-53b9-4c9c-95a1-955d9139f6e8/manager/0.log" Oct 06 11:08:26 crc kubenswrapper[4824]: I1006 11:08:26.595499 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-658588b8c9-6884h_98cddaff-3395-473c-afd4-02c849951d51/kube-rbac-proxy/0.log" Oct 06 11:08:26 crc kubenswrapper[4824]: I1006 11:08:26.735620 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-2fsf8" podUID="c94c3574-b827-4045-869d-c9148e62cb46" containerName="registry-server" probeResult="failure" output=< Oct 06 11:08:26 crc kubenswrapper[4824]: timeout: failed to connect service ":50051" within 1s Oct 06 11:08:26 crc kubenswrapper[4824]: > Oct 06 11:08:26 crc kubenswrapper[4824]: I1006 11:08:26.832575 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-699b87f775-8mvml_09c09b76-0deb-44f1-bc93-b4479b7baa36/kube-rbac-proxy/0.log" Oct 06 11:08:26 crc kubenswrapper[4824]: I1006 11:08:26.855877 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-699b87f775-8mvml_09c09b76-0deb-44f1-bc93-b4479b7baa36/manager/0.log" Oct 06 11:08:26 crc kubenswrapper[4824]: I1006 11:08:26.921419 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-658588b8c9-6884h_98cddaff-3395-473c-afd4-02c849951d51/manager/0.log" Oct 06 11:08:27 crc kubenswrapper[4824]: I1006 11:08:27.589137 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-655d88ccb9-t8sjr_dd382562-c508-4b17-ae00-5db8abe20596/kube-rbac-proxy/0.log" Oct 06 11:08:27 crc kubenswrapper[4824]: I1006 11:08:27.639348 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-655d88ccb9-t8sjr_dd382562-c508-4b17-ae00-5db8abe20596/manager/0.log" Oct 06 11:08:27 crc kubenswrapper[4824]: I1006 11:08:27.673616 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-65d89cfd9f-gts97_21123da0-cde6-45dd-9f50-f3e44a1f78c2/kube-rbac-proxy/0.log" Oct 06 11:08:27 crc kubenswrapper[4824]: I1006 11:08:27.796537 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-65d89cfd9f-gts97_21123da0-cde6-45dd-9f50-f3e44a1f78c2/manager/0.log" Oct 06 11:08:27 crc kubenswrapper[4824]: I1006 11:08:27.828697 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-6cd6d7bdf5-4zc5b_84a43a6c-6456-47ed-ae93-e4476d3cc978/kube-rbac-proxy/0.log" Oct 06 11:08:27 crc kubenswrapper[4824]: I1006 11:08:27.855494 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-6cd6d7bdf5-4zc5b_84a43a6c-6456-47ed-ae93-e4476d3cc978/manager/0.log" Oct 06 11:08:27 crc kubenswrapper[4824]: I1006 11:08:27.989001 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-8d984cc4d-82bf8_36dbefcc-0dac-4435-aaac-5248cd2eb209/kube-rbac-proxy/0.log" Oct 06 11:08:28 crc kubenswrapper[4824]: I1006 11:08:28.073331 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-7c7fc454ff-fjjsf_dff987b5-f7ce-4eb5-9778-3cb34663824a/kube-rbac-proxy/0.log" Oct 06 11:08:28 crc kubenswrapper[4824]: I1006 11:08:28.107157 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-8d984cc4d-82bf8_36dbefcc-0dac-4435-aaac-5248cd2eb209/manager/0.log" Oct 06 11:08:28 crc kubenswrapper[4824]: I1006 11:08:28.246679 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-7c7fc454ff-fjjsf_dff987b5-f7ce-4eb5-9778-3cb34663824a/manager/0.log" Oct 06 11:08:28 crc kubenswrapper[4824]: I1006 11:08:28.267393 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-7468f855d8-kpbcv_c0da600c-cfd5-405a-82ca-ebbf3c474e44/kube-rbac-proxy/0.log" Oct 06 11:08:28 crc kubenswrapper[4824]: I1006 11:08:28.300648 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-7468f855d8-kpbcv_c0da600c-cfd5-405a-82ca-ebbf3c474e44/manager/0.log" Oct 06 11:08:28 crc kubenswrapper[4824]: I1006 11:08:28.449025 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-5dfbbd665cxcwhp_fa5da9bb-805d-420f-adc6-e948cf910b21/kube-rbac-proxy/0.log" Oct 06 11:08:28 crc kubenswrapper[4824]: I1006 11:08:28.462024 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-5dfbbd665cxcwhp_fa5da9bb-805d-420f-adc6-e948cf910b21/manager/0.log" Oct 06 11:08:28 crc kubenswrapper[4824]: I1006 11:08:28.616933 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-844bbc956-tjxqg_993a8c7b-2aac-4de6-934d-b591cb6bfffe/kube-rbac-proxy/0.log" Oct 06 11:08:28 crc kubenswrapper[4824]: I1006 11:08:28.770486 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-58f9547c95-dncvl_c1fed37a-a4ba-43eb-82fa-5b04f6a39ead/kube-rbac-proxy/0.log" Oct 06 11:08:28 crc kubenswrapper[4824]: I1006 11:08:28.958068 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-58f9547c95-dncvl_c1fed37a-a4ba-43eb-82fa-5b04f6a39ead/operator/0.log" Oct 06 11:08:29 crc kubenswrapper[4824]: I1006 11:08:29.063040 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-dlhm7_bd9ffa83-7e81-4322-9507-e92aeec0276d/registry-server/0.log" Oct 06 11:08:29 crc kubenswrapper[4824]: I1006 11:08:29.270198 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-579449c7d5-w7k8s_12802c32-cb1d-452b-8240-9a5e1e40a163/kube-rbac-proxy/0.log" Oct 06 11:08:29 crc kubenswrapper[4824]: I1006 11:08:29.357771 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-579449c7d5-w7k8s_12802c32-cb1d-452b-8240-9a5e1e40a163/manager/0.log" Oct 06 11:08:29 crc kubenswrapper[4824]: I1006 11:08:29.361462 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-54689d9f88-8qfvp_1778e737-b572-417b-931d-0eec04a89ac9/kube-rbac-proxy/0.log" Oct 06 11:08:29 crc kubenswrapper[4824]: I1006 11:08:29.543908 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-54689d9f88-8qfvp_1778e737-b572-417b-931d-0eec04a89ac9/manager/0.log" Oct 06 11:08:29 crc kubenswrapper[4824]: I1006 11:08:29.631640 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-5f97d8c699-c8wx2_1f00bc57-93de-431a-adcf-d1e1221121a1/operator/0.log" Oct 06 11:08:29 crc kubenswrapper[4824]: I1006 11:08:29.810409 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-6859f9b676-n84pw_b32a09cc-901f-41f2-9912-628bbc33da7b/kube-rbac-proxy/0.log" Oct 06 11:08:29 crc kubenswrapper[4824]: I1006 11:08:29.915082 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-84c678c567-wzzrh_7c4f21a3-07f8-46d8-bd7b-e2a24bf280d8/kube-rbac-proxy/0.log" Oct 06 11:08:29 crc kubenswrapper[4824]: I1006 11:08:29.942900 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-844bbc956-tjxqg_993a8c7b-2aac-4de6-934d-b591cb6bfffe/manager/0.log" Oct 06 11:08:29 crc kubenswrapper[4824]: I1006 11:08:29.951575 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-6859f9b676-n84pw_b32a09cc-901f-41f2-9912-628bbc33da7b/manager/0.log" Oct 06 11:08:30 crc kubenswrapper[4824]: I1006 11:08:30.109096 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-84c678c567-wzzrh_7c4f21a3-07f8-46d8-bd7b-e2a24bf280d8/manager/0.log" Oct 06 11:08:30 crc kubenswrapper[4824]: I1006 11:08:30.132684 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5cd5cb47d7-6qgln_755275e4-c198-49df-90b5-0688a4fb8228/kube-rbac-proxy/0.log" Oct 06 11:08:30 crc kubenswrapper[4824]: I1006 11:08:30.373056 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5cd5cb47d7-6qgln_755275e4-c198-49df-90b5-0688a4fb8228/manager/0.log" Oct 06 11:08:30 crc kubenswrapper[4824]: I1006 11:08:30.483260 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-6cbc6dd547-dglbk_2a48640f-172c-4e24-8c75-5c36d26ae1aa/manager/0.log" Oct 06 11:08:30 crc kubenswrapper[4824]: I1006 11:08:30.515138 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-6cbc6dd547-dglbk_2a48640f-172c-4e24-8c75-5c36d26ae1aa/kube-rbac-proxy/0.log" Oct 06 11:08:36 crc kubenswrapper[4824]: I1006 11:08:36.747286 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-2fsf8" podUID="c94c3574-b827-4045-869d-c9148e62cb46" containerName="registry-server" probeResult="failure" output=< Oct 06 11:08:36 crc kubenswrapper[4824]: timeout: failed to connect service ":50051" within 1s Oct 06 11:08:36 crc kubenswrapper[4824]: > Oct 06 11:08:45 crc kubenswrapper[4824]: I1006 11:08:45.756477 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-2fsf8" Oct 06 11:08:45 crc kubenswrapper[4824]: I1006 11:08:45.837574 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-2fsf8" Oct 06 11:08:46 crc kubenswrapper[4824]: I1006 11:08:46.522258 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-2fsf8"] Oct 06 11:08:47 crc kubenswrapper[4824]: I1006 11:08:47.020852 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-2fsf8" podUID="c94c3574-b827-4045-869d-c9148e62cb46" containerName="registry-server" containerID="cri-o://d5fcb607eb9652f0b0b0277d746e580d5e7cc05f99f1260efb9fba869e4069a0" gracePeriod=2 Oct 06 11:08:47 crc kubenswrapper[4824]: I1006 11:08:47.643721 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2fsf8" Oct 06 11:08:47 crc kubenswrapper[4824]: I1006 11:08:47.752573 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zmnpr\" (UniqueName: \"kubernetes.io/projected/c94c3574-b827-4045-869d-c9148e62cb46-kube-api-access-zmnpr\") pod \"c94c3574-b827-4045-869d-c9148e62cb46\" (UID: \"c94c3574-b827-4045-869d-c9148e62cb46\") " Oct 06 11:08:47 crc kubenswrapper[4824]: I1006 11:08:47.753130 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c94c3574-b827-4045-869d-c9148e62cb46-utilities\") pod \"c94c3574-b827-4045-869d-c9148e62cb46\" (UID: \"c94c3574-b827-4045-869d-c9148e62cb46\") " Oct 06 11:08:47 crc kubenswrapper[4824]: I1006 11:08:47.753173 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c94c3574-b827-4045-869d-c9148e62cb46-catalog-content\") pod \"c94c3574-b827-4045-869d-c9148e62cb46\" (UID: \"c94c3574-b827-4045-869d-c9148e62cb46\") " Oct 06 11:08:47 crc kubenswrapper[4824]: I1006 11:08:47.754301 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c94c3574-b827-4045-869d-c9148e62cb46-utilities" (OuterVolumeSpecName: "utilities") pod "c94c3574-b827-4045-869d-c9148e62cb46" (UID: "c94c3574-b827-4045-869d-c9148e62cb46"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 11:08:47 crc kubenswrapper[4824]: I1006 11:08:47.760288 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c94c3574-b827-4045-869d-c9148e62cb46-kube-api-access-zmnpr" (OuterVolumeSpecName: "kube-api-access-zmnpr") pod "c94c3574-b827-4045-869d-c9148e62cb46" (UID: "c94c3574-b827-4045-869d-c9148e62cb46"). InnerVolumeSpecName "kube-api-access-zmnpr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 11:08:47 crc kubenswrapper[4824]: I1006 11:08:47.844040 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c94c3574-b827-4045-869d-c9148e62cb46-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c94c3574-b827-4045-869d-c9148e62cb46" (UID: "c94c3574-b827-4045-869d-c9148e62cb46"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 11:08:47 crc kubenswrapper[4824]: I1006 11:08:47.856325 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c94c3574-b827-4045-869d-c9148e62cb46-utilities\") on node \"crc\" DevicePath \"\"" Oct 06 11:08:47 crc kubenswrapper[4824]: I1006 11:08:47.856366 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c94c3574-b827-4045-869d-c9148e62cb46-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 06 11:08:47 crc kubenswrapper[4824]: I1006 11:08:47.856385 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zmnpr\" (UniqueName: \"kubernetes.io/projected/c94c3574-b827-4045-869d-c9148e62cb46-kube-api-access-zmnpr\") on node \"crc\" DevicePath \"\"" Oct 06 11:08:48 crc kubenswrapper[4824]: I1006 11:08:48.031935 4824 generic.go:334] "Generic (PLEG): container finished" podID="c94c3574-b827-4045-869d-c9148e62cb46" containerID="d5fcb607eb9652f0b0b0277d746e580d5e7cc05f99f1260efb9fba869e4069a0" exitCode=0 Oct 06 11:08:48 crc kubenswrapper[4824]: I1006 11:08:48.032101 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2fsf8" event={"ID":"c94c3574-b827-4045-869d-c9148e62cb46","Type":"ContainerDied","Data":"d5fcb607eb9652f0b0b0277d746e580d5e7cc05f99f1260efb9fba869e4069a0"} Oct 06 11:08:48 crc kubenswrapper[4824]: I1006 11:08:48.032153 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2fsf8" event={"ID":"c94c3574-b827-4045-869d-c9148e62cb46","Type":"ContainerDied","Data":"ef306fa45572f601b30716017d532db8d23d8bd370029fc9efad7a8c317ef1cc"} Oct 06 11:08:48 crc kubenswrapper[4824]: I1006 11:08:48.032185 4824 scope.go:117] "RemoveContainer" containerID="d5fcb607eb9652f0b0b0277d746e580d5e7cc05f99f1260efb9fba869e4069a0" Oct 06 11:08:48 crc kubenswrapper[4824]: I1006 11:08:48.032386 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2fsf8" Oct 06 11:08:48 crc kubenswrapper[4824]: I1006 11:08:48.057424 4824 scope.go:117] "RemoveContainer" containerID="a3e2116ff10f74ff5ef60ef85e83dc0d363e0dd959fba625952bc302c3db66f6" Oct 06 11:08:48 crc kubenswrapper[4824]: I1006 11:08:48.094778 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-2fsf8"] Oct 06 11:08:48 crc kubenswrapper[4824]: I1006 11:08:48.108848 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-2fsf8"] Oct 06 11:08:48 crc kubenswrapper[4824]: I1006 11:08:48.126043 4824 scope.go:117] "RemoveContainer" containerID="585fefe35e4e5f30dc9a4a134301ff569dbf2fe2bd57e16995e08e33a29b62fe" Oct 06 11:08:48 crc kubenswrapper[4824]: I1006 11:08:48.155561 4824 scope.go:117] "RemoveContainer" containerID="d5fcb607eb9652f0b0b0277d746e580d5e7cc05f99f1260efb9fba869e4069a0" Oct 06 11:08:48 crc kubenswrapper[4824]: E1006 11:08:48.156085 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d5fcb607eb9652f0b0b0277d746e580d5e7cc05f99f1260efb9fba869e4069a0\": container with ID starting with d5fcb607eb9652f0b0b0277d746e580d5e7cc05f99f1260efb9fba869e4069a0 not found: ID does not exist" containerID="d5fcb607eb9652f0b0b0277d746e580d5e7cc05f99f1260efb9fba869e4069a0" Oct 06 11:08:48 crc kubenswrapper[4824]: I1006 11:08:48.156117 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d5fcb607eb9652f0b0b0277d746e580d5e7cc05f99f1260efb9fba869e4069a0"} err="failed to get container status \"d5fcb607eb9652f0b0b0277d746e580d5e7cc05f99f1260efb9fba869e4069a0\": rpc error: code = NotFound desc = could not find container \"d5fcb607eb9652f0b0b0277d746e580d5e7cc05f99f1260efb9fba869e4069a0\": container with ID starting with d5fcb607eb9652f0b0b0277d746e580d5e7cc05f99f1260efb9fba869e4069a0 not found: ID does not exist" Oct 06 11:08:48 crc kubenswrapper[4824]: I1006 11:08:48.156139 4824 scope.go:117] "RemoveContainer" containerID="a3e2116ff10f74ff5ef60ef85e83dc0d363e0dd959fba625952bc302c3db66f6" Oct 06 11:08:48 crc kubenswrapper[4824]: E1006 11:08:48.156346 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a3e2116ff10f74ff5ef60ef85e83dc0d363e0dd959fba625952bc302c3db66f6\": container with ID starting with a3e2116ff10f74ff5ef60ef85e83dc0d363e0dd959fba625952bc302c3db66f6 not found: ID does not exist" containerID="a3e2116ff10f74ff5ef60ef85e83dc0d363e0dd959fba625952bc302c3db66f6" Oct 06 11:08:48 crc kubenswrapper[4824]: I1006 11:08:48.156375 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a3e2116ff10f74ff5ef60ef85e83dc0d363e0dd959fba625952bc302c3db66f6"} err="failed to get container status \"a3e2116ff10f74ff5ef60ef85e83dc0d363e0dd959fba625952bc302c3db66f6\": rpc error: code = NotFound desc = could not find container \"a3e2116ff10f74ff5ef60ef85e83dc0d363e0dd959fba625952bc302c3db66f6\": container with ID starting with a3e2116ff10f74ff5ef60ef85e83dc0d363e0dd959fba625952bc302c3db66f6 not found: ID does not exist" Oct 06 11:08:48 crc kubenswrapper[4824]: I1006 11:08:48.156390 4824 scope.go:117] "RemoveContainer" containerID="585fefe35e4e5f30dc9a4a134301ff569dbf2fe2bd57e16995e08e33a29b62fe" Oct 06 11:08:48 crc kubenswrapper[4824]: E1006 11:08:48.156611 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"585fefe35e4e5f30dc9a4a134301ff569dbf2fe2bd57e16995e08e33a29b62fe\": container with ID starting with 585fefe35e4e5f30dc9a4a134301ff569dbf2fe2bd57e16995e08e33a29b62fe not found: ID does not exist" containerID="585fefe35e4e5f30dc9a4a134301ff569dbf2fe2bd57e16995e08e33a29b62fe" Oct 06 11:08:48 crc kubenswrapper[4824]: I1006 11:08:48.156635 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"585fefe35e4e5f30dc9a4a134301ff569dbf2fe2bd57e16995e08e33a29b62fe"} err="failed to get container status \"585fefe35e4e5f30dc9a4a134301ff569dbf2fe2bd57e16995e08e33a29b62fe\": rpc error: code = NotFound desc = could not find container \"585fefe35e4e5f30dc9a4a134301ff569dbf2fe2bd57e16995e08e33a29b62fe\": container with ID starting with 585fefe35e4e5f30dc9a4a134301ff569dbf2fe2bd57e16995e08e33a29b62fe not found: ID does not exist" Oct 06 11:08:49 crc kubenswrapper[4824]: I1006 11:08:49.286066 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c94c3574-b827-4045-869d-c9148e62cb46" path="/var/lib/kubelet/pods/c94c3574-b827-4045-869d-c9148e62cb46/volumes" Oct 06 11:08:49 crc kubenswrapper[4824]: I1006 11:08:49.932438 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-fn7dj_2a76ccaa-9249-46ba-bbc5-dbbbc33b06bb/control-plane-machine-set-operator/0.log" Oct 06 11:08:50 crc kubenswrapper[4824]: I1006 11:08:50.136340 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-fjdwv_b349b0af-0ba8-409f-a915-7fb06b2e3774/kube-rbac-proxy/0.log" Oct 06 11:08:50 crc kubenswrapper[4824]: I1006 11:08:50.186323 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-fjdwv_b349b0af-0ba8-409f-a915-7fb06b2e3774/machine-api-operator/0.log" Oct 06 11:09:05 crc kubenswrapper[4824]: I1006 11:09:05.403221 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-bdn2f_bde1f651-2eda-4d2d-8ab5-5fcdcd4580a4/cert-manager-controller/0.log" Oct 06 11:09:05 crc kubenswrapper[4824]: I1006 11:09:05.595101 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-htx98_d6f420d5-ed6d-4201-b78f-0bfa304e3b6b/cert-manager-webhook/0.log" Oct 06 11:09:05 crc kubenswrapper[4824]: I1006 11:09:05.619213 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-bbqpq_b5759913-9cb4-4012-bffb-fe2f8085543c/cert-manager-cainjector/0.log" Oct 06 11:09:21 crc kubenswrapper[4824]: I1006 11:09:21.192664 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-6b874cbd85-7nwh2_00271786-ab42-4507-a545-8c5bdf1b976d/nmstate-console-plugin/0.log" Oct 06 11:09:21 crc kubenswrapper[4824]: I1006 11:09:21.368191 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-jbv5q_3048d3eb-7f41-4b26-92b3-6d60f5a3fa18/nmstate-handler/0.log" Oct 06 11:09:21 crc kubenswrapper[4824]: I1006 11:09:21.379217 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-6vncz_355b7d6a-3ab4-4d16-8923-f15646b3af54/kube-rbac-proxy/0.log" Oct 06 11:09:21 crc kubenswrapper[4824]: I1006 11:09:21.403837 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-6vncz_355b7d6a-3ab4-4d16-8923-f15646b3af54/nmstate-metrics/0.log" Oct 06 11:09:22 crc kubenswrapper[4824]: I1006 11:09:22.303692 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-858ddd8f98-grzj6_230d69e7-8f29-41a3-af2f-b1b93c58e8c2/nmstate-operator/0.log" Oct 06 11:09:22 crc kubenswrapper[4824]: I1006 11:09:22.333829 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-6cdbc54649-b7dd2_b706331c-d771-4212-950f-6e3fd16faa43/nmstate-webhook/0.log" Oct 06 11:09:38 crc kubenswrapper[4824]: I1006 11:09:38.290665 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-8qff2"] Oct 06 11:09:38 crc kubenswrapper[4824]: E1006 11:09:38.292120 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c94c3574-b827-4045-869d-c9148e62cb46" containerName="extract-content" Oct 06 11:09:38 crc kubenswrapper[4824]: I1006 11:09:38.292134 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="c94c3574-b827-4045-869d-c9148e62cb46" containerName="extract-content" Oct 06 11:09:38 crc kubenswrapper[4824]: E1006 11:09:38.292173 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c94c3574-b827-4045-869d-c9148e62cb46" containerName="extract-utilities" Oct 06 11:09:38 crc kubenswrapper[4824]: I1006 11:09:38.292180 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="c94c3574-b827-4045-869d-c9148e62cb46" containerName="extract-utilities" Oct 06 11:09:38 crc kubenswrapper[4824]: E1006 11:09:38.292194 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c7079f3-84d9-481d-b5fd-090ad21f4678" containerName="container-00" Oct 06 11:09:38 crc kubenswrapper[4824]: I1006 11:09:38.292199 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c7079f3-84d9-481d-b5fd-090ad21f4678" containerName="container-00" Oct 06 11:09:38 crc kubenswrapper[4824]: E1006 11:09:38.292211 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c94c3574-b827-4045-869d-c9148e62cb46" containerName="registry-server" Oct 06 11:09:38 crc kubenswrapper[4824]: I1006 11:09:38.292227 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="c94c3574-b827-4045-869d-c9148e62cb46" containerName="registry-server" Oct 06 11:09:38 crc kubenswrapper[4824]: I1006 11:09:38.292445 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c7079f3-84d9-481d-b5fd-090ad21f4678" containerName="container-00" Oct 06 11:09:38 crc kubenswrapper[4824]: I1006 11:09:38.292475 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="c94c3574-b827-4045-869d-c9148e62cb46" containerName="registry-server" Oct 06 11:09:38 crc kubenswrapper[4824]: I1006 11:09:38.298428 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8qff2" Oct 06 11:09:38 crc kubenswrapper[4824]: I1006 11:09:38.306242 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8qff2"] Oct 06 11:09:38 crc kubenswrapper[4824]: I1006 11:09:38.420779 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a9ba516-e081-4baf-aee0-232275560746-utilities\") pod \"redhat-marketplace-8qff2\" (UID: \"0a9ba516-e081-4baf-aee0-232275560746\") " pod="openshift-marketplace/redhat-marketplace-8qff2" Oct 06 11:09:38 crc kubenswrapper[4824]: I1006 11:09:38.420841 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a9ba516-e081-4baf-aee0-232275560746-catalog-content\") pod \"redhat-marketplace-8qff2\" (UID: \"0a9ba516-e081-4baf-aee0-232275560746\") " pod="openshift-marketplace/redhat-marketplace-8qff2" Oct 06 11:09:38 crc kubenswrapper[4824]: I1006 11:09:38.420911 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pntwr\" (UniqueName: \"kubernetes.io/projected/0a9ba516-e081-4baf-aee0-232275560746-kube-api-access-pntwr\") pod \"redhat-marketplace-8qff2\" (UID: \"0a9ba516-e081-4baf-aee0-232275560746\") " pod="openshift-marketplace/redhat-marketplace-8qff2" Oct 06 11:09:38 crc kubenswrapper[4824]: I1006 11:09:38.522897 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a9ba516-e081-4baf-aee0-232275560746-utilities\") pod \"redhat-marketplace-8qff2\" (UID: \"0a9ba516-e081-4baf-aee0-232275560746\") " pod="openshift-marketplace/redhat-marketplace-8qff2" Oct 06 11:09:38 crc kubenswrapper[4824]: I1006 11:09:38.522968 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a9ba516-e081-4baf-aee0-232275560746-catalog-content\") pod \"redhat-marketplace-8qff2\" (UID: \"0a9ba516-e081-4baf-aee0-232275560746\") " pod="openshift-marketplace/redhat-marketplace-8qff2" Oct 06 11:09:38 crc kubenswrapper[4824]: I1006 11:09:38.523062 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pntwr\" (UniqueName: \"kubernetes.io/projected/0a9ba516-e081-4baf-aee0-232275560746-kube-api-access-pntwr\") pod \"redhat-marketplace-8qff2\" (UID: \"0a9ba516-e081-4baf-aee0-232275560746\") " pod="openshift-marketplace/redhat-marketplace-8qff2" Oct 06 11:09:38 crc kubenswrapper[4824]: I1006 11:09:38.524040 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a9ba516-e081-4baf-aee0-232275560746-utilities\") pod \"redhat-marketplace-8qff2\" (UID: \"0a9ba516-e081-4baf-aee0-232275560746\") " pod="openshift-marketplace/redhat-marketplace-8qff2" Oct 06 11:09:38 crc kubenswrapper[4824]: I1006 11:09:38.524270 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a9ba516-e081-4baf-aee0-232275560746-catalog-content\") pod \"redhat-marketplace-8qff2\" (UID: \"0a9ba516-e081-4baf-aee0-232275560746\") " pod="openshift-marketplace/redhat-marketplace-8qff2" Oct 06 11:09:38 crc kubenswrapper[4824]: I1006 11:09:38.546357 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pntwr\" (UniqueName: \"kubernetes.io/projected/0a9ba516-e081-4baf-aee0-232275560746-kube-api-access-pntwr\") pod \"redhat-marketplace-8qff2\" (UID: \"0a9ba516-e081-4baf-aee0-232275560746\") " pod="openshift-marketplace/redhat-marketplace-8qff2" Oct 06 11:09:38 crc kubenswrapper[4824]: I1006 11:09:38.637971 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8qff2" Oct 06 11:09:39 crc kubenswrapper[4824]: I1006 11:09:39.161828 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8qff2"] Oct 06 11:09:39 crc kubenswrapper[4824]: I1006 11:09:39.655290 4824 generic.go:334] "Generic (PLEG): container finished" podID="0a9ba516-e081-4baf-aee0-232275560746" containerID="5cfade46529993684361ae0a8380b1c9f50a5af30917170cd53d28c0e26de47e" exitCode=0 Oct 06 11:09:39 crc kubenswrapper[4824]: I1006 11:09:39.655419 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8qff2" event={"ID":"0a9ba516-e081-4baf-aee0-232275560746","Type":"ContainerDied","Data":"5cfade46529993684361ae0a8380b1c9f50a5af30917170cd53d28c0e26de47e"} Oct 06 11:09:39 crc kubenswrapper[4824]: I1006 11:09:39.655702 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8qff2" event={"ID":"0a9ba516-e081-4baf-aee0-232275560746","Type":"ContainerStarted","Data":"8e2ecf03fe0f1775cc5b14044a507b17f6410de0f2b81cdb6cd16bd51676f5ce"} Oct 06 11:09:39 crc kubenswrapper[4824]: I1006 11:09:39.658815 4824 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 06 11:09:40 crc kubenswrapper[4824]: I1006 11:09:40.175892 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-sldcd_129eb734-6e0d-4785-b25a-4c68af80d8c6/kube-rbac-proxy/0.log" Oct 06 11:09:40 crc kubenswrapper[4824]: I1006 11:09:40.295895 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-sldcd_129eb734-6e0d-4785-b25a-4c68af80d8c6/controller/0.log" Oct 06 11:09:40 crc kubenswrapper[4824]: I1006 11:09:40.387368 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-64bf5d555-d9kq5_e0836691-2fe5-4e7c-a374-abe71aeb2278/frr-k8s-webhook-server/0.log" Oct 06 11:09:40 crc kubenswrapper[4824]: I1006 11:09:40.532004 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-x6lsw_2f2ec220-4ca6-4b43-9e89-fdae758ab063/cp-frr-files/0.log" Oct 06 11:09:40 crc kubenswrapper[4824]: I1006 11:09:40.669013 4824 generic.go:334] "Generic (PLEG): container finished" podID="0a9ba516-e081-4baf-aee0-232275560746" containerID="b7bdd47b6c820611480651189220a1c77ae43bece7853f4ff37ee64c1b814007" exitCode=0 Oct 06 11:09:40 crc kubenswrapper[4824]: I1006 11:09:40.669070 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8qff2" event={"ID":"0a9ba516-e081-4baf-aee0-232275560746","Type":"ContainerDied","Data":"b7bdd47b6c820611480651189220a1c77ae43bece7853f4ff37ee64c1b814007"} Oct 06 11:09:40 crc kubenswrapper[4824]: I1006 11:09:40.748394 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-x6lsw_2f2ec220-4ca6-4b43-9e89-fdae758ab063/cp-frr-files/0.log" Oct 06 11:09:40 crc kubenswrapper[4824]: I1006 11:09:40.757043 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-x6lsw_2f2ec220-4ca6-4b43-9e89-fdae758ab063/cp-metrics/0.log" Oct 06 11:09:40 crc kubenswrapper[4824]: I1006 11:09:40.802246 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-x6lsw_2f2ec220-4ca6-4b43-9e89-fdae758ab063/cp-reloader/0.log" Oct 06 11:09:40 crc kubenswrapper[4824]: I1006 11:09:40.829212 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-x6lsw_2f2ec220-4ca6-4b43-9e89-fdae758ab063/cp-reloader/0.log" Oct 06 11:09:41 crc kubenswrapper[4824]: I1006 11:09:41.003045 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-x6lsw_2f2ec220-4ca6-4b43-9e89-fdae758ab063/cp-frr-files/0.log" Oct 06 11:09:41 crc kubenswrapper[4824]: I1006 11:09:41.010925 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-x6lsw_2f2ec220-4ca6-4b43-9e89-fdae758ab063/cp-reloader/0.log" Oct 06 11:09:41 crc kubenswrapper[4824]: I1006 11:09:41.059259 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-x6lsw_2f2ec220-4ca6-4b43-9e89-fdae758ab063/cp-metrics/0.log" Oct 06 11:09:41 crc kubenswrapper[4824]: I1006 11:09:41.060023 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-x6lsw_2f2ec220-4ca6-4b43-9e89-fdae758ab063/cp-metrics/0.log" Oct 06 11:09:41 crc kubenswrapper[4824]: I1006 11:09:41.308927 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-x6lsw_2f2ec220-4ca6-4b43-9e89-fdae758ab063/cp-frr-files/0.log" Oct 06 11:09:41 crc kubenswrapper[4824]: I1006 11:09:41.325277 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-x6lsw_2f2ec220-4ca6-4b43-9e89-fdae758ab063/cp-metrics/0.log" Oct 06 11:09:41 crc kubenswrapper[4824]: I1006 11:09:41.325486 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-x6lsw_2f2ec220-4ca6-4b43-9e89-fdae758ab063/cp-reloader/0.log" Oct 06 11:09:41 crc kubenswrapper[4824]: I1006 11:09:41.367960 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-x6lsw_2f2ec220-4ca6-4b43-9e89-fdae758ab063/controller/0.log" Oct 06 11:09:41 crc kubenswrapper[4824]: I1006 11:09:41.552034 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-x6lsw_2f2ec220-4ca6-4b43-9e89-fdae758ab063/frr-metrics/0.log" Oct 06 11:09:41 crc kubenswrapper[4824]: I1006 11:09:41.621534 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-x6lsw_2f2ec220-4ca6-4b43-9e89-fdae758ab063/kube-rbac-proxy-frr/0.log" Oct 06 11:09:41 crc kubenswrapper[4824]: I1006 11:09:41.647532 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-x6lsw_2f2ec220-4ca6-4b43-9e89-fdae758ab063/kube-rbac-proxy/0.log" Oct 06 11:09:41 crc kubenswrapper[4824]: I1006 11:09:41.685218 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8qff2" event={"ID":"0a9ba516-e081-4baf-aee0-232275560746","Type":"ContainerStarted","Data":"75adef83167d6e695d4d508fd140fa3b4dcb685202cfa3731a44898a84c025c7"} Oct 06 11:09:41 crc kubenswrapper[4824]: I1006 11:09:41.714969 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-8qff2" podStartSLOduration=2.337778684 podStartE2EDuration="3.714946923s" podCreationTimestamp="2025-10-06 11:09:38 +0000 UTC" firstStartedPulling="2025-10-06 11:09:39.658528875 +0000 UTC m=+4289.022951726" lastFinishedPulling="2025-10-06 11:09:41.035697104 +0000 UTC m=+4290.400119965" observedRunningTime="2025-10-06 11:09:41.702153564 +0000 UTC m=+4291.066576425" watchObservedRunningTime="2025-10-06 11:09:41.714946923 +0000 UTC m=+4291.079369774" Oct 06 11:09:41 crc kubenswrapper[4824]: I1006 11:09:41.895742 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-x6lsw_2f2ec220-4ca6-4b43-9e89-fdae758ab063/reloader/0.log" Oct 06 11:09:41 crc kubenswrapper[4824]: I1006 11:09:41.966989 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-57744c8fb4-85v7z_d8226542-256f-4e73-9d78-0d8f261108d4/manager/0.log" Oct 06 11:09:42 crc kubenswrapper[4824]: I1006 11:09:42.181846 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-56788496c6-xdgkc_1d2065c1-e8a8-4dfd-bf6e-a8f701cbe060/webhook-server/0.log" Oct 06 11:09:42 crc kubenswrapper[4824]: I1006 11:09:42.442263 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-tcgrw_5c0a8aa4-2b23-4265-8b07-e075b355fc6c/kube-rbac-proxy/0.log" Oct 06 11:09:42 crc kubenswrapper[4824]: I1006 11:09:42.967072 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-tcgrw_5c0a8aa4-2b23-4265-8b07-e075b355fc6c/speaker/0.log" Oct 06 11:09:43 crc kubenswrapper[4824]: I1006 11:09:43.145512 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-x6lsw_2f2ec220-4ca6-4b43-9e89-fdae758ab063/frr/0.log" Oct 06 11:09:43 crc kubenswrapper[4824]: I1006 11:09:43.916319 4824 patch_prober.go:28] interesting pod/machine-config-daemon-khgzw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 11:09:43 crc kubenswrapper[4824]: I1006 11:09:43.916813 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 11:09:48 crc kubenswrapper[4824]: I1006 11:09:48.638692 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-8qff2" Oct 06 11:09:48 crc kubenswrapper[4824]: I1006 11:09:48.639570 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-8qff2" Oct 06 11:09:48 crc kubenswrapper[4824]: I1006 11:09:48.712349 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-8qff2" Oct 06 11:09:48 crc kubenswrapper[4824]: I1006 11:09:48.857732 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-8qff2" Oct 06 11:09:48 crc kubenswrapper[4824]: I1006 11:09:48.980392 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-8qff2"] Oct 06 11:09:50 crc kubenswrapper[4824]: I1006 11:09:50.796183 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-8qff2" podUID="0a9ba516-e081-4baf-aee0-232275560746" containerName="registry-server" containerID="cri-o://75adef83167d6e695d4d508fd140fa3b4dcb685202cfa3731a44898a84c025c7" gracePeriod=2 Oct 06 11:09:51 crc kubenswrapper[4824]: I1006 11:09:51.584041 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8qff2" Oct 06 11:09:51 crc kubenswrapper[4824]: I1006 11:09:51.668927 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a9ba516-e081-4baf-aee0-232275560746-catalog-content\") pod \"0a9ba516-e081-4baf-aee0-232275560746\" (UID: \"0a9ba516-e081-4baf-aee0-232275560746\") " Oct 06 11:09:51 crc kubenswrapper[4824]: I1006 11:09:51.669791 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a9ba516-e081-4baf-aee0-232275560746-utilities\") pod \"0a9ba516-e081-4baf-aee0-232275560746\" (UID: \"0a9ba516-e081-4baf-aee0-232275560746\") " Oct 06 11:09:51 crc kubenswrapper[4824]: I1006 11:09:51.669892 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pntwr\" (UniqueName: \"kubernetes.io/projected/0a9ba516-e081-4baf-aee0-232275560746-kube-api-access-pntwr\") pod \"0a9ba516-e081-4baf-aee0-232275560746\" (UID: \"0a9ba516-e081-4baf-aee0-232275560746\") " Oct 06 11:09:51 crc kubenswrapper[4824]: I1006 11:09:51.671553 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0a9ba516-e081-4baf-aee0-232275560746-utilities" (OuterVolumeSpecName: "utilities") pod "0a9ba516-e081-4baf-aee0-232275560746" (UID: "0a9ba516-e081-4baf-aee0-232275560746"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 11:09:51 crc kubenswrapper[4824]: I1006 11:09:51.685213 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0a9ba516-e081-4baf-aee0-232275560746-kube-api-access-pntwr" (OuterVolumeSpecName: "kube-api-access-pntwr") pod "0a9ba516-e081-4baf-aee0-232275560746" (UID: "0a9ba516-e081-4baf-aee0-232275560746"). InnerVolumeSpecName "kube-api-access-pntwr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 11:09:51 crc kubenswrapper[4824]: I1006 11:09:51.688683 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0a9ba516-e081-4baf-aee0-232275560746-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0a9ba516-e081-4baf-aee0-232275560746" (UID: "0a9ba516-e081-4baf-aee0-232275560746"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 11:09:51 crc kubenswrapper[4824]: I1006 11:09:51.772693 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a9ba516-e081-4baf-aee0-232275560746-utilities\") on node \"crc\" DevicePath \"\"" Oct 06 11:09:51 crc kubenswrapper[4824]: I1006 11:09:51.772742 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pntwr\" (UniqueName: \"kubernetes.io/projected/0a9ba516-e081-4baf-aee0-232275560746-kube-api-access-pntwr\") on node \"crc\" DevicePath \"\"" Oct 06 11:09:51 crc kubenswrapper[4824]: I1006 11:09:51.772756 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a9ba516-e081-4baf-aee0-232275560746-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 06 11:09:51 crc kubenswrapper[4824]: I1006 11:09:51.812531 4824 generic.go:334] "Generic (PLEG): container finished" podID="0a9ba516-e081-4baf-aee0-232275560746" containerID="75adef83167d6e695d4d508fd140fa3b4dcb685202cfa3731a44898a84c025c7" exitCode=0 Oct 06 11:09:51 crc kubenswrapper[4824]: I1006 11:09:51.812599 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8qff2" event={"ID":"0a9ba516-e081-4baf-aee0-232275560746","Type":"ContainerDied","Data":"75adef83167d6e695d4d508fd140fa3b4dcb685202cfa3731a44898a84c025c7"} Oct 06 11:09:51 crc kubenswrapper[4824]: I1006 11:09:51.812638 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8qff2" event={"ID":"0a9ba516-e081-4baf-aee0-232275560746","Type":"ContainerDied","Data":"8e2ecf03fe0f1775cc5b14044a507b17f6410de0f2b81cdb6cd16bd51676f5ce"} Oct 06 11:09:51 crc kubenswrapper[4824]: I1006 11:09:51.812663 4824 scope.go:117] "RemoveContainer" containerID="75adef83167d6e695d4d508fd140fa3b4dcb685202cfa3731a44898a84c025c7" Oct 06 11:09:51 crc kubenswrapper[4824]: I1006 11:09:51.812694 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8qff2" Oct 06 11:09:51 crc kubenswrapper[4824]: I1006 11:09:51.850746 4824 scope.go:117] "RemoveContainer" containerID="b7bdd47b6c820611480651189220a1c77ae43bece7853f4ff37ee64c1b814007" Oct 06 11:09:51 crc kubenswrapper[4824]: I1006 11:09:51.865188 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-8qff2"] Oct 06 11:09:51 crc kubenswrapper[4824]: I1006 11:09:51.883316 4824 scope.go:117] "RemoveContainer" containerID="5cfade46529993684361ae0a8380b1c9f50a5af30917170cd53d28c0e26de47e" Oct 06 11:09:51 crc kubenswrapper[4824]: I1006 11:09:51.883352 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-8qff2"] Oct 06 11:09:51 crc kubenswrapper[4824]: I1006 11:09:51.939563 4824 scope.go:117] "RemoveContainer" containerID="75adef83167d6e695d4d508fd140fa3b4dcb685202cfa3731a44898a84c025c7" Oct 06 11:09:51 crc kubenswrapper[4824]: E1006 11:09:51.940750 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"75adef83167d6e695d4d508fd140fa3b4dcb685202cfa3731a44898a84c025c7\": container with ID starting with 75adef83167d6e695d4d508fd140fa3b4dcb685202cfa3731a44898a84c025c7 not found: ID does not exist" containerID="75adef83167d6e695d4d508fd140fa3b4dcb685202cfa3731a44898a84c025c7" Oct 06 11:09:51 crc kubenswrapper[4824]: I1006 11:09:51.940799 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"75adef83167d6e695d4d508fd140fa3b4dcb685202cfa3731a44898a84c025c7"} err="failed to get container status \"75adef83167d6e695d4d508fd140fa3b4dcb685202cfa3731a44898a84c025c7\": rpc error: code = NotFound desc = could not find container \"75adef83167d6e695d4d508fd140fa3b4dcb685202cfa3731a44898a84c025c7\": container with ID starting with 75adef83167d6e695d4d508fd140fa3b4dcb685202cfa3731a44898a84c025c7 not found: ID does not exist" Oct 06 11:09:51 crc kubenswrapper[4824]: I1006 11:09:51.940824 4824 scope.go:117] "RemoveContainer" containerID="b7bdd47b6c820611480651189220a1c77ae43bece7853f4ff37ee64c1b814007" Oct 06 11:09:51 crc kubenswrapper[4824]: E1006 11:09:51.941153 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b7bdd47b6c820611480651189220a1c77ae43bece7853f4ff37ee64c1b814007\": container with ID starting with b7bdd47b6c820611480651189220a1c77ae43bece7853f4ff37ee64c1b814007 not found: ID does not exist" containerID="b7bdd47b6c820611480651189220a1c77ae43bece7853f4ff37ee64c1b814007" Oct 06 11:09:51 crc kubenswrapper[4824]: I1006 11:09:51.941178 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b7bdd47b6c820611480651189220a1c77ae43bece7853f4ff37ee64c1b814007"} err="failed to get container status \"b7bdd47b6c820611480651189220a1c77ae43bece7853f4ff37ee64c1b814007\": rpc error: code = NotFound desc = could not find container \"b7bdd47b6c820611480651189220a1c77ae43bece7853f4ff37ee64c1b814007\": container with ID starting with b7bdd47b6c820611480651189220a1c77ae43bece7853f4ff37ee64c1b814007 not found: ID does not exist" Oct 06 11:09:51 crc kubenswrapper[4824]: I1006 11:09:51.941198 4824 scope.go:117] "RemoveContainer" containerID="5cfade46529993684361ae0a8380b1c9f50a5af30917170cd53d28c0e26de47e" Oct 06 11:09:51 crc kubenswrapper[4824]: E1006 11:09:51.941562 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5cfade46529993684361ae0a8380b1c9f50a5af30917170cd53d28c0e26de47e\": container with ID starting with 5cfade46529993684361ae0a8380b1c9f50a5af30917170cd53d28c0e26de47e not found: ID does not exist" containerID="5cfade46529993684361ae0a8380b1c9f50a5af30917170cd53d28c0e26de47e" Oct 06 11:09:51 crc kubenswrapper[4824]: I1006 11:09:51.941628 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5cfade46529993684361ae0a8380b1c9f50a5af30917170cd53d28c0e26de47e"} err="failed to get container status \"5cfade46529993684361ae0a8380b1c9f50a5af30917170cd53d28c0e26de47e\": rpc error: code = NotFound desc = could not find container \"5cfade46529993684361ae0a8380b1c9f50a5af30917170cd53d28c0e26de47e\": container with ID starting with 5cfade46529993684361ae0a8380b1c9f50a5af30917170cd53d28c0e26de47e not found: ID does not exist" Oct 06 11:09:53 crc kubenswrapper[4824]: I1006 11:09:53.285802 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0a9ba516-e081-4baf-aee0-232275560746" path="/var/lib/kubelet/pods/0a9ba516-e081-4baf-aee0-232275560746/volumes" Oct 06 11:09:58 crc kubenswrapper[4824]: I1006 11:09:58.744229 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2frg7w_b186e32b-a095-4c20-816c-9e83ea7e261f/util/0.log" Oct 06 11:09:58 crc kubenswrapper[4824]: I1006 11:09:58.983925 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2frg7w_b186e32b-a095-4c20-816c-9e83ea7e261f/pull/0.log" Oct 06 11:09:58 crc kubenswrapper[4824]: I1006 11:09:58.984363 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2frg7w_b186e32b-a095-4c20-816c-9e83ea7e261f/util/0.log" Oct 06 11:09:59 crc kubenswrapper[4824]: I1006 11:09:59.009536 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2frg7w_b186e32b-a095-4c20-816c-9e83ea7e261f/pull/0.log" Oct 06 11:09:59 crc kubenswrapper[4824]: I1006 11:09:59.225502 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2frg7w_b186e32b-a095-4c20-816c-9e83ea7e261f/util/0.log" Oct 06 11:09:59 crc kubenswrapper[4824]: I1006 11:09:59.230892 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2frg7w_b186e32b-a095-4c20-816c-9e83ea7e261f/pull/0.log" Oct 06 11:09:59 crc kubenswrapper[4824]: I1006 11:09:59.257890 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2frg7w_b186e32b-a095-4c20-816c-9e83ea7e261f/extract/0.log" Oct 06 11:09:59 crc kubenswrapper[4824]: I1006 11:09:59.439482 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-mblh7_2dc6af8a-35a7-4fa8-9e03-eb6df841f127/extract-utilities/0.log" Oct 06 11:09:59 crc kubenswrapper[4824]: I1006 11:09:59.623501 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-mblh7_2dc6af8a-35a7-4fa8-9e03-eb6df841f127/extract-content/0.log" Oct 06 11:09:59 crc kubenswrapper[4824]: I1006 11:09:59.624379 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-mblh7_2dc6af8a-35a7-4fa8-9e03-eb6df841f127/extract-content/0.log" Oct 06 11:09:59 crc kubenswrapper[4824]: I1006 11:09:59.642279 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-mblh7_2dc6af8a-35a7-4fa8-9e03-eb6df841f127/extract-utilities/0.log" Oct 06 11:09:59 crc kubenswrapper[4824]: I1006 11:09:59.913697 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-mblh7_2dc6af8a-35a7-4fa8-9e03-eb6df841f127/extract-content/0.log" Oct 06 11:09:59 crc kubenswrapper[4824]: I1006 11:09:59.914411 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-mblh7_2dc6af8a-35a7-4fa8-9e03-eb6df841f127/extract-utilities/0.log" Oct 06 11:10:00 crc kubenswrapper[4824]: I1006 11:10:00.120708 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-s2nbh_4519644a-379b-451a-9dd9-fb528ced61bf/extract-utilities/0.log" Oct 06 11:10:00 crc kubenswrapper[4824]: I1006 11:10:00.471812 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-s2nbh_4519644a-379b-451a-9dd9-fb528ced61bf/extract-content/0.log" Oct 06 11:10:00 crc kubenswrapper[4824]: I1006 11:10:00.510795 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-s2nbh_4519644a-379b-451a-9dd9-fb528ced61bf/extract-content/0.log" Oct 06 11:10:00 crc kubenswrapper[4824]: I1006 11:10:00.529959 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-s2nbh_4519644a-379b-451a-9dd9-fb528ced61bf/extract-utilities/0.log" Oct 06 11:10:00 crc kubenswrapper[4824]: I1006 11:10:00.584662 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-mblh7_2dc6af8a-35a7-4fa8-9e03-eb6df841f127/registry-server/0.log" Oct 06 11:10:00 crc kubenswrapper[4824]: I1006 11:10:00.734084 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-s2nbh_4519644a-379b-451a-9dd9-fb528ced61bf/extract-utilities/0.log" Oct 06 11:10:00 crc kubenswrapper[4824]: I1006 11:10:00.752287 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-s2nbh_4519644a-379b-451a-9dd9-fb528ced61bf/extract-content/0.log" Oct 06 11:10:01 crc kubenswrapper[4824]: I1006 11:10:01.009401 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cxdt8w_5e47aa0e-c660-4f88-8fb1-8c873268879d/util/0.log" Oct 06 11:10:01 crc kubenswrapper[4824]: I1006 11:10:01.331125 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cxdt8w_5e47aa0e-c660-4f88-8fb1-8c873268879d/pull/0.log" Oct 06 11:10:01 crc kubenswrapper[4824]: I1006 11:10:01.341493 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cxdt8w_5e47aa0e-c660-4f88-8fb1-8c873268879d/util/0.log" Oct 06 11:10:01 crc kubenswrapper[4824]: I1006 11:10:01.362842 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cxdt8w_5e47aa0e-c660-4f88-8fb1-8c873268879d/pull/0.log" Oct 06 11:10:01 crc kubenswrapper[4824]: I1006 11:10:01.677315 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cxdt8w_5e47aa0e-c660-4f88-8fb1-8c873268879d/extract/0.log" Oct 06 11:10:01 crc kubenswrapper[4824]: I1006 11:10:01.680364 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cxdt8w_5e47aa0e-c660-4f88-8fb1-8c873268879d/util/0.log" Oct 06 11:10:01 crc kubenswrapper[4824]: I1006 11:10:01.789298 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cxdt8w_5e47aa0e-c660-4f88-8fb1-8c873268879d/pull/0.log" Oct 06 11:10:01 crc kubenswrapper[4824]: I1006 11:10:01.995062 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-s2nbh_4519644a-379b-451a-9dd9-fb528ced61bf/registry-server/0.log" Oct 06 11:10:02 crc kubenswrapper[4824]: I1006 11:10:02.683065 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-db6vn_a3a514b8-f221-465e-bbb2-3c88bcb4f622/marketplace-operator/0.log" Oct 06 11:10:02 crc kubenswrapper[4824]: I1006 11:10:02.724029 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-xdszq_504bf3d4-9589-4a5e-ad25-e16f783e0115/extract-utilities/0.log" Oct 06 11:10:02 crc kubenswrapper[4824]: I1006 11:10:02.919412 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-xdszq_504bf3d4-9589-4a5e-ad25-e16f783e0115/extract-content/0.log" Oct 06 11:10:02 crc kubenswrapper[4824]: I1006 11:10:02.946007 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-xdszq_504bf3d4-9589-4a5e-ad25-e16f783e0115/extract-utilities/0.log" Oct 06 11:10:02 crc kubenswrapper[4824]: I1006 11:10:02.987499 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-xdszq_504bf3d4-9589-4a5e-ad25-e16f783e0115/extract-content/0.log" Oct 06 11:10:03 crc kubenswrapper[4824]: I1006 11:10:03.207531 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-d8s5f_bcb79b86-65c0-44c1-9ce2-588ad66206a6/extract-utilities/0.log" Oct 06 11:10:03 crc kubenswrapper[4824]: I1006 11:10:03.221706 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-xdszq_504bf3d4-9589-4a5e-ad25-e16f783e0115/extract-utilities/0.log" Oct 06 11:10:03 crc kubenswrapper[4824]: I1006 11:10:03.235504 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-xdszq_504bf3d4-9589-4a5e-ad25-e16f783e0115/extract-content/0.log" Oct 06 11:10:03 crc kubenswrapper[4824]: I1006 11:10:03.344715 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-xdszq_504bf3d4-9589-4a5e-ad25-e16f783e0115/registry-server/0.log" Oct 06 11:10:03 crc kubenswrapper[4824]: I1006 11:10:03.487586 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-d8s5f_bcb79b86-65c0-44c1-9ce2-588ad66206a6/extract-content/0.log" Oct 06 11:10:03 crc kubenswrapper[4824]: I1006 11:10:03.489888 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-d8s5f_bcb79b86-65c0-44c1-9ce2-588ad66206a6/extract-utilities/0.log" Oct 06 11:10:03 crc kubenswrapper[4824]: I1006 11:10:03.494604 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-d8s5f_bcb79b86-65c0-44c1-9ce2-588ad66206a6/extract-content/0.log" Oct 06 11:10:03 crc kubenswrapper[4824]: I1006 11:10:03.658658 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-d8s5f_bcb79b86-65c0-44c1-9ce2-588ad66206a6/extract-utilities/0.log" Oct 06 11:10:03 crc kubenswrapper[4824]: I1006 11:10:03.709531 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-d8s5f_bcb79b86-65c0-44c1-9ce2-588ad66206a6/extract-content/0.log" Oct 06 11:10:03 crc kubenswrapper[4824]: I1006 11:10:03.962261 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-d8s5f_bcb79b86-65c0-44c1-9ce2-588ad66206a6/registry-server/0.log" Oct 06 11:10:13 crc kubenswrapper[4824]: I1006 11:10:13.914869 4824 patch_prober.go:28] interesting pod/machine-config-daemon-khgzw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 11:10:13 crc kubenswrapper[4824]: I1006 11:10:13.917300 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 11:10:36 crc kubenswrapper[4824]: E1006 11:10:36.390014 4824 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.50:41774->38.102.83.50:33571: write tcp 38.102.83.50:41774->38.102.83.50:33571: write: broken pipe Oct 06 11:10:43 crc kubenswrapper[4824]: I1006 11:10:43.914992 4824 patch_prober.go:28] interesting pod/machine-config-daemon-khgzw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 11:10:43 crc kubenswrapper[4824]: I1006 11:10:43.915737 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 11:10:43 crc kubenswrapper[4824]: I1006 11:10:43.915799 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" Oct 06 11:10:43 crc kubenswrapper[4824]: I1006 11:10:43.916671 4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"4fa20ecee3b1c9e1a0579597134d4ac9934abd73fbbea1c39577d842a5d800c4"} pod="openshift-machine-config-operator/machine-config-daemon-khgzw" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 06 11:10:43 crc kubenswrapper[4824]: I1006 11:10:43.916725 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" containerName="machine-config-daemon" containerID="cri-o://4fa20ecee3b1c9e1a0579597134d4ac9934abd73fbbea1c39577d842a5d800c4" gracePeriod=600 Oct 06 11:10:44 crc kubenswrapper[4824]: I1006 11:10:44.436043 4824 generic.go:334] "Generic (PLEG): container finished" podID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" containerID="4fa20ecee3b1c9e1a0579597134d4ac9934abd73fbbea1c39577d842a5d800c4" exitCode=0 Oct 06 11:10:44 crc kubenswrapper[4824]: I1006 11:10:44.436052 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" event={"ID":"e1140c8c-93f1-4fce-8c06-a69561ff8a6d","Type":"ContainerDied","Data":"4fa20ecee3b1c9e1a0579597134d4ac9934abd73fbbea1c39577d842a5d800c4"} Oct 06 11:10:44 crc kubenswrapper[4824]: I1006 11:10:44.436760 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" event={"ID":"e1140c8c-93f1-4fce-8c06-a69561ff8a6d","Type":"ContainerStarted","Data":"ab600e16b7115731e5996d1f6491119cc9b18fcff9ff9742d61072463acfb2f4"} Oct 06 11:10:44 crc kubenswrapper[4824]: I1006 11:10:44.436829 4824 scope.go:117] "RemoveContainer" containerID="5a47a5d2e5b1ba0ad6ad09ca0bd7f5568851664618ce5ad15f9b3204fecd76c5" Oct 06 11:12:20 crc kubenswrapper[4824]: I1006 11:12:20.785054 4824 generic.go:334] "Generic (PLEG): container finished" podID="215303b4-ee21-44b7-a2a3-cd3ff4c6265e" containerID="551a46d712b0750162e9fdff376ae3c1fdecf89e52dd8b5b60fe053ec4f4b731" exitCode=0 Oct 06 11:12:20 crc kubenswrapper[4824]: I1006 11:12:20.785168 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-x5ksk/must-gather-zn6wq" event={"ID":"215303b4-ee21-44b7-a2a3-cd3ff4c6265e","Type":"ContainerDied","Data":"551a46d712b0750162e9fdff376ae3c1fdecf89e52dd8b5b60fe053ec4f4b731"} Oct 06 11:12:20 crc kubenswrapper[4824]: I1006 11:12:20.787342 4824 scope.go:117] "RemoveContainer" containerID="551a46d712b0750162e9fdff376ae3c1fdecf89e52dd8b5b60fe053ec4f4b731" Oct 06 11:12:21 crc kubenswrapper[4824]: I1006 11:12:21.513236 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-x5ksk_must-gather-zn6wq_215303b4-ee21-44b7-a2a3-cd3ff4c6265e/gather/0.log" Oct 06 11:12:30 crc kubenswrapper[4824]: I1006 11:12:30.805248 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-k65bj"] Oct 06 11:12:30 crc kubenswrapper[4824]: E1006 11:12:30.807033 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a9ba516-e081-4baf-aee0-232275560746" containerName="extract-utilities" Oct 06 11:12:30 crc kubenswrapper[4824]: I1006 11:12:30.807062 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a9ba516-e081-4baf-aee0-232275560746" containerName="extract-utilities" Oct 06 11:12:30 crc kubenswrapper[4824]: E1006 11:12:30.807091 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a9ba516-e081-4baf-aee0-232275560746" containerName="registry-server" Oct 06 11:12:30 crc kubenswrapper[4824]: I1006 11:12:30.807107 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a9ba516-e081-4baf-aee0-232275560746" containerName="registry-server" Oct 06 11:12:30 crc kubenswrapper[4824]: E1006 11:12:30.807135 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a9ba516-e081-4baf-aee0-232275560746" containerName="extract-content" Oct 06 11:12:30 crc kubenswrapper[4824]: I1006 11:12:30.807147 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a9ba516-e081-4baf-aee0-232275560746" containerName="extract-content" Oct 06 11:12:30 crc kubenswrapper[4824]: I1006 11:12:30.807552 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a9ba516-e081-4baf-aee0-232275560746" containerName="registry-server" Oct 06 11:12:30 crc kubenswrapper[4824]: I1006 11:12:30.811294 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-k65bj" Oct 06 11:12:30 crc kubenswrapper[4824]: I1006 11:12:30.868439 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-k65bj"] Oct 06 11:12:30 crc kubenswrapper[4824]: I1006 11:12:30.929245 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/baee1ed3-6bd7-4f45-a914-ae37c34aa28f-catalog-content\") pod \"certified-operators-k65bj\" (UID: \"baee1ed3-6bd7-4f45-a914-ae37c34aa28f\") " pod="openshift-marketplace/certified-operators-k65bj" Oct 06 11:12:30 crc kubenswrapper[4824]: I1006 11:12:30.929310 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gdcgh\" (UniqueName: \"kubernetes.io/projected/baee1ed3-6bd7-4f45-a914-ae37c34aa28f-kube-api-access-gdcgh\") pod \"certified-operators-k65bj\" (UID: \"baee1ed3-6bd7-4f45-a914-ae37c34aa28f\") " pod="openshift-marketplace/certified-operators-k65bj" Oct 06 11:12:30 crc kubenswrapper[4824]: I1006 11:12:30.929347 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/baee1ed3-6bd7-4f45-a914-ae37c34aa28f-utilities\") pod \"certified-operators-k65bj\" (UID: \"baee1ed3-6bd7-4f45-a914-ae37c34aa28f\") " pod="openshift-marketplace/certified-operators-k65bj" Oct 06 11:12:31 crc kubenswrapper[4824]: I1006 11:12:31.031771 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/baee1ed3-6bd7-4f45-a914-ae37c34aa28f-catalog-content\") pod \"certified-operators-k65bj\" (UID: \"baee1ed3-6bd7-4f45-a914-ae37c34aa28f\") " pod="openshift-marketplace/certified-operators-k65bj" Oct 06 11:12:31 crc kubenswrapper[4824]: I1006 11:12:31.031847 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gdcgh\" (UniqueName: \"kubernetes.io/projected/baee1ed3-6bd7-4f45-a914-ae37c34aa28f-kube-api-access-gdcgh\") pod \"certified-operators-k65bj\" (UID: \"baee1ed3-6bd7-4f45-a914-ae37c34aa28f\") " pod="openshift-marketplace/certified-operators-k65bj" Oct 06 11:12:31 crc kubenswrapper[4824]: I1006 11:12:31.031904 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/baee1ed3-6bd7-4f45-a914-ae37c34aa28f-utilities\") pod \"certified-operators-k65bj\" (UID: \"baee1ed3-6bd7-4f45-a914-ae37c34aa28f\") " pod="openshift-marketplace/certified-operators-k65bj" Oct 06 11:12:31 crc kubenswrapper[4824]: I1006 11:12:31.032425 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/baee1ed3-6bd7-4f45-a914-ae37c34aa28f-catalog-content\") pod \"certified-operators-k65bj\" (UID: \"baee1ed3-6bd7-4f45-a914-ae37c34aa28f\") " pod="openshift-marketplace/certified-operators-k65bj" Oct 06 11:12:31 crc kubenswrapper[4824]: I1006 11:12:31.032707 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/baee1ed3-6bd7-4f45-a914-ae37c34aa28f-utilities\") pod \"certified-operators-k65bj\" (UID: \"baee1ed3-6bd7-4f45-a914-ae37c34aa28f\") " pod="openshift-marketplace/certified-operators-k65bj" Oct 06 11:12:31 crc kubenswrapper[4824]: I1006 11:12:31.059262 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gdcgh\" (UniqueName: \"kubernetes.io/projected/baee1ed3-6bd7-4f45-a914-ae37c34aa28f-kube-api-access-gdcgh\") pod \"certified-operators-k65bj\" (UID: \"baee1ed3-6bd7-4f45-a914-ae37c34aa28f\") " pod="openshift-marketplace/certified-operators-k65bj" Oct 06 11:12:31 crc kubenswrapper[4824]: I1006 11:12:31.144131 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-k65bj" Oct 06 11:12:31 crc kubenswrapper[4824]: I1006 11:12:31.726797 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-k65bj"] Oct 06 11:12:32 crc kubenswrapper[4824]: W1006 11:12:32.277664 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbaee1ed3_6bd7_4f45_a914_ae37c34aa28f.slice/crio-422b839ae9c1d4aac6fc2c885e13d7ae9c636c3f2ddb797b52752c6e81ea1115 WatchSource:0}: Error finding container 422b839ae9c1d4aac6fc2c885e13d7ae9c636c3f2ddb797b52752c6e81ea1115: Status 404 returned error can't find the container with id 422b839ae9c1d4aac6fc2c885e13d7ae9c636c3f2ddb797b52752c6e81ea1115 Oct 06 11:12:32 crc kubenswrapper[4824]: I1006 11:12:32.984086 4824 generic.go:334] "Generic (PLEG): container finished" podID="baee1ed3-6bd7-4f45-a914-ae37c34aa28f" containerID="177215f8f914616f92da47ce3835b96873552cb9ad949ac47b53df6e49d473d2" exitCode=0 Oct 06 11:12:32 crc kubenswrapper[4824]: I1006 11:12:32.984157 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k65bj" event={"ID":"baee1ed3-6bd7-4f45-a914-ae37c34aa28f","Type":"ContainerDied","Data":"177215f8f914616f92da47ce3835b96873552cb9ad949ac47b53df6e49d473d2"} Oct 06 11:12:32 crc kubenswrapper[4824]: I1006 11:12:32.985300 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k65bj" event={"ID":"baee1ed3-6bd7-4f45-a914-ae37c34aa28f","Type":"ContainerStarted","Data":"422b839ae9c1d4aac6fc2c885e13d7ae9c636c3f2ddb797b52752c6e81ea1115"} Oct 06 11:12:34 crc kubenswrapper[4824]: I1006 11:12:34.000649 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k65bj" event={"ID":"baee1ed3-6bd7-4f45-a914-ae37c34aa28f","Type":"ContainerStarted","Data":"01cd6db9770dee99fd6b5c658673ee9e2a4f715240b77e21aa52fd81d7ff8350"} Oct 06 11:12:34 crc kubenswrapper[4824]: I1006 11:12:34.281717 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-x5ksk/must-gather-zn6wq"] Oct 06 11:12:34 crc kubenswrapper[4824]: I1006 11:12:34.281917 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-x5ksk/must-gather-zn6wq" podUID="215303b4-ee21-44b7-a2a3-cd3ff4c6265e" containerName="copy" containerID="cri-o://018ec10e14724ca059b1c230ec932142c368814b8f1315ec80e308d2f7d9a3bf" gracePeriod=2 Oct 06 11:12:34 crc kubenswrapper[4824]: I1006 11:12:34.295310 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-x5ksk/must-gather-zn6wq"] Oct 06 11:12:34 crc kubenswrapper[4824]: I1006 11:12:34.849794 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-x5ksk_must-gather-zn6wq_215303b4-ee21-44b7-a2a3-cd3ff4c6265e/copy/0.log" Oct 06 11:12:34 crc kubenswrapper[4824]: I1006 11:12:34.850843 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-x5ksk/must-gather-zn6wq" Oct 06 11:12:34 crc kubenswrapper[4824]: I1006 11:12:34.932926 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-64g8x\" (UniqueName: \"kubernetes.io/projected/215303b4-ee21-44b7-a2a3-cd3ff4c6265e-kube-api-access-64g8x\") pod \"215303b4-ee21-44b7-a2a3-cd3ff4c6265e\" (UID: \"215303b4-ee21-44b7-a2a3-cd3ff4c6265e\") " Oct 06 11:12:34 crc kubenswrapper[4824]: I1006 11:12:34.933020 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/215303b4-ee21-44b7-a2a3-cd3ff4c6265e-must-gather-output\") pod \"215303b4-ee21-44b7-a2a3-cd3ff4c6265e\" (UID: \"215303b4-ee21-44b7-a2a3-cd3ff4c6265e\") " Oct 06 11:12:34 crc kubenswrapper[4824]: I1006 11:12:34.941520 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/215303b4-ee21-44b7-a2a3-cd3ff4c6265e-kube-api-access-64g8x" (OuterVolumeSpecName: "kube-api-access-64g8x") pod "215303b4-ee21-44b7-a2a3-cd3ff4c6265e" (UID: "215303b4-ee21-44b7-a2a3-cd3ff4c6265e"). InnerVolumeSpecName "kube-api-access-64g8x". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 11:12:35 crc kubenswrapper[4824]: I1006 11:12:35.016882 4824 generic.go:334] "Generic (PLEG): container finished" podID="baee1ed3-6bd7-4f45-a914-ae37c34aa28f" containerID="01cd6db9770dee99fd6b5c658673ee9e2a4f715240b77e21aa52fd81d7ff8350" exitCode=0 Oct 06 11:12:35 crc kubenswrapper[4824]: I1006 11:12:35.017155 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k65bj" event={"ID":"baee1ed3-6bd7-4f45-a914-ae37c34aa28f","Type":"ContainerDied","Data":"01cd6db9770dee99fd6b5c658673ee9e2a4f715240b77e21aa52fd81d7ff8350"} Oct 06 11:12:35 crc kubenswrapper[4824]: I1006 11:12:35.021588 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-x5ksk_must-gather-zn6wq_215303b4-ee21-44b7-a2a3-cd3ff4c6265e/copy/0.log" Oct 06 11:12:35 crc kubenswrapper[4824]: I1006 11:12:35.022639 4824 generic.go:334] "Generic (PLEG): container finished" podID="215303b4-ee21-44b7-a2a3-cd3ff4c6265e" containerID="018ec10e14724ca059b1c230ec932142c368814b8f1315ec80e308d2f7d9a3bf" exitCode=143 Oct 06 11:12:35 crc kubenswrapper[4824]: I1006 11:12:35.022723 4824 scope.go:117] "RemoveContainer" containerID="018ec10e14724ca059b1c230ec932142c368814b8f1315ec80e308d2f7d9a3bf" Oct 06 11:12:35 crc kubenswrapper[4824]: I1006 11:12:35.022750 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-x5ksk/must-gather-zn6wq" Oct 06 11:12:35 crc kubenswrapper[4824]: I1006 11:12:35.036292 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-64g8x\" (UniqueName: \"kubernetes.io/projected/215303b4-ee21-44b7-a2a3-cd3ff4c6265e-kube-api-access-64g8x\") on node \"crc\" DevicePath \"\"" Oct 06 11:12:35 crc kubenswrapper[4824]: I1006 11:12:35.068949 4824 scope.go:117] "RemoveContainer" containerID="551a46d712b0750162e9fdff376ae3c1fdecf89e52dd8b5b60fe053ec4f4b731" Oct 06 11:12:35 crc kubenswrapper[4824]: I1006 11:12:35.121020 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/215303b4-ee21-44b7-a2a3-cd3ff4c6265e-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "215303b4-ee21-44b7-a2a3-cd3ff4c6265e" (UID: "215303b4-ee21-44b7-a2a3-cd3ff4c6265e"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 11:12:35 crc kubenswrapper[4824]: I1006 11:12:35.138572 4824 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/215303b4-ee21-44b7-a2a3-cd3ff4c6265e-must-gather-output\") on node \"crc\" DevicePath \"\"" Oct 06 11:12:35 crc kubenswrapper[4824]: I1006 11:12:35.139874 4824 scope.go:117] "RemoveContainer" containerID="018ec10e14724ca059b1c230ec932142c368814b8f1315ec80e308d2f7d9a3bf" Oct 06 11:12:35 crc kubenswrapper[4824]: E1006 11:12:35.140790 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"018ec10e14724ca059b1c230ec932142c368814b8f1315ec80e308d2f7d9a3bf\": container with ID starting with 018ec10e14724ca059b1c230ec932142c368814b8f1315ec80e308d2f7d9a3bf not found: ID does not exist" containerID="018ec10e14724ca059b1c230ec932142c368814b8f1315ec80e308d2f7d9a3bf" Oct 06 11:12:35 crc kubenswrapper[4824]: I1006 11:12:35.140859 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"018ec10e14724ca059b1c230ec932142c368814b8f1315ec80e308d2f7d9a3bf"} err="failed to get container status \"018ec10e14724ca059b1c230ec932142c368814b8f1315ec80e308d2f7d9a3bf\": rpc error: code = NotFound desc = could not find container \"018ec10e14724ca059b1c230ec932142c368814b8f1315ec80e308d2f7d9a3bf\": container with ID starting with 018ec10e14724ca059b1c230ec932142c368814b8f1315ec80e308d2f7d9a3bf not found: ID does not exist" Oct 06 11:12:35 crc kubenswrapper[4824]: I1006 11:12:35.140894 4824 scope.go:117] "RemoveContainer" containerID="551a46d712b0750162e9fdff376ae3c1fdecf89e52dd8b5b60fe053ec4f4b731" Oct 06 11:12:35 crc kubenswrapper[4824]: E1006 11:12:35.141562 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"551a46d712b0750162e9fdff376ae3c1fdecf89e52dd8b5b60fe053ec4f4b731\": container with ID starting with 551a46d712b0750162e9fdff376ae3c1fdecf89e52dd8b5b60fe053ec4f4b731 not found: ID does not exist" containerID="551a46d712b0750162e9fdff376ae3c1fdecf89e52dd8b5b60fe053ec4f4b731" Oct 06 11:12:35 crc kubenswrapper[4824]: I1006 11:12:35.141673 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"551a46d712b0750162e9fdff376ae3c1fdecf89e52dd8b5b60fe053ec4f4b731"} err="failed to get container status \"551a46d712b0750162e9fdff376ae3c1fdecf89e52dd8b5b60fe053ec4f4b731\": rpc error: code = NotFound desc = could not find container \"551a46d712b0750162e9fdff376ae3c1fdecf89e52dd8b5b60fe053ec4f4b731\": container with ID starting with 551a46d712b0750162e9fdff376ae3c1fdecf89e52dd8b5b60fe053ec4f4b731 not found: ID does not exist" Oct 06 11:12:35 crc kubenswrapper[4824]: I1006 11:12:35.291453 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="215303b4-ee21-44b7-a2a3-cd3ff4c6265e" path="/var/lib/kubelet/pods/215303b4-ee21-44b7-a2a3-cd3ff4c6265e/volumes" Oct 06 11:12:36 crc kubenswrapper[4824]: I1006 11:12:36.066170 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k65bj" event={"ID":"baee1ed3-6bd7-4f45-a914-ae37c34aa28f","Type":"ContainerStarted","Data":"9c73b8f5b91a36876a5ffd0aa4846ab89da0e59432ac3dc62a04ed1e42c834aa"} Oct 06 11:12:36 crc kubenswrapper[4824]: I1006 11:12:36.130069 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-k65bj" podStartSLOduration=3.638827091 podStartE2EDuration="6.130046373s" podCreationTimestamp="2025-10-06 11:12:30 +0000 UTC" firstStartedPulling="2025-10-06 11:12:32.986918053 +0000 UTC m=+4462.351340914" lastFinishedPulling="2025-10-06 11:12:35.478137325 +0000 UTC m=+4464.842560196" observedRunningTime="2025-10-06 11:12:36.125703815 +0000 UTC m=+4465.490126676" watchObservedRunningTime="2025-10-06 11:12:36.130046373 +0000 UTC m=+4465.494469234" Oct 06 11:12:41 crc kubenswrapper[4824]: I1006 11:12:41.145285 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-k65bj" Oct 06 11:12:41 crc kubenswrapper[4824]: I1006 11:12:41.147290 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-k65bj" Oct 06 11:12:41 crc kubenswrapper[4824]: I1006 11:12:41.253442 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-k65bj" Oct 06 11:12:42 crc kubenswrapper[4824]: I1006 11:12:42.208333 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-k65bj" Oct 06 11:12:42 crc kubenswrapper[4824]: I1006 11:12:42.286641 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-k65bj"] Oct 06 11:12:44 crc kubenswrapper[4824]: I1006 11:12:44.180243 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-k65bj" podUID="baee1ed3-6bd7-4f45-a914-ae37c34aa28f" containerName="registry-server" containerID="cri-o://9c73b8f5b91a36876a5ffd0aa4846ab89da0e59432ac3dc62a04ed1e42c834aa" gracePeriod=2 Oct 06 11:12:45 crc kubenswrapper[4824]: I1006 11:12:45.196026 4824 generic.go:334] "Generic (PLEG): container finished" podID="baee1ed3-6bd7-4f45-a914-ae37c34aa28f" containerID="9c73b8f5b91a36876a5ffd0aa4846ab89da0e59432ac3dc62a04ed1e42c834aa" exitCode=0 Oct 06 11:12:45 crc kubenswrapper[4824]: I1006 11:12:45.196166 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k65bj" event={"ID":"baee1ed3-6bd7-4f45-a914-ae37c34aa28f","Type":"ContainerDied","Data":"9c73b8f5b91a36876a5ffd0aa4846ab89da0e59432ac3dc62a04ed1e42c834aa"} Oct 06 11:12:45 crc kubenswrapper[4824]: I1006 11:12:45.537338 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-k65bj" Oct 06 11:12:45 crc kubenswrapper[4824]: I1006 11:12:45.718056 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/baee1ed3-6bd7-4f45-a914-ae37c34aa28f-catalog-content\") pod \"baee1ed3-6bd7-4f45-a914-ae37c34aa28f\" (UID: \"baee1ed3-6bd7-4f45-a914-ae37c34aa28f\") " Oct 06 11:12:45 crc kubenswrapper[4824]: I1006 11:12:45.718217 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/baee1ed3-6bd7-4f45-a914-ae37c34aa28f-utilities\") pod \"baee1ed3-6bd7-4f45-a914-ae37c34aa28f\" (UID: \"baee1ed3-6bd7-4f45-a914-ae37c34aa28f\") " Oct 06 11:12:45 crc kubenswrapper[4824]: I1006 11:12:45.718305 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gdcgh\" (UniqueName: \"kubernetes.io/projected/baee1ed3-6bd7-4f45-a914-ae37c34aa28f-kube-api-access-gdcgh\") pod \"baee1ed3-6bd7-4f45-a914-ae37c34aa28f\" (UID: \"baee1ed3-6bd7-4f45-a914-ae37c34aa28f\") " Oct 06 11:12:45 crc kubenswrapper[4824]: I1006 11:12:45.719718 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/baee1ed3-6bd7-4f45-a914-ae37c34aa28f-utilities" (OuterVolumeSpecName: "utilities") pod "baee1ed3-6bd7-4f45-a914-ae37c34aa28f" (UID: "baee1ed3-6bd7-4f45-a914-ae37c34aa28f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 11:12:45 crc kubenswrapper[4824]: I1006 11:12:45.735385 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/baee1ed3-6bd7-4f45-a914-ae37c34aa28f-kube-api-access-gdcgh" (OuterVolumeSpecName: "kube-api-access-gdcgh") pod "baee1ed3-6bd7-4f45-a914-ae37c34aa28f" (UID: "baee1ed3-6bd7-4f45-a914-ae37c34aa28f"). InnerVolumeSpecName "kube-api-access-gdcgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 11:12:45 crc kubenswrapper[4824]: I1006 11:12:45.794418 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/baee1ed3-6bd7-4f45-a914-ae37c34aa28f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "baee1ed3-6bd7-4f45-a914-ae37c34aa28f" (UID: "baee1ed3-6bd7-4f45-a914-ae37c34aa28f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 11:12:45 crc kubenswrapper[4824]: I1006 11:12:45.822198 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gdcgh\" (UniqueName: \"kubernetes.io/projected/baee1ed3-6bd7-4f45-a914-ae37c34aa28f-kube-api-access-gdcgh\") on node \"crc\" DevicePath \"\"" Oct 06 11:12:45 crc kubenswrapper[4824]: I1006 11:12:45.822257 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/baee1ed3-6bd7-4f45-a914-ae37c34aa28f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 06 11:12:45 crc kubenswrapper[4824]: I1006 11:12:45.822271 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/baee1ed3-6bd7-4f45-a914-ae37c34aa28f-utilities\") on node \"crc\" DevicePath \"\"" Oct 06 11:12:46 crc kubenswrapper[4824]: I1006 11:12:46.210493 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k65bj" event={"ID":"baee1ed3-6bd7-4f45-a914-ae37c34aa28f","Type":"ContainerDied","Data":"422b839ae9c1d4aac6fc2c885e13d7ae9c636c3f2ddb797b52752c6e81ea1115"} Oct 06 11:12:46 crc kubenswrapper[4824]: I1006 11:12:46.210570 4824 scope.go:117] "RemoveContainer" containerID="9c73b8f5b91a36876a5ffd0aa4846ab89da0e59432ac3dc62a04ed1e42c834aa" Oct 06 11:12:46 crc kubenswrapper[4824]: I1006 11:12:46.210581 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-k65bj" Oct 06 11:12:46 crc kubenswrapper[4824]: I1006 11:12:46.239730 4824 scope.go:117] "RemoveContainer" containerID="01cd6db9770dee99fd6b5c658673ee9e2a4f715240b77e21aa52fd81d7ff8350" Oct 06 11:12:46 crc kubenswrapper[4824]: I1006 11:12:46.252835 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-k65bj"] Oct 06 11:12:46 crc kubenswrapper[4824]: I1006 11:12:46.269518 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-k65bj"] Oct 06 11:12:46 crc kubenswrapper[4824]: I1006 11:12:46.272499 4824 scope.go:117] "RemoveContainer" containerID="177215f8f914616f92da47ce3835b96873552cb9ad949ac47b53df6e49d473d2" Oct 06 11:12:46 crc kubenswrapper[4824]: E1006 11:12:46.472632 4824 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbaee1ed3_6bd7_4f45_a914_ae37c34aa28f.slice\": RecentStats: unable to find data in memory cache]" Oct 06 11:12:47 crc kubenswrapper[4824]: I1006 11:12:47.292479 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="baee1ed3-6bd7-4f45-a914-ae37c34aa28f" path="/var/lib/kubelet/pods/baee1ed3-6bd7-4f45-a914-ae37c34aa28f/volumes" Oct 06 11:13:13 crc kubenswrapper[4824]: I1006 11:13:13.915154 4824 patch_prober.go:28] interesting pod/machine-config-daemon-khgzw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 11:13:13 crc kubenswrapper[4824]: I1006 11:13:13.917409 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 11:13:43 crc kubenswrapper[4824]: I1006 11:13:43.915412 4824 patch_prober.go:28] interesting pod/machine-config-daemon-khgzw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 11:13:43 crc kubenswrapper[4824]: I1006 11:13:43.916462 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 11:14:13 crc kubenswrapper[4824]: I1006 11:14:13.915884 4824 patch_prober.go:28] interesting pod/machine-config-daemon-khgzw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 11:14:13 crc kubenswrapper[4824]: I1006 11:14:13.916795 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 11:14:13 crc kubenswrapper[4824]: I1006 11:14:13.916871 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" Oct 06 11:14:13 crc kubenswrapper[4824]: I1006 11:14:13.917794 4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ab600e16b7115731e5996d1f6491119cc9b18fcff9ff9742d61072463acfb2f4"} pod="openshift-machine-config-operator/machine-config-daemon-khgzw" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 06 11:14:13 crc kubenswrapper[4824]: I1006 11:14:13.917880 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" containerName="machine-config-daemon" containerID="cri-o://ab600e16b7115731e5996d1f6491119cc9b18fcff9ff9742d61072463acfb2f4" gracePeriod=600 Oct 06 11:14:14 crc kubenswrapper[4824]: E1006 11:14:14.077570 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-khgzw_openshift-machine-config-operator(e1140c8c-93f1-4fce-8c06-a69561ff8a6d)\"" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" Oct 06 11:14:14 crc kubenswrapper[4824]: I1006 11:14:14.443444 4824 generic.go:334] "Generic (PLEG): container finished" podID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" containerID="ab600e16b7115731e5996d1f6491119cc9b18fcff9ff9742d61072463acfb2f4" exitCode=0 Oct 06 11:14:14 crc kubenswrapper[4824]: I1006 11:14:14.443530 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" event={"ID":"e1140c8c-93f1-4fce-8c06-a69561ff8a6d","Type":"ContainerDied","Data":"ab600e16b7115731e5996d1f6491119cc9b18fcff9ff9742d61072463acfb2f4"} Oct 06 11:14:14 crc kubenswrapper[4824]: I1006 11:14:14.443832 4824 scope.go:117] "RemoveContainer" containerID="4fa20ecee3b1c9e1a0579597134d4ac9934abd73fbbea1c39577d842a5d800c4" Oct 06 11:14:14 crc kubenswrapper[4824]: I1006 11:14:14.444709 4824 scope.go:117] "RemoveContainer" containerID="ab600e16b7115731e5996d1f6491119cc9b18fcff9ff9742d61072463acfb2f4" Oct 06 11:14:14 crc kubenswrapper[4824]: E1006 11:14:14.445224 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-khgzw_openshift-machine-config-operator(e1140c8c-93f1-4fce-8c06-a69561ff8a6d)\"" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" Oct 06 11:14:26 crc kubenswrapper[4824]: I1006 11:14:26.275134 4824 scope.go:117] "RemoveContainer" containerID="ab600e16b7115731e5996d1f6491119cc9b18fcff9ff9742d61072463acfb2f4" Oct 06 11:14:26 crc kubenswrapper[4824]: E1006 11:14:26.275985 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-khgzw_openshift-machine-config-operator(e1140c8c-93f1-4fce-8c06-a69561ff8a6d)\"" pod="openshift-machine-config-operator/machine-config-daemon-khgzw" podUID="e1140c8c-93f1-4fce-8c06-a69561ff8a6d" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515070722237024452 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015070722237017367 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015070710752016511 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015070710752015461 5ustar corecore